首页 > 最新文献

2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)最新文献

英文 中文
Discovering Future Malware Variants By Generating New Malware Samples Using Generative Adversarial Network 通过使用生成对抗网络生成新的恶意软件样本来发现未来的恶意软件变体
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8964913
Zahra Moti, S. Hashemi, Amir Namavar
Detecting malware sample is one of the most important issues in computer security. Malware variants are growing exponentially by more usage of computer in industries, homes, and other places. Among different types of malware samples, zero-day samples are more challenging. The conventional antivirus systems, which rely on known malware patterns, cannot detect zero-day samples since did not see them before. As reported in [1], in 2018, 76% of successful attacks on organization endpoints were based on zero-day samples. Therefore, predicting these types of attacks and preparing a solution is an open challenge.This paper presents a deep generative adversarial network to generate the signature of unseen malware samples; The generated signature is potentially similar to the malware samples that may be released in the future. After generating the samples, these generated data were added to the dataset to train a robust classifier against new variants of malware. Also, neural network is applied for extracting high-level features from raw bytes for detection. In the proposed method, only the header of the executable file was used for detection, which is a small piece of the file that contains some information about the file. To validate our method, we used three classification algorithms and classified the raw and new representation using them. Also, we compared our work with another malware detection using the PE header. The results of this paper show that the generated data improves the accuracy of classification algorithms by at least 1%.
恶意软件样本检测是计算机安全领域的重要课题之一。随着工业、家庭和其他地方越来越多地使用计算机,恶意软件变种呈指数级增长。在不同类型的恶意软件样本中,零日样本更具挑战性。传统的反病毒系统依赖于已知的恶意软件模式,无法检测到零日样本,因为之前没有看到它们。据[1]报道,2018年,76%的对组织端点的成功攻击是基于零日样本的。因此,预测这些类型的攻击并准备解决方案是一个公开的挑战。本文提出了一种深度生成对抗网络来生成不可见恶意软件样本的签名;生成的签名可能与将来可能发布的恶意软件样本相似。在生成样本后,这些生成的数据被添加到数据集中,以训练针对新恶意软件变体的鲁棒分类器。同时,利用神经网络从原始字节中提取高级特征进行检测。在提出的方法中,仅使用可执行文件的头文件进行检测,头文件是文件的一小部分,包含有关文件的一些信息。为了验证我们的方法,我们使用了三种分类算法,并使用它们对原始表示和新表示进行了分类。此外,我们还将我们的工作与使用PE头的另一种恶意软件检测进行了比较。本文的结果表明,生成的数据使分类算法的准确率提高了至少1%。
{"title":"Discovering Future Malware Variants By Generating New Malware Samples Using Generative Adversarial Network","authors":"Zahra Moti, S. Hashemi, Amir Namavar","doi":"10.1109/ICCKE48569.2019.8964913","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8964913","url":null,"abstract":"Detecting malware sample is one of the most important issues in computer security. Malware variants are growing exponentially by more usage of computer in industries, homes, and other places. Among different types of malware samples, zero-day samples are more challenging. The conventional antivirus systems, which rely on known malware patterns, cannot detect zero-day samples since did not see them before. As reported in [1], in 2018, 76% of successful attacks on organization endpoints were based on zero-day samples. Therefore, predicting these types of attacks and preparing a solution is an open challenge.This paper presents a deep generative adversarial network to generate the signature of unseen malware samples; The generated signature is potentially similar to the malware samples that may be released in the future. After generating the samples, these generated data were added to the dataset to train a robust classifier against new variants of malware. Also, neural network is applied for extracting high-level features from raw bytes for detection. In the proposed method, only the header of the executable file was used for detection, which is a small piece of the file that contains some information about the file. To validate our method, we used three classification algorithms and classified the raw and new representation using them. Also, we compared our work with another malware detection using the PE header. The results of this paper show that the generated data improves the accuracy of classification algorithms by at least 1%.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"20 1","pages":"319-324"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88121241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Detection and Counting of Lipid Droplets in Adipocyte Differentiation of Bone Marrow-Derived Mesenchymal Stem Cells Using a Tiny Convolutional Network and Image Processing 基于微卷积网络和图像处理的骨髓间充质干细胞脂肪细胞分化过程中脂滴的检测与计数
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8965200
Leila Hassanlou, S. Meshgini, E. Alizadeh, A. Farzamnia
Stem cells are a bunch of cells that are considered as encouraging cells for treating patients because of their ability to regenerate themselves and also their potential for differentiation into several lineages. When stem cells are differentiated into adipose tissues, a great variety of lipid droplets usually grow in these cells and can be observed by oil red O staining, which is typically used for evaluating adipocyte differentiation status. For numerous differentiation experiments, counting and calculation of the population of lipid droplets are necessary. The disadvantages of conducting experiments for identification and investigation of lipid droplets include being expensive, time-consuming and subjective. There are few studies carried out in the field of machine learning and image processing for the automatic detection and counting of lipid droplets in intracellular images. In this study, to demonstrate the adipocyte differentiation of mesenchymal stem cells, their microscopic images were prepared. After the preprocessing operation, the images were fed to a tiny convolutional neural network. Images created within the network output were examined using two image processing methods. Finally, the number of lipid droplets was obtained with acceptable accuracy, and their exact location was displayed.
干细胞是一群被认为是治疗病人的鼓励细胞,因为它们有自我再生的能力,也有分化成几个谱系的潜力。当干细胞分化为脂肪组织时,通常会在这些细胞中生长各种各样的脂滴,油红O染色可以观察到这些脂滴,通常用于评估脂肪细胞分化状态。对于大量的分化实验,脂滴数量的计数和计算是必要的。进行脂滴鉴定和研究实验的缺点是昂贵、耗时和主观。对于细胞内图像中脂滴的自动检测和计数,在机器学习和图像处理领域的研究很少。为了证明间充质干细胞的脂肪细胞分化,本研究制备了间充质干细胞的显微图像。经过预处理操作后,图像被送入一个微小的卷积神经网络。使用两种图像处理方法检查在网络输出中创建的图像。最后,获得了精度可接受的脂滴数量,并显示了它们的确切位置。
{"title":"Detection and Counting of Lipid Droplets in Adipocyte Differentiation of Bone Marrow-Derived Mesenchymal Stem Cells Using a Tiny Convolutional Network and Image Processing","authors":"Leila Hassanlou, S. Meshgini, E. Alizadeh, A. Farzamnia","doi":"10.1109/ICCKE48569.2019.8965200","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8965200","url":null,"abstract":"Stem cells are a bunch of cells that are considered as encouraging cells for treating patients because of their ability to regenerate themselves and also their potential for differentiation into several lineages. When stem cells are differentiated into adipose tissues, a great variety of lipid droplets usually grow in these cells and can be observed by oil red O staining, which is typically used for evaluating adipocyte differentiation status. For numerous differentiation experiments, counting and calculation of the population of lipid droplets are necessary. The disadvantages of conducting experiments for identification and investigation of lipid droplets include being expensive, time-consuming and subjective. There are few studies carried out in the field of machine learning and image processing for the automatic detection and counting of lipid droplets in intracellular images. In this study, to demonstrate the adipocyte differentiation of mesenchymal stem cells, their microscopic images were prepared. After the preprocessing operation, the images were fed to a tiny convolutional neural network. Images created within the network output were examined using two image processing methods. Finally, the number of lipid droplets was obtained with acceptable accuracy, and their exact location was displayed.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"24 1","pages":"176-181"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87854943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Case Study for Presenting Bank Recommender Systems based on Bon Card Transaction Data 基于银行卡交易数据的银行推荐系统案例研究
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8964698
Abdorreza Sharifihosseini
As with many other businesses, banking industry tends to digitalize its working processes and use state-of-the-art technique in the financial and commercial areas in its business. The main core of the bank business is managing communication with customers which eventually results in investment on customers. In this paper, the structure of a recommender system is described, whereby using the recommender technology the places for purchase in which so far, the customers have not used any special type of Bon cards but are probable to buy from them is estimated and proposed to the customers.Matrix factorization is a type of method for collaborative filtering based on models which is widely used for rating prediction concept. Generally, bank products are not rated by customers; these products are usually purchased or offered to customers by the bank. Therefore, to determine the rating, RFM 1 method which is an instrument for analysis in marketing is used along with clustering algorithm to determine the customer value and place. If a place does not have any value, i.e. the data have missing values, it suggests that we do not know whether the customer prefers the place for purchase or not. In this paper, a hybrid method based on dimension reduction technique is presented. This method is able to predict the missing values in data to offer recommendation to customers. Assessment of the proposed model through Root Mean Square Error 2 indicates that the architecture in this paper has less error in comparison to common collaborative filtering methods.
与许多其他业务一样,银行业倾向于将其工作流程数字化,并在其业务的金融和商业领域使用最先进的技术。银行业务的核心是管理与客户的沟通,最终形成对客户的投资。本文描述了一个推荐系统的结构,通过使用推荐技术,对顾客目前还没有使用过任何特殊类型的Bon卡的购买地点进行估计,并向顾客提出可能购买的地点。矩阵分解是一种基于模型的协同过滤方法,广泛应用于评级预测概念。一般来说,银行产品没有客户评级;这些产品通常由银行购买或提供给客户。因此,为了确定评级,使用市场营销中的分析工具RFM 1方法与聚类算法一起确定客户价值和位置。如果一个地方没有任何值,即数据缺少值,则表明我们不知道客户是否喜欢购买该地方。本文提出了一种基于降维技术的混合方法。该方法能够预测数据中的缺失值,为客户提供推荐。通过均方根误差2对所提出的模型进行评估表明,与常见的协同过滤方法相比,本文的架构具有更小的误差。
{"title":"A Case Study for Presenting Bank Recommender Systems based on Bon Card Transaction Data","authors":"Abdorreza Sharifihosseini","doi":"10.1109/ICCKE48569.2019.8964698","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8964698","url":null,"abstract":"As with many other businesses, banking industry tends to digitalize its working processes and use state-of-the-art technique in the financial and commercial areas in its business. The main core of the bank business is managing communication with customers which eventually results in investment on customers. In this paper, the structure of a recommender system is described, whereby using the recommender technology the places for purchase in which so far, the customers have not used any special type of Bon cards but are probable to buy from them is estimated and proposed to the customers.Matrix factorization is a type of method for collaborative filtering based on models which is widely used for rating prediction concept. Generally, bank products are not rated by customers; these products are usually purchased or offered to customers by the bank. Therefore, to determine the rating, RFM 1 method which is an instrument for analysis in marketing is used along with clustering algorithm to determine the customer value and place. If a place does not have any value, i.e. the data have missing values, it suggests that we do not know whether the customer prefers the place for purchase or not. In this paper, a hybrid method based on dimension reduction technique is presented. This method is able to predict the missing values in data to offer recommendation to customers. Assessment of the proposed model through Root Mean Square Error 2 indicates that the architecture in this paper has less error in comparison to common collaborative filtering methods.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"29 1","pages":"72-77"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73598463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
List Scheduling for Heterogeneous Computing Systems Introducing a Performance-Effective Definition for Critical Path 异构计算系统的列表调度——引入关键路径的性能有效定义
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8964738
Farzam Dorostkar, S. Mirzakuchaki
The emergence of heterogeneous computing systems has been accompanied by serious design issues. Being a highly influential factor on performance in these systems, application scheduling is one of the major design considerations. In this paper, we propose a new critical path-oriented list scheduling heuristic algorithm called Communication-Intensive Path on a Processor (CIPOP) for heterogeneous computing environments. It is a modification of the well-known CPOP algorithm that presented the idea of scheduling the most costly entry-exit path of tasks, commonly known as the critical path, on a single processor. Generally, this processor selection strategy has different potential impacts on computation and communication costs along a selected path in the produced schedule. However, these probably different effects are not considered in the common definition of a critical path. Differentiating between these two types of costs, the proposed algorithm introduces a novel performance-effective definition for a critical path that is compatible with this processor selection strategy. CIPOP has a time complexity the same as that of the state-of-the-art list scheduling heuristic algorithms, which is of the order O(v2.× p) for v tasks and p processors. The conducted comprehensive experiment based on a wide variety of randomly generated application DAGs demonstrates the performance improvement of the proposed algorithm.
异构计算系统的出现伴随着严重的设计问题。在这些系统中,应用程序调度是对性能影响很大的因素,是主要的设计考虑因素之一。在本文中,我们提出了一种新的面向关键路径的列表调度启发式算法,称为处理器上的通信密集型路径(CIPOP)。它是对著名的CPOP算法的改进,CPOP算法提出了在单个处理器上调度成本最高的任务进出路径(通常称为关键路径)的思想。通常,这种处理器选择策略对生产调度中选定路径的计算和通信成本有不同的潜在影响。然而,这些可能不同的影响在关键路径的通用定义中没有被考虑。为了区分这两种类型的成本,该算法引入了一种新的性能有效的关键路径定义,该定义与该处理器选择策略兼容。CIPOP的时间复杂度与最先进的列表调度启发式算法相同,为O(v2)阶。xp)表示v个任务和p个处理器。基于各种随机生成的应用dag进行的综合实验证明了该算法的性能改进。
{"title":"List Scheduling for Heterogeneous Computing Systems Introducing a Performance-Effective Definition for Critical Path","authors":"Farzam Dorostkar, S. Mirzakuchaki","doi":"10.1109/ICCKE48569.2019.8964738","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8964738","url":null,"abstract":"The emergence of heterogeneous computing systems has been accompanied by serious design issues. Being a highly influential factor on performance in these systems, application scheduling is one of the major design considerations. In this paper, we propose a new critical path-oriented list scheduling heuristic algorithm called Communication-Intensive Path on a Processor (CIPOP) for heterogeneous computing environments. It is a modification of the well-known CPOP algorithm that presented the idea of scheduling the most costly entry-exit path of tasks, commonly known as the critical path, on a single processor. Generally, this processor selection strategy has different potential impacts on computation and communication costs along a selected path in the produced schedule. However, these probably different effects are not considered in the common definition of a critical path. Differentiating between these two types of costs, the proposed algorithm introduces a novel performance-effective definition for a critical path that is compatible with this processor selection strategy. CIPOP has a time complexity the same as that of the state-of-the-art list scheduling heuristic algorithms, which is of the order O(v2.× p) for v tasks and p processors. The conducted comprehensive experiment based on a wide variety of randomly generated application DAGs demonstrates the performance improvement of the proposed algorithm.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"10 1","pages":"356-362"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78915929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Predicting Serious Outcomes in Syncope Patients Using Data Mining Techniques 使用数据挖掘技术预测晕厥患者的严重后果
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8965047
Ardeshir Mansouri, Mohammad Ordikhani, M. S. Abadeh, Masih Tajdini
Syncope or fainting refers to a temporary loss of consciousness usually related to insufficient blood flow to the brain and can be due to several causes, which are either simple or serious conditions. Syncope can be caused by life-threatening conditions not evident in the first evaluations, which can lead to serious outcomes, including death, after discharge from the hospital. We have developed a decision tool to identify syncope patients with 18 years of age or higher who are at risk of a serious event within 30 days after discharge from the hospital.We used the data provided by the Tehran Heart Clinic. In this dataset adults with 18 years old or above with syncope signs are enrolled. The patients presented themselves within 24 hours after the event to the THC. Standardized variables from clinical evaluation and investigations have been collected. Serious adverse events included death, Intracerebral hemorrhage (ICH) or Subarachnoid hemorrhage (SAH), Cerebrovascular accident (CVA), Device Implantation, myocardial infarction, arrhythmia, traumatic syncope or cardiac surgery within 30 days. 356 patients were enrolled with syncope; the mean age was 44.5 years and 53.6% were women. Serious events occurred among 26 (7.3%) of the patients within 30 days of discharge from the hospital.Different machine learning algorithms such as Decision Tree, SMO, Neural Networks, Naïve Bayes and Random Forest have been used on the dataset to predict patients with serious adverse outcomes and the WEKA program has been used to validate the results.Results show that when using Random Forrest Algorithm, the accuracy rate and ROC Area reached 91.09% and 0.90. However, previous statistical risk scores such as the San Francisco Score resulted in lower ROC Area readings.
晕厥或昏厥指的是一种暂时性的意识丧失,通常与大脑供血不足有关,可能是由几种原因引起的,这些原因有简单的也有严重的。晕厥可由危及生命的疾病引起,在最初的评估中不明显,这可能导致出院后的严重后果,包括死亡。我们开发了一种决策工具来识别18岁或以上的晕厥患者,他们在出院后30天内有严重事件的风险。我们使用了德黑兰心脏诊所提供的数据。在这个数据集中,18岁或以上有晕厥症状的成年人被纳入。患者在事件发生后24小时内向THC自首。收集了来自临床评估和调查的标准化变量。严重不良事件包括30天内死亡、脑出血(ICH)或蛛网膜下腔出血(SAH)、脑血管意外(CVA)、器械植入、心肌梗死、心律失常、外伤性晕厥或心脏手术。356例晕厥患者入组;平均年龄44.5岁,女性53.6%。出院后30天内发生严重事件26例(7.3%)。不同的机器学习算法,如决策树,SMO,神经网络,Naïve贝叶斯和随机森林已经在数据集上使用,以预测严重不良后果的患者,并使用WEKA程序验证结果。结果表明,使用随机福雷斯特算法时,准确率达到91.09%,ROC面积达到0.90。然而,以前的统计风险评分,如旧金山评分导致较低的ROC区域读数。
{"title":"Predicting Serious Outcomes in Syncope Patients Using Data Mining Techniques","authors":"Ardeshir Mansouri, Mohammad Ordikhani, M. S. Abadeh, Masih Tajdini","doi":"10.1109/ICCKE48569.2019.8965047","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8965047","url":null,"abstract":"Syncope or fainting refers to a temporary loss of consciousness usually related to insufficient blood flow to the brain and can be due to several causes, which are either simple or serious conditions. Syncope can be caused by life-threatening conditions not evident in the first evaluations, which can lead to serious outcomes, including death, after discharge from the hospital. We have developed a decision tool to identify syncope patients with 18 years of age or higher who are at risk of a serious event within 30 days after discharge from the hospital.We used the data provided by the Tehran Heart Clinic. In this dataset adults with 18 years old or above with syncope signs are enrolled. The patients presented themselves within 24 hours after the event to the THC. Standardized variables from clinical evaluation and investigations have been collected. Serious adverse events included death, Intracerebral hemorrhage (ICH) or Subarachnoid hemorrhage (SAH), Cerebrovascular accident (CVA), Device Implantation, myocardial infarction, arrhythmia, traumatic syncope or cardiac surgery within 30 days. 356 patients were enrolled with syncope; the mean age was 44.5 years and 53.6% were women. Serious events occurred among 26 (7.3%) of the patients within 30 days of discharge from the hospital.Different machine learning algorithms such as Decision Tree, SMO, Neural Networks, Naïve Bayes and Random Forest have been used on the dataset to predict patients with serious adverse outcomes and the WEKA program has been used to validate the results.Results show that when using Random Forrest Algorithm, the accuracy rate and ROC Area reached 91.09% and 0.90. However, previous statistical risk scores such as the San Francisco Score resulted in lower ROC Area readings.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"18 1","pages":"409-413"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80297258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Metasearch engine result optimization using reformed genetic algorithm 基于改进遗传算法的元搜索引擎结果优化
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8964735
Somayeh Adeli, M. P. Aghababa
Metasearch engine is a system that applies several different search engines, merges the returned results from the search engines and presents the best results. Principal component of the metasearch engine is the method applied for merging the given results. The most of existing merging algorithms are relied on the information achieved by ranking scores which is integrated with the results of different search engines. In this paper, a reformed genetic algorithm (RGA) is proposed for aggregating results of different search engines. In the RGA, a chaotic sequence is applied to select the parents to mate, preventing the RGA to get stuck in local optima. The combination of pitch adjustment rule and uniform crossover (CPARU) is also proposed to further mutate of chromosomes. In the problem of optimizing search engine results, the proposed method tries to find weights of documents’ place to allocate each document to the best place. Therefore, the only required information is to know the number of the search engines that finds each document in the corresponding place. Accordingly, this technique works independently of the different search engines’ ranking scores. The experimental results have depicted that the RGA outperforms the genetic algorithm (GA), Borda method, Kendall-tau genetic algorithm (GKTu) and Spearmen's footrule genetic algorithm (GSFD) methods.
元搜索引擎是一个应用几个不同的搜索引擎,合并从搜索引擎返回的结果,并呈现最佳结果的系统。元搜索引擎的主成分是用于合并给定结果的方法。现有的合并算法大多依赖于排序分数所获得的信息,这些信息与不同搜索引擎的结果相结合。本文提出了一种改进的遗传算法(RGA),用于聚合不同搜索引擎的搜索结果。在RGA中,采用混沌序列选择亲本进行交配,避免了RGA陷入局部最优状态。还提出了基音调整规则和均匀交叉(CPARU)相结合的方法来进一步实现染色体的突变。在优化搜索引擎结果的问题中,该方法试图找到文档位置的权重,将每个文档分配到最佳位置。因此,唯一需要的信息是知道在相应位置找到每个文档的搜索引擎的数量。因此,这种技术独立于不同搜索引擎的排名分数而工作。实验结果表明,RGA优于遗传算法(GA)、Borda方法、Kendall-tau遗传算法(GKTu)和Spearmen's footrule遗传算法(GSFD)。
{"title":"Metasearch engine result optimization using reformed genetic algorithm","authors":"Somayeh Adeli, M. P. Aghababa","doi":"10.1109/ICCKE48569.2019.8964735","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8964735","url":null,"abstract":"Metasearch engine is a system that applies several different search engines, merges the returned results from the search engines and presents the best results. Principal component of the metasearch engine is the method applied for merging the given results. The most of existing merging algorithms are relied on the information achieved by ranking scores which is integrated with the results of different search engines. In this paper, a reformed genetic algorithm (RGA) is proposed for aggregating results of different search engines. In the RGA, a chaotic sequence is applied to select the parents to mate, preventing the RGA to get stuck in local optima. The combination of pitch adjustment rule and uniform crossover (CPARU) is also proposed to further mutate of chromosomes. In the problem of optimizing search engine results, the proposed method tries to find weights of documents’ place to allocate each document to the best place. Therefore, the only required information is to know the number of the search engines that finds each document in the corresponding place. Accordingly, this technique works independently of the different search engines’ ranking scores. The experimental results have depicted that the RGA outperforms the genetic algorithm (GA), Borda method, Kendall-tau genetic algorithm (GKTu) and Spearmen's footrule genetic algorithm (GSFD) methods.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"37 1","pages":"18-25"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80213011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Object Tracking Methods:A Review 对象跟踪方法综述
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8964761
Zahra Soleimanitaleb, Mohammad Ali Keyvanrad, Ali Jafari
Object tracking is one of the most important tasks in computer vision that has many practical applications such as traffic monitoring, robotics, autonomous vehicle tracking, and so on. Different researches have been done in recent years, but because of different challenges such as occlusion, illumination variations, fast motion, etc. researches in this area continues. In this paper, various methods of tracking objects are examined and a comprehensive classification is presented that classified tracking methods into four main categories of feature-based, segmentation-based, estimation-based, and learning-based methods that each of which has its own sub-categories. The main focus of this paper is on learning-based methods, which are classified into three categories of generative methods, discriminative methods, and reinforcement learning. One of the sub-categories of the discriminative model is deep learning. Because of high-performance, deep learning has recently been very much considered.
目标跟踪是计算机视觉中最重要的任务之一,在交通监控、机器人、自动驾驶车辆跟踪等领域有许多实际应用。近年来,由于遮挡、光照变化、快速运动等不同的挑战,该领域的研究仍在继续。本文研究了各种跟踪对象的方法,并提出了一种全面的分类方法,将跟踪方法分为基于特征的、基于分割的、基于估计的和基于学习的四大类,每种方法都有自己的子类别。本文主要关注基于学习的方法,将其分为三类:生成方法、判别方法和强化学习。判别模型的一个子类别是深度学习。由于其高性能,深度学习最近受到了广泛的关注。
{"title":"Object Tracking Methods:A Review","authors":"Zahra Soleimanitaleb, Mohammad Ali Keyvanrad, Ali Jafari","doi":"10.1109/ICCKE48569.2019.8964761","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8964761","url":null,"abstract":"Object tracking is one of the most important tasks in computer vision that has many practical applications such as traffic monitoring, robotics, autonomous vehicle tracking, and so on. Different researches have been done in recent years, but because of different challenges such as occlusion, illumination variations, fast motion, etc. researches in this area continues. In this paper, various methods of tracking objects are examined and a comprehensive classification is presented that classified tracking methods into four main categories of feature-based, segmentation-based, estimation-based, and learning-based methods that each of which has its own sub-categories. The main focus of this paper is on learning-based methods, which are classified into three categories of generative methods, discriminative methods, and reinforcement learning. One of the sub-categories of the discriminative model is deep learning. Because of high-performance, deep learning has recently been very much considered.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"1 1","pages":"282-288"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90869203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Automatic Classification of Galaxies Based on SVM 基于SVM的星系自动分类
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8965020
A. Bastanfard, Dariush Amirkhani, Moslem abbasiasl
Viewing heavenly objects in the sky helps astronomers understand how the world is shaped. Regarding the large number of objects observed by modern telescopes, it is very difficult to manually analyze it manually. An important part of galactic research is classification based on Hubble's design. The purpose of this research is to classify images of the stars using machine learning and neural networks. Particularly in this study, the galaxy's image is employed. The galaxies are divided into regular two-dimensional Hubble designs and an irregular bunch. The regular bands that are presented in the shape of the Hubble design are divided into two distinct spiral and elliptical galaxies. Spiral galaxies can be considered as elliptical or circular galaxies depending on the shape of the spiral, so the identification or classification of the spiral galaxy is considered important from other galaxies. In the proposed algorithm, the Sloan Digital Sky is used for testing, including 570 images. In the first step, its preprocessing operation is performed to remove image noise. In the next step, extracting the attribute from the galactic images takes place in a total of 827 properties using the sub-windows, the moments of different color spaces and the properties of the local configuration patterns. Then the classification is performed after extracting the property using a Support vector machine. And then compared with other methods, which indicate that our approach has worked better. In this study, the experiments were carried out in two spiral and elliptic classes and three spiral, elliptic and zinc-edged classes with a precision of 96 and 94 respectively.
观察天空中的天体有助于天文学家了解世界是如何形成的。对于现代望远镜观测到的大量天体,人工分析是非常困难的。银河研究的一个重要部分是基于哈勃设计的分类。这项研究的目的是使用机器学习和神经网络对恒星的图像进行分类。特别是在这项研究中,星系的图像被使用。这些星系被分为规则的二维哈勃星系和不规则的星系群。以哈勃设计的形状呈现的规则带分为两个不同的螺旋星系和椭圆星系。根据螺旋星系的形状,螺旋星系可以被认为是椭圆星系或圆形星系,因此将螺旋星系与其他星系区分开来被认为是重要的。在算法中,使用斯隆数字天空进行测试,包括570张图像。第一步,对图像进行预处理,去除图像噪声。下一步,利用子窗口、不同颜色空间的矩和局部构型模式的属性,从星系图像中提取总共827个属性。然后使用支持向量机提取属性后进行分类。并与其他方法进行了比较,结果表明我们的方法效果更好。本研究采用2个螺旋椭圆类和3个螺旋椭圆锌边类进行实验,实验精度分别为96和94。
{"title":"Automatic Classification of Galaxies Based on SVM","authors":"A. Bastanfard, Dariush Amirkhani, Moslem abbasiasl","doi":"10.1109/ICCKE48569.2019.8965020","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8965020","url":null,"abstract":"Viewing heavenly objects in the sky helps astronomers understand how the world is shaped. Regarding the large number of objects observed by modern telescopes, it is very difficult to manually analyze it manually. An important part of galactic research is classification based on Hubble's design. The purpose of this research is to classify images of the stars using machine learning and neural networks. Particularly in this study, the galaxy's image is employed. The galaxies are divided into regular two-dimensional Hubble designs and an irregular bunch. The regular bands that are presented in the shape of the Hubble design are divided into two distinct spiral and elliptical galaxies. Spiral galaxies can be considered as elliptical or circular galaxies depending on the shape of the spiral, so the identification or classification of the spiral galaxy is considered important from other galaxies. In the proposed algorithm, the Sloan Digital Sky is used for testing, including 570 images. In the first step, its preprocessing operation is performed to remove image noise. In the next step, extracting the attribute from the galactic images takes place in a total of 827 properties using the sub-windows, the moments of different color spaces and the properties of the local configuration patterns. Then the classification is performed after extracting the property using a Support vector machine. And then compared with other methods, which indicate that our approach has worked better. In this study, the experiments were carried out in two spiral and elliptic classes and three spiral, elliptic and zinc-edged classes with a precision of 96 and 94 respectively.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"1 1","pages":"32-39"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84367839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Resource Provisioning in IaaS Clouds; Auto-Scale RAM memory issue IaaS云中的资源配置自动缩放RAM内存问题
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8964932
Zolfaghar Salmanian, Habib Izadkhah, A. Isazadeh
In the Infrastructure-as-a-Service model of the cloud computing paradigm, virtual machines are deployed on bare-metal servers called hosts. The host is responsible for the allocation of required resources such as CPU, RAM memory, and network bandwidth for the virtual machine. Thus, the problem of resource allocation reduces to how to place the virtual machines on physical hosts. In this paper, we propose CTMC modeling based on the birth-death process of the queueing systems for the performance of the data center. We will focus on RAM allocation for virtual machines. In this architecture, a job is defined as RAM assignment for a virtual machine. Job arrivals and their service times are assumed to be based on the Poisson process and exponential distribution, respectively. The purpose of this modeling is to keep the number of running hosts minimal in a scalable datacenter while the quality of service in terms of response time is acceptable due to system utilization.
在云计算范式的基础设施即服务模型中,虚拟机部署在称为主机的裸机服务器上。主机负责分配虚拟机所需的资源,如CPU、RAM、网络带宽等。因此,资源分配问题减少到如何将虚拟机放置在物理主机上。本文提出了基于排队系统生灭过程的CTMC模型,以提高数据中心的性能。我们将重点讨论虚拟机的RAM分配。在这个体系结构中,作业被定义为虚拟机的RAM分配。假设工作到达和服务时间分别基于泊松过程和指数分布。此建模的目的是在可伸缩数据中心中保持运行主机的数量最少,同时由于系统利用率,响应时间方面的服务质量是可接受的。
{"title":"Resource Provisioning in IaaS Clouds; Auto-Scale RAM memory issue","authors":"Zolfaghar Salmanian, Habib Izadkhah, A. Isazadeh","doi":"10.1109/ICCKE48569.2019.8964932","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8964932","url":null,"abstract":"In the Infrastructure-as-a-Service model of the cloud computing paradigm, virtual machines are deployed on bare-metal servers called hosts. The host is responsible for the allocation of required resources such as CPU, RAM memory, and network bandwidth for the virtual machine. Thus, the problem of resource allocation reduces to how to place the virtual machines on physical hosts. In this paper, we propose CTMC modeling based on the birth-death process of the queueing systems for the performance of the data center. We will focus on RAM allocation for virtual machines. In this architecture, a job is defined as RAM assignment for a virtual machine. Job arrivals and their service times are assumed to be based on the Poisson process and exponential distribution, respectively. The purpose of this modeling is to keep the number of running hosts minimal in a scalable datacenter while the quality of service in terms of response time is acceptable due to system utilization.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"48 1","pages":"455-460"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81795000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
POI Recommendation Based on Heterogeneous Graph Embedding 基于异构图嵌入的POI推荐
Pub Date : 2019-10-01 DOI: 10.1109/ICCKE48569.2019.8964762
Sima Naderi Mighan, M. Kahani, F. Pourgholamali
With the development and popularity of social networks, many human beings prefer to share their experiences on these networks. There are various methods proposed by the researcher which utilized user-generated content in the location-based social networks (LBSN) and recommend locations to users. However, there is a high sparsity in the user check-in information makes it tough to recommend the appropriate and accurate location to the user. To fix this issue, we put forward a proposal as a framework which utilizes a wide range of information available in these networks, each of which has its own type and provides appropriate recommendation. For this purpose, we encode the information as a number of entities and its attributes in the form of a heterogeneous graph, then graph embedding methods are used to embed all nodes in unified semantic representation space. As a result, we are able to model relations between users and venues in an efficient way and ameliorate the accuracy of the proposed method that recommends a place to a user. Our method is implemented and evaluated using Foursquare dataset, and the evaluation results depict that our work, boost performance in terms of precision, recall, and f-measure compared to the baseline work.
随着社交网络的发展和普及,许多人喜欢在这些网络上分享他们的经历。研究者提出了多种方法,利用基于位置的社交网络(LBSN)中的用户生成内容,向用户推荐位置。然而,用户签入信息的高度稀疏性使得向用户推荐适当和准确的位置变得困难。为了解决这个问题,我们提出了一个建议,作为一个框架,利用这些网络中可用的广泛信息,每个网络都有自己的类型,并提供适当的建议。为此,我们以异构图的形式将信息编码为多个实体及其属性,然后使用图嵌入方法将所有节点嵌入到统一的语义表示空间中。因此,我们能够以一种有效的方式对用户和场地之间的关系进行建模,并提高向用户推荐地点的方法的准确性。我们的方法是使用Foursquare数据集实现和评估的,评估结果表明,与基线工作相比,我们的工作在精度、召回率和f-measure方面提高了性能。
{"title":"POI Recommendation Based on Heterogeneous Graph Embedding","authors":"Sima Naderi Mighan, M. Kahani, F. Pourgholamali","doi":"10.1109/ICCKE48569.2019.8964762","DOIUrl":"https://doi.org/10.1109/ICCKE48569.2019.8964762","url":null,"abstract":"With the development and popularity of social networks, many human beings prefer to share their experiences on these networks. There are various methods proposed by the researcher which utilized user-generated content in the location-based social networks (LBSN) and recommend locations to users. However, there is a high sparsity in the user check-in information makes it tough to recommend the appropriate and accurate location to the user. To fix this issue, we put forward a proposal as a framework which utilizes a wide range of information available in these networks, each of which has its own type and provides appropriate recommendation. For this purpose, we encode the information as a number of entities and its attributes in the form of a heterogeneous graph, then graph embedding methods are used to embed all nodes in unified semantic representation space. As a result, we are able to model relations between users and venues in an efficient way and ameliorate the accuracy of the proposed method that recommends a place to a user. Our method is implemented and evaluated using Foursquare dataset, and the evaluation results depict that our work, boost performance in terms of precision, recall, and f-measure compared to the baseline work.","PeriodicalId":6685,"journal":{"name":"2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"1 1","pages":"188-193"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83943320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2019 9th International Conference on Computer and Knowledge Engineering (ICCKE)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1