首页 > 最新文献

2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)最新文献

英文 中文
Cryptocurrency Trading Agent Using Deep Reinforcement Learning 使用深度强化学习的加密货币交易代理
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068485
Uwais Suliman, Terence L. van Zyl, A. Paskaramoorthy
Cryptocurrencies are peer-to-peer digital assets monitored and organised by a blockchain network. Price prediction has been a significant focus point with various machine learning algorithms, especially concerning cryptocurrency. This work addresses the challenge faced by traders of short-term profit maximisation. The study presents a deep reinforcement learning algorithm to trade in cryptocurrency markets, Duelling DQN. The environment has been designed to simulate actual trading behaviour, observing historical price movements and taking action on real-time prices. The proposed algorithm was tested with Bitcoin, Ethereum, and Litecoin. The respective portfolio returns are used as a metric to measure the algorithm's performance against the buy-and-hold benchmark, with the buy-and-hold outperforming the results produced by the Duelling DQN agent.
加密货币是由区块链网络监控和组织的点对点数字资产。价格预测一直是各种机器学习算法的一个重要焦点,特别是在加密货币方面。这项工作解决了交易者面临的短期利润最大化的挑战。该研究提出了一种用于加密货币市场交易的深度强化学习算法Duelling DQN。该环境旨在模拟实际交易行为,观察历史价格走势并对实时价格采取行动。提出的算法在比特币、以太坊和莱特币上进行了测试。各自的投资组合回报被用作衡量算法相对于买入并持有基准的表现的指标,买入并持有的表现优于Duelling DQN代理产生的结果。
{"title":"Cryptocurrency Trading Agent Using Deep Reinforcement Learning","authors":"Uwais Suliman, Terence L. van Zyl, A. Paskaramoorthy","doi":"10.1109/ISCMI56532.2022.10068485","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068485","url":null,"abstract":"Cryptocurrencies are peer-to-peer digital assets monitored and organised by a blockchain network. Price prediction has been a significant focus point with various machine learning algorithms, especially concerning cryptocurrency. This work addresses the challenge faced by traders of short-term profit maximisation. The study presents a deep reinforcement learning algorithm to trade in cryptocurrency markets, Duelling DQN. The environment has been designed to simulate actual trading behaviour, observing historical price movements and taking action on real-time prices. The proposed algorithm was tested with Bitcoin, Ethereum, and Litecoin. The respective portfolio returns are used as a metric to measure the algorithm's performance against the buy-and-hold benchmark, with the buy-and-hold outperforming the results produced by the Duelling DQN agent.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115109828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fake News Detection Using Deep Learning and Natural Language Processing 使用深度学习和自然语言处理的假新闻检测
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068440
Anand Matheven, B. V. D. Kumar
The rise of social media has brought the rise of fake news and this fake news comes with negative consequences. With fake news being such a huge issue, efforts should be made to identify any forms of fake news however it is not so simple. Manually identifying fake news can be extremely subjective as determining the accuracy of the information in a story is complex and difficult to perform, even for experts. On the other hand, an automated solution would require a good understanding of NLP which is also complex and may have difficulties producing an accurate output. Therefore, the main problem focused on this project is the viability of developing a system that can effectively and accurately detect and identify fake news. Finding a solution would be a significant benefit to the media industry, particularly the social media industry as this is where a large proportion of fake news is published and spread. In order to find a solution to this problem, this project proposed the development of a fake news identification system using deep learning and natural language processing. The system was developed using a Word2vec model combined with a Long Short-Term Memory model in order to showcase the compatibility of the two models in a whole system. This system was trained and tested using two different dataset collections that each consisted of one real news dataset and one fake news dataset. Furthermore, three independent variables were chosen which were the number of training cycles, data diversity and vector size to analyze the relationship between these variables and the accuracy levels of the system. It was found that these three variables did have a significant effect on the accuracy of the system. From this, the system was then trained and tested with the optimal variables and was able to achieve the minimum expected accuracy level of 90%. The achieving of this accuracy levels confirms the compatibility of the LSTM and Word2vec model and their capability to be synergized into a single system that is able to identify fake news with a high level of accuracy.
社交媒体的兴起带来了假新闻的兴起,这些假新闻带来了负面后果。假新闻是一个巨大的问题,应该努力识别任何形式的假新闻,但这并不是那么简单。人工识别假新闻可能非常主观,因为确定新闻中信息的准确性非常复杂,即使是专家也很难做到。另一方面,自动化解决方案需要对NLP有很好的理解,这也很复杂,可能难以产生准确的输出。因此,本项目关注的主要问题是开发一个能够有效准确地检测和识别假新闻的系统的可行性。找到一个解决方案将对媒体行业,特别是社交媒体行业大有裨益,因为这是很大一部分假新闻发布和传播的地方。为了解决这一问题,本项目提出利用深度学习和自然语言处理技术开发假新闻识别系统。本系统采用Word2vec模型与长短期记忆模型相结合的方式开发,以展示两种模型在整个系统中的兼容性。该系统使用两个不同的数据集集进行训练和测试,每个数据集集由一个真实新闻数据集和一个假新闻数据集组成。进一步,选取训练周期数、数据多样性和向量大小三个自变量,分析这些变量与系统准确率水平之间的关系。结果表明,这三个变量对系统的精度有显著影响。然后,用最优变量对系统进行训练和测试,并能够达到90%的最低预期精度水平。这种准确性水平的实现证实了LSTM和Word2vec模型的兼容性,以及它们协同成一个能够以高准确性识别假新闻的单一系统的能力。
{"title":"Fake News Detection Using Deep Learning and Natural Language Processing","authors":"Anand Matheven, B. V. D. Kumar","doi":"10.1109/ISCMI56532.2022.10068440","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068440","url":null,"abstract":"The rise of social media has brought the rise of fake news and this fake news comes with negative consequences. With fake news being such a huge issue, efforts should be made to identify any forms of fake news however it is not so simple. Manually identifying fake news can be extremely subjective as determining the accuracy of the information in a story is complex and difficult to perform, even for experts. On the other hand, an automated solution would require a good understanding of NLP which is also complex and may have difficulties producing an accurate output. Therefore, the main problem focused on this project is the viability of developing a system that can effectively and accurately detect and identify fake news. Finding a solution would be a significant benefit to the media industry, particularly the social media industry as this is where a large proportion of fake news is published and spread. In order to find a solution to this problem, this project proposed the development of a fake news identification system using deep learning and natural language processing. The system was developed using a Word2vec model combined with a Long Short-Term Memory model in order to showcase the compatibility of the two models in a whole system. This system was trained and tested using two different dataset collections that each consisted of one real news dataset and one fake news dataset. Furthermore, three independent variables were chosen which were the number of training cycles, data diversity and vector size to analyze the relationship between these variables and the accuracy levels of the system. It was found that these three variables did have a significant effect on the accuracy of the system. From this, the system was then trained and tested with the optimal variables and was able to achieve the minimum expected accuracy level of 90%. The achieving of this accuracy levels confirms the compatibility of the LSTM and Word2vec model and their capability to be synergized into a single system that is able to identify fake news with a high level of accuracy.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114497320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Layer Sensitivity Aware CNN Quantization for Resource Constrained Edge Devices 资源受限边缘设备的分层敏感CNN量化
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068464
Alptekin Vardar, Li Zhang, Susu Hu, Saiyam Bherulal Jain, Shaown Mojumder, N. Laleni, A. Shrivastava, S. De, T. Kämpfe
Edge computing is rapidly becoming the defacto method for AI applications. However, the latency area and energy continue to be the main bottlenecks. To solve this problem, a hardware-aware approach has to be adopted. Quantizing the activations vastly reduces the number of Multiply-Accumulate (MAC) operations, resulting in with better latency and energy consumption while quantizing the weights decreases both memory footprint and the number of MAC operations, also helping with area reduction. In this paper, it is demonstrated that adapting an intra-layer mixed quantization training technique for both weights and activations, concerning layer sensitivities, in a Resnet-20 architecture with CIFAR-10 data set, a memory reduction of 73% can be achieved compared to even its all 8bits counterpart while sacrificing only around 2.3% accuracy. Moreover, it is demonstrated that, depending on the needs of the application, the balance between accuracy and resource usage can easily be arranged using different mixed-quantization schemes.
边缘计算正迅速成为人工智能应用的实际方法。然而,延迟面积和能量仍然是主要的瓶颈。为了解决这个问题,必须采用一种硬件感知的方法。量化激活大大减少了乘法累积(MAC)操作的数量,从而降低了延迟和能耗,同时量化权重减少了内存占用和MAC操作的数量,还有助于减少面积。在本文中,证明了在具有CIFAR-10数据集的Resnet-20架构中,对权重和激活采用层内混合量化训练技术,涉及层灵敏度,与所有8位相比,可以实现内存减少73%,同时仅牺牲约2.3%的准确性。此外,根据应用的需要,使用不同的混合量化方案可以很容易地安排精度和资源使用之间的平衡。
{"title":"Layer Sensitivity Aware CNN Quantization for Resource Constrained Edge Devices","authors":"Alptekin Vardar, Li Zhang, Susu Hu, Saiyam Bherulal Jain, Shaown Mojumder, N. Laleni, A. Shrivastava, S. De, T. Kämpfe","doi":"10.1109/ISCMI56532.2022.10068464","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068464","url":null,"abstract":"Edge computing is rapidly becoming the defacto method for AI applications. However, the latency area and energy continue to be the main bottlenecks. To solve this problem, a hardware-aware approach has to be adopted. Quantizing the activations vastly reduces the number of Multiply-Accumulate (MAC) operations, resulting in with better latency and energy consumption while quantizing the weights decreases both memory footprint and the number of MAC operations, also helping with area reduction. In this paper, it is demonstrated that adapting an intra-layer mixed quantization training technique for both weights and activations, concerning layer sensitivities, in a Resnet-20 architecture with CIFAR-10 data set, a memory reduction of 73% can be achieved compared to even its all 8bits counterpart while sacrificing only around 2.3% accuracy. Moreover, it is demonstrated that, depending on the needs of the application, the balance between accuracy and resource usage can easily be arranged using different mixed-quantization schemes.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123932277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pipeline Leakage Detection and Characterisation with Adaptive Surrogate Modelling Using Particle Swarm Optimisation 基于粒子群优化的自适应代理模型管道泄漏检测与表征
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068436
M. Adegboye, Aditya Karnik, W. Fung, R. Prabhu
Pipelines are often subject to leakage due to ageing, corrosion, and weld defects, and it is difficult to avoid as the sources of leakages are diverse. Several studies have demonstrated the applicability of the machine learning model for the timely prediction of pipeline leakage. However, most of these studies rely on a large training data set for training accurate models. The cost of collecting experimental data for model training is huge, while simulation data is computationally expensive and time-consuming. To tackle this problem, the present study proposes a novel data sampling optimisation method, named adaptive particle swarm optimisation (PSO) assisted surrogate model, which was used to train the machine learning models with a limited dataset and achieved good accuracy. The proposed model incorporates the population density of training data samples and model prediction fitness to determine new data samples for improved model fitting accuracy. The proposed method is applied to 3-D pipeline leakage detection and characterisation. The result shows that the predicted leak sizes and location match the actual leakage. The significance of this study is two-fold: the practical application allows for pipeline leak prediction with limited training samples and provides a general framework for computational efficiency improvement using adaptive surrogate modelling in various real-life applications.
管道经常因老化、腐蚀和焊接缺陷而发生泄漏,泄漏来源多种多样,难以避免。一些研究已经证明了机器学习模型在及时预测管道泄漏方面的适用性。然而,这些研究大多依赖于一个大的训练数据集来训练准确的模型。用于模型训练的实验数据的收集成本巨大,而仿真数据的计算成本高且耗时长。为了解决这一问题,本研究提出了一种新的数据采样优化方法——自适应粒子群优化(PSO)辅助代理模型,并将其用于有限数据集的机器学习模型的训练,取得了较好的精度。该模型结合训练数据样本的总体密度和模型预测适应度来确定新的数据样本,以提高模型拟合精度。将该方法应用于三维管道泄漏检测与表征。结果表明,预测的泄漏尺寸和泄漏位置与实际泄漏相吻合。这项研究的意义有两方面:实际应用允许在有限的训练样本下进行管道泄漏预测,并为在各种实际应用中使用自适应代理模型提高计算效率提供了一个总体框架。
{"title":"Pipeline Leakage Detection and Characterisation with Adaptive Surrogate Modelling Using Particle Swarm Optimisation","authors":"M. Adegboye, Aditya Karnik, W. Fung, R. Prabhu","doi":"10.1109/ISCMI56532.2022.10068436","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068436","url":null,"abstract":"Pipelines are often subject to leakage due to ageing, corrosion, and weld defects, and it is difficult to avoid as the sources of leakages are diverse. Several studies have demonstrated the applicability of the machine learning model for the timely prediction of pipeline leakage. However, most of these studies rely on a large training data set for training accurate models. The cost of collecting experimental data for model training is huge, while simulation data is computationally expensive and time-consuming. To tackle this problem, the present study proposes a novel data sampling optimisation method, named adaptive particle swarm optimisation (PSO) assisted surrogate model, which was used to train the machine learning models with a limited dataset and achieved good accuracy. The proposed model incorporates the population density of training data samples and model prediction fitness to determine new data samples for improved model fitting accuracy. The proposed method is applied to 3-D pipeline leakage detection and characterisation. The result shows that the predicted leak sizes and location match the actual leakage. The significance of this study is two-fold: the practical application allows for pipeline leak prediction with limited training samples and provides a general framework for computational efficiency improvement using adaptive surrogate modelling in various real-life applications.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121579236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modeling and Optimization of Two-Chamber Muffler by Genetic Algorithm 基于遗传算法的双腔消声器建模与优化
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068448
Jing-Fung Lin
In this study, Taguchi design method is used to optimize the acoustic performance of two-chamber muffler. The excellent parameter combination for high signal to noise ratio (S/N) of transmission loss (TL) is obtained by the range analysis, and influence sequence of four parameters on TL is determined. TL is evaluated by a COMSOL software based on the finite element method. Further, by the modification on the radius of hole in the baffle, a revised parameter combination for better S/N is found. Finally, the stepwise regression method is used to decide a statistically significant model with a high correlation coefficient. A potential muffler is obtained by the use of genetic algorithm and has high S/N ratio of 27.097 and average value of 33.21 dB for TL in a frequency range from 10 Hz to 1400 Hz.
本研究采用田口设计方法对双腔消声器的声学性能进行优化。通过极差分析,获得了传输损耗高信噪比(S/N)的最佳参数组合,确定了4个参数对传输损耗的影响顺序。利用COMSOL软件基于有限元法对其进行了TL计算。进一步,通过对挡板孔半径的修正,找到了更好信噪比的修正参数组合。最后,采用逐步回归方法,确定具有高相关系数的统计显著模型。利用遗传算法得到的潜在消声器在10 ~ 1400 Hz频率范围内具有27.097的高信噪比和33.21 dB的平均TL值。
{"title":"Modeling and Optimization of Two-Chamber Muffler by Genetic Algorithm","authors":"Jing-Fung Lin","doi":"10.1109/ISCMI56532.2022.10068448","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068448","url":null,"abstract":"In this study, Taguchi design method is used to optimize the acoustic performance of two-chamber muffler. The excellent parameter combination for high signal to noise ratio (S/N) of transmission loss (TL) is obtained by the range analysis, and influence sequence of four parameters on TL is determined. TL is evaluated by a COMSOL software based on the finite element method. Further, by the modification on the radius of hole in the baffle, a revised parameter combination for better S/N is found. Finally, the stepwise regression method is used to decide a statistically significant model with a high correlation coefficient. A potential muffler is obtained by the use of genetic algorithm and has high S/N ratio of 27.097 and average value of 33.21 dB for TL in a frequency range from 10 Hz to 1400 Hz.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"132 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114875669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Classification Performance for Android Small Sample Malicious Families Using Hybrid RGB Image Augmentation Method 基于混合RGB图像增强方法增强Android小样本恶意家族分类性能
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068453
Yi-Hsuan Ting, Yi-ming Chen, Li-Kai Chen
With the improvement of computer computing speed, many researches use deep learning for Android malware detection. In addition to malware detection, malware family classification will help malware researchers understand the behavior of the malware families to optimize detection and prevent However, the new malware family has few samples, which lead to bad classification results. GAN-based method can improve the classification results, but minor data will still lead to the unstable quality of the data generated by the deep learning augmentation method, which will limit the improvement of classification results. In the study, we will propose a hybrid augmentation method, first extracting malware features and converting them into RGB images, and then the minor families will augment by the gaussian noise augmentation method, and then combined with the deep convolutional generative adversarial network (DCGAN) which have better effect on image augmentation, and finally input to CNN for family classification. The experimental results show that using the hybrid augmentation method proposed in the study, compared to no augmentation and augmentation with only using the deep convolutional generative adversarial network, the F1-Score increased between 7%~34% and 2%~7%.
随着计算机计算速度的提高,许多研究将深度学习用于Android恶意软件检测。除了恶意软件检测之外,恶意软件家族分类还有助于恶意软件研究人员了解恶意软件家族的行为,从而优化检测和预防。然而,新的恶意软件家族样本较少,导致分类结果较差。基于gan的方法可以提高分类结果,但少量数据仍然会导致深度学习增强方法生成的数据质量不稳定,限制了分类结果的提高。在研究中,我们将提出一种混合增强方法,首先提取恶意软件特征并将其转换为RGB图像,然后通过高斯噪声增强方法对次要家族进行增强,然后结合对图像增强效果较好的深度卷积生成对抗网络(DCGAN),最后输入CNN进行家族分类。实验结果表明,采用本文提出的混合增强方法,与不增强和仅使用深度卷积生成对抗网络增强相比,F1-Score分别提高了7%~34%和2%~7%。
{"title":"Enhancing Classification Performance for Android Small Sample Malicious Families Using Hybrid RGB Image Augmentation Method","authors":"Yi-Hsuan Ting, Yi-ming Chen, Li-Kai Chen","doi":"10.1109/ISCMI56532.2022.10068453","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068453","url":null,"abstract":"With the improvement of computer computing speed, many researches use deep learning for Android malware detection. In addition to malware detection, malware family classification will help malware researchers understand the behavior of the malware families to optimize detection and prevent However, the new malware family has few samples, which lead to bad classification results. GAN-based method can improve the classification results, but minor data will still lead to the unstable quality of the data generated by the deep learning augmentation method, which will limit the improvement of classification results. In the study, we will propose a hybrid augmentation method, first extracting malware features and converting them into RGB images, and then the minor families will augment by the gaussian noise augmentation method, and then combined with the deep convolutional generative adversarial network (DCGAN) which have better effect on image augmentation, and finally input to CNN for family classification. The experimental results show that using the hybrid augmentation method proposed in the study, compared to no augmentation and augmentation with only using the deep convolutional generative adversarial network, the F1-Score increased between 7%~34% and 2%~7%.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134339489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Hybrid Joint Embedding with Intra-Modality Loss for Image-Text Matching 基于模态内损失的混合联合嵌入图像-文本匹配
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068471
Doaa B. Ebaid, A. El-Zoghabi, Magda M. Madbouly
Image-text(caption) matching has become a regular evaluation of joint-embedding models that combine vision and language. This task comprises ranking the data of one modality (images) based on a Text query (Image Retrieval), and ranking texts by relevance for an image query (Text Retrieval). The current joint embedding approaches use symmetric similarity measurement, due to that order embedding is not taken in consideration. In addition to that, in image-text matching, the used losses ignore the intra similarity in a certain modality that explores the relation between the candidates in the same modality. In spite of, the important role of intra information in the embedding. In this paper, we proposed a hybrid joint embedding approach that combines between distance preserving which based on symmetric distance and order preserving that based on asymmetric distance to improve image-text matching. In addition to that we propose an intra loss function to enrich the embedding with intra-modality information. We evaluate our embedding approach on the baseline model on Flickr30K dataset. The proposed loss shows a significant enhancement in matching task.
图像-文本(标题)匹配已经成为视觉与语言相结合的联合嵌入模型的常规评价方法。该任务包括基于文本查询(图像检索)对一种模式(图像)的数据进行排序,以及根据图像查询(文本检索)的相关性对文本进行排序。目前的联合嵌入方法采用对称相似度量,由于没有考虑顺序嵌入。此外,在图像-文本匹配中,所使用的损失忽略了某一情态的内部相似性,而是探索同一情态下候选词之间的关系。尽管如此,内部信息在嵌入中起着重要的作用。本文提出了一种将基于对称距离的距离保持和基于非对称距离的顺序保持相结合的混合联合嵌入方法,以改善图像-文本的匹配。此外,我们提出了一个内损失函数来丰富嵌入的模态内信息。我们在Flickr30K数据集的基线模型上评估了我们的嵌入方法。所提出的损失算法对匹配任务有显著的提高。
{"title":"Hybrid Joint Embedding with Intra-Modality Loss for Image-Text Matching","authors":"Doaa B. Ebaid, A. El-Zoghabi, Magda M. Madbouly","doi":"10.1109/ISCMI56532.2022.10068471","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068471","url":null,"abstract":"Image-text(caption) matching has become a regular evaluation of joint-embedding models that combine vision and language. This task comprises ranking the data of one modality (images) based on a Text query (Image Retrieval), and ranking texts by relevance for an image query (Text Retrieval). The current joint embedding approaches use symmetric similarity measurement, due to that order embedding is not taken in consideration. In addition to that, in image-text matching, the used losses ignore the intra similarity in a certain modality that explores the relation between the candidates in the same modality. In spite of, the important role of intra information in the embedding. In this paper, we proposed a hybrid joint embedding approach that combines between distance preserving which based on symmetric distance and order preserving that based on asymmetric distance to improve image-text matching. In addition to that we propose an intra loss function to enrich the embedding with intra-modality information. We evaluate our embedding approach on the baseline model on Flickr30K dataset. The proposed loss shows a significant enhancement in matching task.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"37 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123470191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multiple Severity-Level Classifications for IT Incident Risk Prediction 面向IT事件风险预测的多严重级别分类
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068477
Salman Ahmed, Muskaan Singh, Brendan Doherty, E. Ramlan, Kathryn Harkin, Damien Coyle
The adoption of Artificial Intelligence (AI) is now widespread in Information Technology (IT) support. A particular area of interest is in the automation of IT incident management (i.e., the handling of an unusual event that hampers the quality of IT services in the most optimized manner). In this paper, we propose a framework using state-of-art algorithms to classify and predict the severity of such incidents (commonly labeled as High, Medium, and Low severity). We argue that the proposed framework would accelerate the process of handling IT incidents with improved accuracy. The experimentation was performed on the IT Service Management (ITSM) dataset containing 500,000 real-time incident descriptions with their encoded labels (Dataset 1) from a reputable IT firm. Our results showed that the Transformer models outperformed machine learning (ML) and other deep learning (DL) models with a 98% AUC score to predict the three severity classes. We tested our framework with an open-access dataset (Dataset 2) to further validate our findings. Our framework produced a 44% improvement in AUC score compared to the existing benchmark approaches. The results show the plausibility of AI algorithms in automating the prioritization of incident processing in large IT systems.
人工智能(AI)的采用现在在信息技术(IT)支持中得到广泛应用。一个特别感兴趣的领域是IT事件管理的自动化(即,以最优化的方式处理妨碍IT服务质量的异常事件)。在本文中,我们提出了一个框架,使用最先进的算法来分类和预测此类事件的严重程度(通常标记为高、中、低严重程度)。我们认为,建议的框架将加快处理IT事件的过程,并提高准确性。实验是在IT服务管理(ITSM)数据集上进行的,该数据集包含来自一家知名IT公司的500,000个实时事件描述及其编码标签(数据集1)。我们的结果表明,Transformer模型在预测三种严重程度类别方面的AUC得分为98%,优于机器学习(ML)和其他深度学习(DL)模型。我们用一个开放获取的数据集(数据集2)测试了我们的框架,以进一步验证我们的发现。与现有的基准方法相比,我们的框架在AUC得分方面提高了44%。结果表明,人工智能算法在大型IT系统中自动化事件处理优先级方面是可行的。
{"title":"Multiple Severity-Level Classifications for IT Incident Risk Prediction","authors":"Salman Ahmed, Muskaan Singh, Brendan Doherty, E. Ramlan, Kathryn Harkin, Damien Coyle","doi":"10.1109/ISCMI56532.2022.10068477","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068477","url":null,"abstract":"The adoption of Artificial Intelligence (AI) is now widespread in Information Technology (IT) support. A particular area of interest is in the automation of IT incident management (i.e., the handling of an unusual event that hampers the quality of IT services in the most optimized manner). In this paper, we propose a framework using state-of-art algorithms to classify and predict the severity of such incidents (commonly labeled as High, Medium, and Low severity). We argue that the proposed framework would accelerate the process of handling IT incidents with improved accuracy. The experimentation was performed on the IT Service Management (ITSM) dataset containing 500,000 real-time incident descriptions with their encoded labels (Dataset 1) from a reputable IT firm. Our results showed that the Transformer models outperformed machine learning (ML) and other deep learning (DL) models with a 98% AUC score to predict the three severity classes. We tested our framework with an open-access dataset (Dataset 2) to further validate our findings. Our framework produced a 44% improvement in AUC score compared to the existing benchmark approaches. The results show the plausibility of AI algorithms in automating the prioritization of incident processing in large IT systems.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130045833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Determining the Optimal Application of Distributed Generation for Grid Integration 探讨分布式发电在电网集成中的最佳应用
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068444
K. Moloi, H. Langa
The energy supply framework has transformed over the recent time from tradition to a more flexible energy topology. This has resulted in the introduction of using distributed generation technologies to improve the sustainability of energy supply. However, the introduction of distributed generation for grid-integration has technical disadvantages, such as increase in power and voltage losses. In this paper, a technique based on discrete wavelet transform (DWT) and the genetic algorithm (GA) is proposed to determine the optimal location and size of the DGs to be connected into the grid for power loss minimisation and voltage improvement. The technique is tested using the 33 and 69 IEEE test bus system. The results obtained show that the power losses are significantly reduced with an increase in the voltage profile.
近年来,能源供应框架已经从传统的能源结构转变为更灵活的能源结构。这导致了采用分布式发电技术来提高能源供应的可持续性。然而,引入分布式发电并网有技术上的缺点,如功率和电压损失的增加。本文提出了一种基于离散小波变换(DWT)和遗传算法(GA)的技术来确定dg并网的最佳位置和大小,以实现功率损耗最小化和电压改善。采用33和69 IEEE测试总线系统对该技术进行了测试。结果表明,随着电压分布的增大,功率损耗显著降低。
{"title":"Towards Determining the Optimal Application of Distributed Generation for Grid Integration","authors":"K. Moloi, H. Langa","doi":"10.1109/ISCMI56532.2022.10068444","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068444","url":null,"abstract":"The energy supply framework has transformed over the recent time from tradition to a more flexible energy topology. This has resulted in the introduction of using distributed generation technologies to improve the sustainability of energy supply. However, the introduction of distributed generation for grid-integration has technical disadvantages, such as increase in power and voltage losses. In this paper, a technique based on discrete wavelet transform (DWT) and the genetic algorithm (GA) is proposed to determine the optimal location and size of the DGs to be connected into the grid for power loss minimisation and voltage improvement. The technique is tested using the 33 and 69 IEEE test bus system. The results obtained show that the power losses are significantly reduced with an increase in the voltage profile.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130594244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Analysis of the Generator Architectures and Loss Functions in Improving the Stability of GANs Training towards Efficient Intrusion Detection 生成器结构和损失函数在提高gan训练稳定性以实现高效入侵检测中的分析
Pub Date : 2022-11-26 DOI: 10.1109/ISCMI56532.2022.10068468
Raha Soleymanzadeh, R. Kashef
Various research studies have been recently introduced in developing generative models, especially in computer vision and image classification. These models are inspired by a generator and discriminator network architecture in a min-max optimization game called Generative Adversarial Networks (GANs). However, GANs-based models suffer from training instability, which means high oscillations during the training, which provides inaccurate results. There are various causes beyond the instability behaviours, such as the adopted generator architecture, loss function, and distance metrics. In this paper, we focus on the impact of the generator architectures and the loss functions on the GANs training. We aim to provide a comparative assessment of various architectures focusing on ensemble and hybrid models and loss functions such as Focal loss, Binary Cross-Entropy and Mean Squared loss function. Experimental results on NSL-KDD and UNSW-NB15 datasets show that the ensemble models are more stable in terms of training and have higher intrusion detection rates. Additionally, the focal loss can improve the performance of detection minority classes. Using Mean squared loss improved the detection rate for discriminator, however with the Binary Cross entropy loss function, the deep features representation is improved and there is more stability in trends for all architectures.
近年来,人们对生成模型的发展进行了大量的研究,特别是在计算机视觉和图像分类方面。这些模型的灵感来自最小-最大优化博弈生成对抗网络(GANs)中的生成器和鉴别器网络架构。然而,基于高斯的模型存在训练不稳定性,这意味着训练过程中的高振荡,从而提供不准确的结果。除不稳定行为外,还有各种原因,如采用的发电机结构、损失函数和距离度量。在本文中,我们重点研究了生成器结构和损失函数对gan训练的影响。我们的目标是提供各种架构的比较评估,重点是集成和混合模型以及损失函数,如焦损失,二元交叉熵和均方损失函数。在NSL-KDD和UNSW-NB15数据集上的实验结果表明,集成模型在训练方面更稳定,入侵检测率更高。此外,焦损可以提高检测少数类的性能。使用均方损失提高了鉴别器的检测率,而使用二元交叉熵损失函数改进了深度特征表示,并且对所有体系结构都具有更大的趋势稳定性。
{"title":"The Analysis of the Generator Architectures and Loss Functions in Improving the Stability of GANs Training towards Efficient Intrusion Detection","authors":"Raha Soleymanzadeh, R. Kashef","doi":"10.1109/ISCMI56532.2022.10068468","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068468","url":null,"abstract":"Various research studies have been recently introduced in developing generative models, especially in computer vision and image classification. These models are inspired by a generator and discriminator network architecture in a min-max optimization game called Generative Adversarial Networks (GANs). However, GANs-based models suffer from training instability, which means high oscillations during the training, which provides inaccurate results. There are various causes beyond the instability behaviours, such as the adopted generator architecture, loss function, and distance metrics. In this paper, we focus on the impact of the generator architectures and the loss functions on the GANs training. We aim to provide a comparative assessment of various architectures focusing on ensemble and hybrid models and loss functions such as Focal loss, Binary Cross-Entropy and Mean Squared loss function. Experimental results on NSL-KDD and UNSW-NB15 datasets show that the ensemble models are more stable in terms of training and have higher intrusion detection rates. Additionally, the focal loss can improve the performance of detection minority classes. Using Mean squared loss improved the detection rate for discriminator, however with the Binary Cross entropy loss function, the deep features representation is improved and there is more stability in trends for all architectures.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"162 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126807916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1