首页 > 最新文献

2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)最新文献

英文 中文
Towards Intelligent Reading through Multimodal and Contextualized Word LookUp 通过多模态和语境化查词实现智能阅读
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00203
Swetha Govindu, Raviteja Vidya Guttula, Swati Kohli, Poonam Patil, Anagha Kulkarni, Ilmi Yoon
This paper presents Koob, an eBook Reader app that coalesces three key ideas to enhance students’ language learning, specifically for ambiguous words. The first idea is to improve the effectiveness of word lookup functionality through contextualization – by incorporating word sense disambiguation (WSD) techniques to show the contextually relevant definition at the top. The second idea is to augment WSD results with crowd-sourcing solutions. The last idea seeks to reinforce students’ learning by augmenting textual information with a visual aid, pictures related to the word, as part of the word lookup functionality. An empirical evaluation demonstrates that existing WSD techniques can successfully employed to dynamically reorder definitions such that the most relevant definition is at the top of the list for more than 80% of the instances.
本文介绍了Koob,一个电子书阅读器应用程序,它结合了三个关键思想来提高学生的语言学习,特别是对于歧义词。第一个想法是通过上下文化来提高单词查找功能的有效性——通过结合词义消歧(WSD)技术在顶部显示与上下文相关的定义。第二个想法是通过众包解决方案来增强水务署的成果。最后一个想法是通过视觉辅助来增加文本信息,与单词相关的图片,作为单词查找功能的一部分,来加强学生的学习。经验评估表明,现有的WSD技术可以成功地用于动态地重新排序定义,以便在超过80%的实例中,最相关的定义位于列表的顶部。
{"title":"Towards Intelligent Reading through Multimodal and Contextualized Word LookUp","authors":"Swetha Govindu, Raviteja Vidya Guttula, Swati Kohli, Poonam Patil, Anagha Kulkarni, Ilmi Yoon","doi":"10.1109/ICMLA52953.2021.00203","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00203","url":null,"abstract":"This paper presents Koob, an eBook Reader app that coalesces three key ideas to enhance students’ language learning, specifically for ambiguous words. The first idea is to improve the effectiveness of word lookup functionality through contextualization – by incorporating word sense disambiguation (WSD) techniques to show the contextually relevant definition at the top. The second idea is to augment WSD results with crowd-sourcing solutions. The last idea seeks to reinforce students’ learning by augmenting textual information with a visual aid, pictures related to the word, as part of the word lookup functionality. An empirical evaluation demonstrates that existing WSD techniques can successfully employed to dynamically reorder definitions such that the most relevant definition is at the top of the list for more than 80% of the instances.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"10 1","pages":"1249-1252"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88640636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Improve Learner-based Recommender System with Learner’s Mood in Online Learning Platform 基于学习者情绪的在线学习平台学习者推荐系统的改进
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00271
Qing Tang, Marie-Hélène Abel, E. Negre
Learning with huge amount of online educational resources is challenging, especially when variety resources come from different online systems. Recommender systems are used to help learners obtain appropriate resources efficiently in online learning. To improve the performance of recommender system, more and more learner’s attributes (e.g. learning style, learning ability, knowledge level, etc.) have been considered. We are committed to proposing a learner-based recommender system, not just consider learner’s physical features, but also learner’s mood while learning. This recommender system can make recommendations according to the links between learners, and can change the recommendation strategy as learner’s mood changes, which will have a certain improvement in recommendation accuracy and makes recommended results more reasonable and interpretable.
利用大量的在线教育资源进行学习是一项挑战,尤其是当各种资源来自不同的在线系统时。在线学习中使用推荐系统来帮助学习者有效地获取适当的资源。为了提高推荐系统的性能,越来越多地考虑了学习者的属性(如学习风格、学习能力、知识水平等)。我们致力于提出一个基于学习者的推荐系统,不仅考虑学习者的身体特征,还考虑学习者在学习时的情绪。该推荐系统可以根据学习者之间的联系进行推荐,并且可以随着学习者情绪的变化而改变推荐策略,这样会在推荐的准确性上有一定的提高,使推荐结果更加合理和可解释性。
{"title":"Improve Learner-based Recommender System with Learner’s Mood in Online Learning Platform","authors":"Qing Tang, Marie-Hélène Abel, E. Negre","doi":"10.1109/ICMLA52953.2021.00271","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00271","url":null,"abstract":"Learning with huge amount of online educational resources is challenging, especially when variety resources come from different online systems. Recommender systems are used to help learners obtain appropriate resources efficiently in online learning. To improve the performance of recommender system, more and more learner’s attributes (e.g. learning style, learning ability, knowledge level, etc.) have been considered. We are committed to proposing a learner-based recommender system, not just consider learner’s physical features, but also learner’s mood while learning. This recommender system can make recommendations according to the links between learners, and can change the recommendation strategy as learner’s mood changes, which will have a certain improvement in recommendation accuracy and makes recommended results more reasonable and interpretable.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"46 1","pages":"1704-1709"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85686811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-Driven State of Charge Estimation of Li-ion Batteries using Supervised Machine Learning Methods 基于监督机器学习方法的锂离子电池充电状态估计
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00144
Yichun Li, Mina Maleki, Shadi Banitaan, Ming-Jie Chen
Recently, electrical vehicles (EVs) have attracted considerable attention from researchers due to the transition of the transportation industry and the increasing demand in the clean energy domain. State of charge (SOC) of Li-ion batteries has a significant role in improving the efficiency, performance, and reliability of EVs. Estimating the SOC of the Li-ion battery cannot be done directly from inner measurements due to the complex and dynamic nature of these kinds of batteries. Several data-driven approaches have recently been used to estimate the SOC of Li-ion batteries, benefiting from the availability of battery data and hardware computing capacity. However, selecting the discriminative features and best supervised machine learning (ML) models for accurate battery states estimation is still challenging. Thus, this paper investigates the effect of different ML models and extracted input features of Li-ion batteries, including Electrochemical Impedance Spectroscopy (EIS) and multi-channel feature set on the SOC prediction. The results on the public Panasonic dataset indicate that using EIS feature set as an input to the deep neural network (DNN) model is more efficient than the multi-channel feature set. Moreover, the DNN model outperforms the Gaussian process regression (GPR) model in terms of the mean squared error, mean absolute error, and root mean squared error rates for the SOC prediction.
近年来,由于交通运输行业的转型和清洁能源领域需求的增加,电动汽车引起了研究人员的广泛关注。锂离子电池的荷电状态(SOC)对提高电动汽车的效率、性能和可靠性具有重要作用。由于锂离子电池的复杂性和动态性,不能直接从内部测量来估计锂离子电池的SOC。得益于电池数据的可用性和硬件计算能力,最近已经使用了几种数据驱动的方法来估计锂离子电池的SOC。然而,选择判别特征和最佳监督机器学习(ML)模型来准确估计电池状态仍然是一个挑战。因此,本文研究了不同的ML模型和提取的锂离子电池输入特征,包括电化学阻抗谱(EIS)和多通道特征集对电池荷电状态预测的影响。在松下公共数据集上的结果表明,使用EIS特征集作为深度神经网络(DNN)模型的输入比多通道特征集更有效。此外,DNN模型在SOC预测的均方误差、平均绝对误差和均方根错误率方面优于高斯过程回归(GPR)模型。
{"title":"Data-Driven State of Charge Estimation of Li-ion Batteries using Supervised Machine Learning Methods","authors":"Yichun Li, Mina Maleki, Shadi Banitaan, Ming-Jie Chen","doi":"10.1109/ICMLA52953.2021.00144","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00144","url":null,"abstract":"Recently, electrical vehicles (EVs) have attracted considerable attention from researchers due to the transition of the transportation industry and the increasing demand in the clean energy domain. State of charge (SOC) of Li-ion batteries has a significant role in improving the efficiency, performance, and reliability of EVs. Estimating the SOC of the Li-ion battery cannot be done directly from inner measurements due to the complex and dynamic nature of these kinds of batteries. Several data-driven approaches have recently been used to estimate the SOC of Li-ion batteries, benefiting from the availability of battery data and hardware computing capacity. However, selecting the discriminative features and best supervised machine learning (ML) models for accurate battery states estimation is still challenging. Thus, this paper investigates the effect of different ML models and extracted input features of Li-ion batteries, including Electrochemical Impedance Spectroscopy (EIS) and multi-channel feature set on the SOC prediction. The results on the public Panasonic dataset indicate that using EIS feature set as an input to the deep neural network (DNN) model is more efficient than the multi-channel feature set. Moreover, the DNN model outperforms the Gaussian process regression (GPR) model in terms of the mean squared error, mean absolute error, and root mean squared error rates for the SOC prediction.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"1 1","pages":"873-878"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86230746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Purrai: A Deep Neural Network based Approach to Interpret Domestic Cat Language Purrai:一种基于深度神经网络的方法来解释家猫的语言
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00104
Weilin Sun, V. Lu, Aaron Truong, Hermione Bossolina, Yuan Lu
Being able to understand and communicate with domestic cats has always been fascinating to humans, although it is considered a difficult task even for phonetics experts. In this paper, we present our approach to this problem: Purrai, a neural-network-based machine learning platform to interpret cat’s language. Our framework consists of two parts. First, we build a comprehensively constructed cat voice dataset that is 3.7x larger than any existing public available dataset [1]. To improve accuracy, we also use several techniques to ensure labeling quality, including rule-based labeling, cross validation, cosine distance, and outlier detection, etc. Second, we design a two-stage neural network structure to interpret what cats express in the context of multiple sounds called sentences. The first stage is a modification of Google’s Vggish architecture [2] [3], which is a Convolutional Neural Network (CNN) architecture that focuses on the classification of nine primary cat sounds. The second stage takes the probability outputs of a sequence of sound classifications from the first stage and determines the emotional meaning of a cat sentence. Our first stage architecture generates a top-l and top-2 accuracy of 74.1% and 92.1%, better than that of the state-of-the-art approach: 64.9% and 83.4% [4]. Our sentence-based AI model achieves an accuracy of 81.1% for emotion prediction.
人类一直对能够听懂家猫的声音并与之交流很感兴趣,尽管即使对语音专家来说,这也被认为是一项艰巨的任务。在本文中,我们提出了解决这个问题的方法:Purrai,一个基于神经网络的机器学习平台,用于解释猫的语言。我们的框架由两部分组成。首先,我们构建了一个全面构建的猫声数据集,该数据集比任何现有的公共可用数据集大3.7倍[1]。为了提高准确性,我们还使用了几种技术来确保标注质量,包括基于规则的标注、交叉验证、余弦距离和离群值检测等。其次,我们设计了一个两阶段的神经网络结构来解释猫在多个声音(称为句子)的背景下表达的内容。第一阶段是对Google的Vggish架构的修改[2][3],这是一个卷积神经网络(CNN)架构,专注于九种主要猫音的分类。第二阶段从第一阶段获得一系列声音分类的概率输出,并确定猫句的情感意义。我们的第一阶段架构产生的top- 1和top-2准确率分别为74.1%和92.1%,优于最先进的方法:64.9%和83.4%[4]。我们基于句子的人工智能模型在情绪预测方面达到了81.1%的准确率。
{"title":"Purrai: A Deep Neural Network based Approach to Interpret Domestic Cat Language","authors":"Weilin Sun, V. Lu, Aaron Truong, Hermione Bossolina, Yuan Lu","doi":"10.1109/ICMLA52953.2021.00104","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00104","url":null,"abstract":"Being able to understand and communicate with domestic cats has always been fascinating to humans, although it is considered a difficult task even for phonetics experts. In this paper, we present our approach to this problem: Purrai, a neural-network-based machine learning platform to interpret cat’s language. Our framework consists of two parts. First, we build a comprehensively constructed cat voice dataset that is 3.7x larger than any existing public available dataset [1]. To improve accuracy, we also use several techniques to ensure labeling quality, including rule-based labeling, cross validation, cosine distance, and outlier detection, etc. Second, we design a two-stage neural network structure to interpret what cats express in the context of multiple sounds called sentences. The first stage is a modification of Google’s Vggish architecture [2] [3], which is a Convolutional Neural Network (CNN) architecture that focuses on the classification of nine primary cat sounds. The second stage takes the probability outputs of a sequence of sound classifications from the first stage and determines the emotional meaning of a cat sentence. Our first stage architecture generates a top-l and top-2 accuracy of 74.1% and 92.1%, better than that of the state-of-the-art approach: 64.9% and 83.4% [4]. Our sentence-based AI model achieves an accuracy of 81.1% for emotion prediction.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"11 1","pages":"622-627"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82592225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tiny Generative Image Compression for Bandwidth-Constrained Sensor Applications 用于带宽受限传感器应用的微小生成图像压缩
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00094
Nikolai Körber, A. Siebert, S. Hauke, Daniel Mueller-Gritschneder
Deep image compression algorithms based on Generative Adversarial Networks (GANs) are a promising direction to address the strict communication bandwidth limitations commonly encountered in IoT sensor networks (e.g. Low Power Wide Area Networks). However, current methods do not consider that the sensor nodes, which perform the image encoding, usually only offer very limited computation and memory capabilities, e.g. a resource-constrained tiny device such as a micro-controller. In this paper, we propose the first tiny generative image compression method specifically designed for image compression on micro-controllers. We base our encoder on the well-known MobileNetV2 network architecture, while keeping the decoder side fixed. To cope with the resulting asymmetric design of the compression pipeline, we investigate the impact of different training strategies (end-to-end, knowledge distillation) and integer quantization techniques (post-training, quantization-aware training) on the GAN-training stability. On the Cityscapes dataset, we achieve a compression performance that is very close to the state-of-the-art, while requiring 99% less SRAM size, 97% smaller flash storage and 87% less multiply-add operations. Our findings suggest that tiny generative image compression is particularly well suited for application-specific domains.
基于生成对抗网络(gan)的深度图像压缩算法是解决物联网传感器网络(例如低功耗广域网)中常见的严格通信带宽限制的一个有前途的方向。然而,目前的方法没有考虑到执行图像编码的传感器节点通常只提供非常有限的计算和存储能力,例如资源受限的微型设备,如微控制器。在本文中,我们提出了第一种专门为微控制器上的图像压缩而设计的微型生成图像压缩方法。我们的编码器基于众所周知的MobileNetV2网络架构,同时保持解码器侧固定。为了应对压缩管道的不对称设计,我们研究了不同的训练策略(端到端、知识蒸馏)和整数量化技术(后训练、量化感知训练)对gan训练稳定性的影响。在cityscape数据集上,我们实现了非常接近最先进的压缩性能,同时所需的SRAM大小减少了99%,闪存减少了97%,乘法添加操作减少了87%。我们的研究结果表明,微小的生成图像压缩特别适合于特定应用领域。
{"title":"Tiny Generative Image Compression for Bandwidth-Constrained Sensor Applications","authors":"Nikolai Körber, A. Siebert, S. Hauke, Daniel Mueller-Gritschneder","doi":"10.1109/ICMLA52953.2021.00094","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00094","url":null,"abstract":"Deep image compression algorithms based on Generative Adversarial Networks (GANs) are a promising direction to address the strict communication bandwidth limitations commonly encountered in IoT sensor networks (e.g. Low Power Wide Area Networks). However, current methods do not consider that the sensor nodes, which perform the image encoding, usually only offer very limited computation and memory capabilities, e.g. a resource-constrained tiny device such as a micro-controller. In this paper, we propose the first tiny generative image compression method specifically designed for image compression on micro-controllers. We base our encoder on the well-known MobileNetV2 network architecture, while keeping the decoder side fixed. To cope with the resulting asymmetric design of the compression pipeline, we investigate the impact of different training strategies (end-to-end, knowledge distillation) and integer quantization techniques (post-training, quantization-aware training) on the GAN-training stability. On the Cityscapes dataset, we achieve a compression performance that is very close to the state-of-the-art, while requiring 99% less SRAM size, 97% smaller flash storage and 87% less multiply-add operations. Our findings suggest that tiny generative image compression is particularly well suited for application-specific domains.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"84 1","pages":"564-569"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89856831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Self-Attention Mechanism in GANs for Molecule Generation gan分子生成的自注意机制
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00017
S. Chinnareddy, Pranav Grandhi, Apurva Narayan
In discrete sequence based Generative Adversarial Networks (GANs), it is important to both land the samples in the initial distribution and drive the generation towards desirable properties. However, in the case of longer molecules, the existing models seem to under-perform in producing new molecules. In this work, we propose the use of Self-Attention mechanism for Generative Adversarial Networks to allow long range dependencies. Self-Attention mechanism has produced improved rewards in novelty and promising results in generating molecules.
在基于离散序列的生成对抗网络(GANs)中,重要的是将样本置于初始分布中并使生成朝着理想的性质发展。然而,在长分子的情况下,现有的模型似乎在产生新分子方面表现不佳。在这项工作中,我们建议在生成对抗网络中使用自注意机制来允许长距离依赖。自注意机制在新颖性奖励和分子生成方面取得了可喜的成果。
{"title":"Self-Attention Mechanism in GANs for Molecule Generation","authors":"S. Chinnareddy, Pranav Grandhi, Apurva Narayan","doi":"10.1109/ICMLA52953.2021.00017","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00017","url":null,"abstract":"In discrete sequence based Generative Adversarial Networks (GANs), it is important to both land the samples in the initial distribution and drive the generation towards desirable properties. However, in the case of longer molecules, the existing models seem to under-perform in producing new molecules. In this work, we propose the use of Self-Attention mechanism for Generative Adversarial Networks to allow long range dependencies. Self-Attention mechanism has produced improved rewards in novelty and promising results in generating molecules.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"64 1","pages":"57-60"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83747506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Con Connections: Detecting Fraud from Abstracts using Topological Data Analysis 连接:利用拓扑数据分析从摘要中检测欺诈
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00069
Sarah Tymochko, Julien Chaput, T. Doster, Emilie Purvine, Jackson Warley, T. Emerson
In this paper we present a novel approach for identifying fraudulent papers from their titles and abstracts. The premise of the approach is that there are holes in the presentation of the approach and findings of fraudulent research papers. As an abstract is intended to highlight key features of the approach as well as important conclusions the authors seek to determine if the assumed existence of holes can be identified from analysis of abstracts alone. The data set considered is derived from papers sharing a single author with labels determined based on a formal linguistic analysis of the complete documents. To detect these logical and literary holes we utilize techniques from topological data analysis which summarizes data based on the presence of multi-dimensional, topological holes. We find that, in fact, topological features derived through a combination of techniques in natural language processing and time-series analysis allow for superior detection of the fraudulent papers than the natural language processing tools alone. Thus we conclude that the connections and holes present in the abstracts of research cons contributes to an ability to infer the scientific validity of the corresponding work.
在本文中,我们提出了一种从标题和摘要中识别欺诈性论文的新方法。该方法的前提是该方法的呈现和欺诈性研究论文的发现存在漏洞。由于摘要旨在突出该方法的关键特征以及重要结论,因此作者试图确定是否可以仅从摘要分析中识别假设存在的漏洞。所考虑的数据集来源于共享单个作者的论文,其标签是根据对完整文档的正式语言分析确定的。为了检测这些逻辑和文字漏洞,我们利用拓扑数据分析技术,该技术基于多维拓扑漏洞的存在来总结数据。我们发现,事实上,通过自然语言处理和时间序列分析技术相结合得出的拓扑特征比单独使用自然语言处理工具更能检测出欺诈性论文。因此,我们得出结论,研究摘要中存在的联系和漏洞有助于推断相应工作的科学有效性。
{"title":"Con Connections: Detecting Fraud from Abstracts using Topological Data Analysis","authors":"Sarah Tymochko, Julien Chaput, T. Doster, Emilie Purvine, Jackson Warley, T. Emerson","doi":"10.1109/ICMLA52953.2021.00069","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00069","url":null,"abstract":"In this paper we present a novel approach for identifying fraudulent papers from their titles and abstracts. The premise of the approach is that there are holes in the presentation of the approach and findings of fraudulent research papers. As an abstract is intended to highlight key features of the approach as well as important conclusions the authors seek to determine if the assumed existence of holes can be identified from analysis of abstracts alone. The data set considered is derived from papers sharing a single author with labels determined based on a formal linguistic analysis of the complete documents. To detect these logical and literary holes we utilize techniques from topological data analysis which summarizes data based on the presence of multi-dimensional, topological holes. We find that, in fact, topological features derived through a combination of techniques in natural language processing and time-series analysis allow for superior detection of the fraudulent papers than the natural language processing tools alone. Thus we conclude that the connections and holes present in the abstracts of research cons contributes to an ability to infer the scientific validity of the corresponding work.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"92 1","pages":"403-408"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74474223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Guided-Generative Network for noise detection in Monte-Carlo rendering 蒙特卡罗渲染中噪声检测的导向生成网络
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00018
Jérôme Buisine, F. Teytaud, S. Delepoulle, C. Renaud
Estimating the features to be extracted from an image for classification tasks are sometimes difficult, especially if images are related to a particular kind of noise. The aim of this paper is to propose a neural network architecture named Guided-Generative Network (GGN) to extract refined information that allows to correctly quantify the noise present in a sliding window of images. GNN tends to find the desired features to address such a problem in order to emit a detection criterion of this noise. The proposed GGN is applied on photorealistic images which are rendered by Monte-Carlo methods by evaluating a large number of samples per pixel. An insufficient number of samples per pixel tends to result in residual noise which is very noticeable to humans. This noise can be reduced by increasing the number of samples, as proven by Monte-Carlo theory, but this involves considerable computational time. Finding the right number of samples needed for human observers to perceive no noise is still an open problem. The results obtained show that GGN can correctly solve the problem without prior knowledge of the noise while being competitive with existing methods.
估计从图像中提取的特征用于分类任务有时是困难的,特别是当图像与特定类型的噪声相关时。本文的目的是提出一种名为引导生成网络(GGN)的神经网络架构,以提取精细信息,从而正确量化图像滑动窗口中存在的噪声。GNN倾向于找到所需的特征来解决这样的问题,以便发出该噪声的检测准则。通过对每像素的大量样本进行评估,将所提出的GGN应用于蒙特卡罗方法渲染的逼真图像。每个像素的样本数量不足往往会导致对人类来说非常明显的残余噪声。正如蒙特卡罗理论所证明的那样,可以通过增加样本数量来减少这种噪声,但这需要大量的计算时间。如何找到正确数量的样本,使人类观察者感觉不到噪音,仍然是一个悬而未决的问题。结果表明,该算法在不需要先验知识的情况下能够正确地解决问题,并与现有方法相竞争。
{"title":"Guided-Generative Network for noise detection in Monte-Carlo rendering","authors":"Jérôme Buisine, F. Teytaud, S. Delepoulle, C. Renaud","doi":"10.1109/ICMLA52953.2021.00018","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00018","url":null,"abstract":"Estimating the features to be extracted from an image for classification tasks are sometimes difficult, especially if images are related to a particular kind of noise. The aim of this paper is to propose a neural network architecture named Guided-Generative Network (GGN) to extract refined information that allows to correctly quantify the noise present in a sliding window of images. GNN tends to find the desired features to address such a problem in order to emit a detection criterion of this noise. The proposed GGN is applied on photorealistic images which are rendered by Monte-Carlo methods by evaluating a large number of samples per pixel. An insufficient number of samples per pixel tends to result in residual noise which is very noticeable to humans. This noise can be reduced by increasing the number of samples, as proven by Monte-Carlo theory, but this involves considerable computational time. Finding the right number of samples needed for human observers to perceive no noise is still an open problem. The results obtained show that GGN can correctly solve the problem without prior knowledge of the noise while being competitive with existing methods.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"17 1","pages":"61-66"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79267064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perceptually Constrained Fast Adversarial Audio Attacks 感知约束快速对抗性音频攻击
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00135
Jason Henry, Mehmet Ergezer, M. Orescanin
Audio adversarial attacks on deep learning models are of great interest given the commercial success and proliferation of these technologies. These types of attacks have been successfully demonstrated, however, artifacts introduced in the adversarial audio are easily detectable by a human observer. In this work, an expansion of the fast audio adversarial perturbation framework is proposed that can produce an adversarial attack that is imperceptible to a human observer in near-real time using black-box attacks. This is achieved by proposing a perceptually motivated penalty function. We propose a perceptual fast audio adversarial perturbation generator (PFAPG) that employs a loudness constrained loss function, in lieu of a conventional L-2 norm, between the adversarial example and original audio signal. We compare the performance of PFAPG against the conventional constraint based on the MSE on three audio recognition datasets: speaker recognition, speech command, and the Ryerson audiovisual database of emotional speech and song. Our results indicate that, on average, PFAPG equipped with the loudness-constrained loss function yields a 11% higher success rate, while reducing the undesirable distortion artifacts in adversarial audio by 10% dB compared to the prevalent MSE constraints.
鉴于这些技术的商业成功和扩散,对深度学习模型的音频对抗性攻击引起了人们的极大兴趣。这些类型的攻击已经被成功证明,然而,在对抗音频中引入的伪影很容易被人类观察者检测到。在这项工作中,提出了快速音频对抗性扰动框架的扩展,该框架可以使用黑盒攻击在近实时的情况下产生对人类观察者难以察觉的对抗性攻击。这是通过提出感知动机惩罚函数来实现的。我们提出了一种感知快速音频对抗性扰动发生器(PFAPG),它在对抗性示例和原始音频信号之间使用响度约束损失函数代替传统的L-2范数。在说话人识别、语音指令和Ryerson情感语音和歌曲的视听数据库三个音频识别数据集上,比较了PFAPG与基于MSE的传统约束的性能。我们的研究结果表明,平均而言,配备了响度约束损失函数的PFAPG的成功率提高了11%,同时与流行的MSE约束相比,对抗性音频中的不良失真伪像减少了10% dB。
{"title":"Perceptually Constrained Fast Adversarial Audio Attacks","authors":"Jason Henry, Mehmet Ergezer, M. Orescanin","doi":"10.1109/ICMLA52953.2021.00135","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00135","url":null,"abstract":"Audio adversarial attacks on deep learning models are of great interest given the commercial success and proliferation of these technologies. These types of attacks have been successfully demonstrated, however, artifacts introduced in the adversarial audio are easily detectable by a human observer. In this work, an expansion of the fast audio adversarial perturbation framework is proposed that can produce an adversarial attack that is imperceptible to a human observer in near-real time using black-box attacks. This is achieved by proposing a perceptually motivated penalty function. We propose a perceptual fast audio adversarial perturbation generator (PFAPG) that employs a loudness constrained loss function, in lieu of a conventional L-2 norm, between the adversarial example and original audio signal. We compare the performance of PFAPG against the conventional constraint based on the MSE on three audio recognition datasets: speaker recognition, speech command, and the Ryerson audiovisual database of emotional speech and song. Our results indicate that, on average, PFAPG equipped with the loudness-constrained loss function yields a 11% higher success rate, while reducing the undesirable distortion artifacts in adversarial audio by 10% dB compared to the prevalent MSE constraints.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"3 1","pages":"819-824"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75289722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Argue to Learn: Accelerated Argumentation-Based Learning
Pub Date : 2021-12-01 DOI: 10.1109/ICMLA52953.2021.00183
H. Ayoobi, M. Cao, R. Verbrugge, B. Verheij
Human agents can acquire knowledge and learn through argumentation. Inspired by this fact, we propose a novel argumentation-based machine learning technique that can be used for online incremental learning scenarios. Existing methods for online incremental learning problems typically do not generalize well from just a few learning instances. Our previous argumentation-based online incremental learning method outperformed state-of-the-art methods in terms of accuracy and learning speed. However, it was neither memory-efficient nor computationally efficient since the algorithm used the power set of the feature values for updating the model. In this paper, we propose an accelerated version of the algorithm, with polynomial instead of exponential complexity, while achieving higher learning accuracy. The proposed method is at least $200times$ faster than the original argumentation-based learning method and is more memory-efficient.
人类代理可以通过论证来获取知识和学习。受这一事实的启发,我们提出了一种新的基于论证的机器学习技术,可用于在线增量学习场景。现有的在线增量学习问题的方法通常不能很好地从几个学习实例中泛化。我们之前基于论证的在线增量学习方法在准确性和学习速度方面优于最先进的方法。然而,由于该算法使用特征值的幂集来更新模型,因此既不节省内存也不节省计算效率。在本文中,我们提出了该算法的加速版本,使用多项式而不是指数复杂度,同时实现了更高的学习精度。提出的方法比原始的基于论证的学习方法至少快200倍,并且更节省内存。
{"title":"Argue to Learn: Accelerated Argumentation-Based Learning","authors":"H. Ayoobi, M. Cao, R. Verbrugge, B. Verheij","doi":"10.1109/ICMLA52953.2021.00183","DOIUrl":"https://doi.org/10.1109/ICMLA52953.2021.00183","url":null,"abstract":"Human agents can acquire knowledge and learn through argumentation. Inspired by this fact, we propose a novel argumentation-based machine learning technique that can be used for online incremental learning scenarios. Existing methods for online incremental learning problems typically do not generalize well from just a few learning instances. Our previous argumentation-based online incremental learning method outperformed state-of-the-art methods in terms of accuracy and learning speed. However, it was neither memory-efficient nor computationally efficient since the algorithm used the power set of the feature values for updating the model. In this paper, we propose an accelerated version of the algorithm, with polynomial instead of exponential complexity, while achieving higher learning accuracy. The proposed method is at least $200times$ faster than the original argumentation-based learning method and is more memory-efficient.","PeriodicalId":6750,"journal":{"name":"2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"54 1","pages":"1118-1123"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80865779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1