首页 > 最新文献

Machine Learning and Applications: An International Journal最新文献

英文 中文
A Machine Learning Method for Prediction of Yogurt Quality and Consumers Preferencesusing Sensory Attributes and Image Processing Techniques 利用感官属性和图像处理技术预测酸奶质量和消费者偏好的机器学习方法
Pub Date : 2023-03-30 DOI: 10.5121/mlaij.2023.10101
Maha Hany, Shaheera Rashwan, Neveen M. Abdelmotilib
Prediction of quality and consumers’ preferences is essential task for food producers to improve their market share and reduce any gap in food safety standards. In this paper, we develop a machine learning method to predict yogurt preferences based on the sensory attributes and analysis of samples’ images using image processing texture and color feature extraction techniques. We compare three unsupervised ML feature selection techniques (Principal Component Analysis and Independent Component Analysis and t-distributed Stochastic Neighbour Embedding) with one supervised ML feature selection technique (Linear Discriminant Analysis) in terms of accuracy of classification. Results show the efficiency of the supervised ML feature selection technique over the traditional feature selection techniques.
预测质量和消费者的偏好是食品生产商提高市场份额和缩小食品安全标准差距的基本任务。在本文中,我们开发了一种机器学习方法来预测酸奶的偏好,该方法基于感官属性和使用图像处理纹理和颜色特征提取技术对样本图像进行分析。我们比较了三种无监督机器学习特征选择技术(主成分分析、独立成分分析和t分布随机邻居嵌入)和一种有监督机器学习特征选择技术(线性判别分析)在分类精度方面的差异。结果表明,有监督的机器学习特征选择技术比传统的特征选择技术更有效。
{"title":"A Machine Learning Method for Prediction of Yogurt Quality and Consumers Preferencesusing Sensory Attributes and Image Processing Techniques","authors":"Maha Hany, Shaheera Rashwan, Neveen M. Abdelmotilib","doi":"10.5121/mlaij.2023.10101","DOIUrl":"https://doi.org/10.5121/mlaij.2023.10101","url":null,"abstract":"Prediction of quality and consumers’ preferences is essential task for food producers to improve their market share and reduce any gap in food safety standards. In this paper, we develop a machine learning method to predict yogurt preferences based on the sensory attributes and analysis of samples’ images using image processing texture and color feature extraction techniques. We compare three unsupervised ML feature selection techniques (Principal Component Analysis and Independent Component Analysis and t-distributed Stochastic Neighbour Embedding) with one supervised ML feature selection technique (Linear Discriminant Analysis) in terms of accuracy of classification. Results show the efficiency of the supervised ML feature selection technique over the traditional feature selection techniques.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127300812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Spectral Classification of Stars using Machine Learning: An Approach based on the use of Unbalanced Data 基于机器学习的恒星光谱自动分类:一种基于非平衡数据的方法
Pub Date : 2022-12-30 DOI: 10.5121/mlaij.2022.9401
Marco Oyarzo Huichaqueo, Renato Andrés Muñoz Orrego
With the increase in astronomical surveys, astronomers are faced with the challenging task of analyzing a large amount of data in order to classify observed objects into hard-to-distinguish classes. This article presents a machine learning-based method for the automatic spectral classification of stars from the latest release of the SDSS database. We propose the combinatorial use of spectral data, derived stellar data, and calculated data to create patterns. Using these patterns as inputs, we develop a Random Forest model that outputs the spectral class of the observed star. Our model is able to classify data into six complex classes: A, F, G, K, M, and Carbon stars. Due to the unbalanced nature of the data, we train our model considering three data use cases: using the original data, using under-sampling, and over-sampling data techniques. We further test our model by using a fixed dataset and a stratified dataset. From this, we analyze the performance of our model through statistical metrics. The experimental results showed that the combinatorial use of data as an input pattern contributes to improve the prediction scores in all data use cases, meanwhile, the model trained with augmented data outperforms the other cases. Our results suggest that machine learning-based spectral classification of stars may be useful for astronomers.
随着天文调查的增加,天文学家面临着分析大量数据以将观测到的天体划分为难以区分的类别的艰巨任务。本文提出了一种基于机器学习的SDSS数据库中恒星光谱自动分类方法。我们建议组合使用光谱数据、导出的恒星数据和计算数据来创建模式。使用这些模式作为输入,我们开发了一个随机森林模型,输出被观测恒星的光谱类别。我们的模型能够将数据分为六个复杂的类别:A、F、G、K、M和碳星。由于数据的不平衡性质,我们考虑了三种数据用例来训练我们的模型:使用原始数据,使用欠采样和过度采样数据技术。我们通过使用固定数据集和分层数据集进一步测试我们的模型。在此基础上,我们通过统计度量来分析模型的性能。实验结果表明,组合使用数据作为输入模式有助于提高所有数据用例的预测分数,同时,用增强数据训练的模型优于其他用例。我们的研究结果表明,基于机器学习的恒星光谱分类可能对天文学家有用。
{"title":"Automatic Spectral Classification of Stars using Machine Learning: An Approach based on the use of Unbalanced Data","authors":"Marco Oyarzo Huichaqueo, Renato Andrés Muñoz Orrego","doi":"10.5121/mlaij.2022.9401","DOIUrl":"https://doi.org/10.5121/mlaij.2022.9401","url":null,"abstract":"With the increase in astronomical surveys, astronomers are faced with the challenging task of analyzing a large amount of data in order to classify observed objects into hard-to-distinguish classes. This article presents a machine learning-based method for the automatic spectral classification of stars from the latest release of the SDSS database. We propose the combinatorial use of spectral data, derived stellar data, and calculated data to create patterns. Using these patterns as inputs, we develop a Random Forest model that outputs the spectral class of the observed star. Our model is able to classify data into six complex classes: A, F, G, K, M, and Carbon stars. Due to the unbalanced nature of the data, we train our model considering three data use cases: using the original data, using under-sampling, and over-sampling data techniques. We further test our model by using a fixed dataset and a stratified dataset. From this, we analyze the performance of our model through statistical metrics. The experimental results showed that the combinatorial use of data as an input pattern contributes to improve the prediction scores in all data use cases, meanwhile, the model trained with augmented data outperforms the other cases. Our results suggest that machine learning-based spectral classification of stars may be useful for astronomers.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"26 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133320543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ai_Birder: Using Artificial Intelligence and Deep Learning to Create a Mobile Application that Automates Bird Classification Ai_Birder:使用人工智能和深度学习创建一个自动鸟类分类的移动应用程序
Pub Date : 2022-09-30 DOI: 10.5121/mlaij.2022.9301
Charles Tian, Yu Sun
Birds are everywhere around us and are easy to spot. However, for many beginner birders, identifying the birds is a hard task [8]. There are many apps that help the birder to identify the birds, but they are often too complicated and require good internet to give a result. A better app is needed so that birders can identify birds while not depending on internet connection. My app, AI_Bider, is mainly built in android studio using flutter and firebase, and the AI engine is coded with TensorFlow and trained with images from the internet [9]. To test my AI engine, I made six different prototypes, each having a different number of times that the code will train from the dataset of pictures. I then selected 5 birds that are in my dataset and found 5 pictures on the internet for each of them, which I then uploaded to the app. My app will then give me 3 bird species that most closely resemble the image, as well as the app’s confidence in its choices, which are listed as percentages [6]. I recorded the percentages of accuracy for each picture. After taking the average percentage of all the models, I selected the most successful model, which had an average percent of accuracy of 79%.
我们周围到处都是鸟,很容易发现。然而,对于许多观鸟初学者来说,识别鸟类是一项艰巨的任务[8]。有许多应用程序可以帮助观鸟者识别鸟类,但它们往往过于复杂,需要良好的网络才能给出结果。需要一个更好的应用程序,这样观鸟者就可以在不依赖网络连接的情况下识别鸟类。我的应用AI_Bider主要是在android studio中使用flutter和firebase构建的,AI引擎是用TensorFlow编码的,并使用来自互联网的图像进行训练[9]。为了测试我的AI引擎,我制作了6个不同的原型,每个原型都有不同的次数,代码将从图片数据集中训练。然后,我在我的数据集中选择了5只鸟,并在互联网上为每只鸟找到了5张图片,然后我将其上传到应用程序。然后,我的应用程序将为我提供与图像最相似的3种鸟类,以及应用程序对其选择的置信度,它们以百分比列出[6]。我记录了每张图片的准确率。在取了所有模型的平均百分比后,我选择了最成功的模型,其平均准确率为79%。
{"title":"Ai_Birder: Using Artificial Intelligence and Deep Learning to Create a Mobile Application that Automates Bird Classification","authors":"Charles Tian, Yu Sun","doi":"10.5121/mlaij.2022.9301","DOIUrl":"https://doi.org/10.5121/mlaij.2022.9301","url":null,"abstract":"Birds are everywhere around us and are easy to spot. However, for many beginner birders, identifying the birds is a hard task [8]. There are many apps that help the birder to identify the birds, but they are often too complicated and require good internet to give a result. A better app is needed so that birders can identify birds while not depending on internet connection. My app, AI_Bider, is mainly built in android studio using flutter and firebase, and the AI engine is coded with TensorFlow and trained with images from the internet [9]. To test my AI engine, I made six different prototypes, each having a different number of times that the code will train from the dataset of pictures. I then selected 5 birds that are in my dataset and found 5 pictures on the internet for each of them, which I then uploaded to the app. My app will then give me 3 bird species that most closely resemble the image, as well as the app’s confidence in its choices, which are listed as percentages [6]. I recorded the percentages of accuracy for each picture. After taking the average percentage of all the models, I selected the most successful model, which had an average percent of accuracy of 79%.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"105 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123086850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multilingual Speech to Text using Deep Learning based on MFCC Features 基于MFCC特征的深度学习多语言语音到文本
Pub Date : 2022-06-30 DOI: 10.5121/mlaij.2022.9202
P. Reddy
The proposed methodology presented in the paper deals with solving the problem of multilingual speech recognition. Current text and speech recognition and translation methods have a very low accuracy in translating sentences which contain a mixture of two or more different languages. The paper proposes a novel approach to tackling this problem and highlights some of the drawbacks of current recognition and translation methods. The proposed approach deals with recognition of audio queries which contain a mixture of words in two different languages - Kannada and English. The novelty in the approach presented, is the use of a next Word Prediction model in combination with a Deep Learning speech recognition model to accurately recognise and convert the input audio query to text. Another method proposed to solve the problem of multilingual speech recognition and translation is the use of cosine similarity between the audio features of words for fast and accurate recognition. The dataset used for training and testing the models was generated manually by the authors as there was no pre-existing audio and text dataset which contained sentences in a mixture of both Kannada and English. The DL speech recognition model in combination with the Word Prediction model gives an accuracy of 71% when tested on the in-house multilingual dataset. This method outperforms other existing translation and recognition solutions for the same test set. Multilingual translation and recognition is an important problem to tackle as there is a tendency for people to speak in a mixture of languages. By solving this problem, the barrier of language and communication can be lifted and thus can help people connect better and more comfortably with each other.
本文提出的方法解决了多语言语音识别问题。目前的文本和语音识别和翻译方法在翻译包含两种或两种以上不同语言的混合句子时准确率很低。本文提出了一种新的方法来解决这一问题,并指出了当前识别和翻译方法的一些缺陷。所提出的方法处理包含两种不同语言(卡纳达语和英语)混合单词的音频查询的识别。该方法的新颖之处在于将下一个单词预测模型与深度学习语音识别模型相结合,以准确识别并将输入的音频查询转换为文本。另一种解决多语言语音识别和翻译问题的方法是利用单词音频特征之间的余弦相似度进行快速准确的识别。用于训练和测试模型的数据集是由作者手动生成的,因为没有预先存在的音频和文本数据集,其中包含卡纳达语和英语混合的句子。当在内部多语言数据集上测试时,DL语音识别模型与Word预测模型相结合的准确率为71%。对于相同的测试集,该方法优于其他现有的翻译和识别解决方案。多语种翻译和识别是一个需要解决的重要问题,因为人们倾向于使用多种语言。通过解决这个问题,语言和沟通的障碍可以被解除,从而可以帮助人们更好,更舒适地相互联系。
{"title":"Multilingual Speech to Text using Deep Learning based on MFCC Features","authors":"P. Reddy","doi":"10.5121/mlaij.2022.9202","DOIUrl":"https://doi.org/10.5121/mlaij.2022.9202","url":null,"abstract":"The proposed methodology presented in the paper deals with solving the problem of multilingual speech recognition. Current text and speech recognition and translation methods have a very low accuracy in translating sentences which contain a mixture of two or more different languages. The paper proposes a novel approach to tackling this problem and highlights some of the drawbacks of current recognition and translation methods. The proposed approach deals with recognition of audio queries which contain a mixture of words in two different languages - Kannada and English. The novelty in the approach presented, is the use of a next Word Prediction model in combination with a Deep Learning speech recognition model to accurately recognise and convert the input audio query to text. Another method proposed to solve the problem of multilingual speech recognition and translation is the use of cosine similarity between the audio features of words for fast and accurate recognition. The dataset used for training and testing the models was generated manually by the authors as there was no pre-existing audio and text dataset which contained sentences in a mixture of both Kannada and English. The DL speech recognition model in combination with the Word Prediction model gives an accuracy of 71% when tested on the in-house multilingual dataset. This method outperforms other existing translation and recognition solutions for the same test set. Multilingual translation and recognition is an important problem to tackle as there is a tendency for people to speak in a mixture of languages. By solving this problem, the barrier of language and communication can be lifted and thus can help people connect better and more comfortably with each other.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127752550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DSAGLSTM-DTA: Prediction of Drug-Target Affinity using Dual Self-Attention and LSTM DSAGLSTM-DTA:利用双自注意和LSTM预测药物-靶标亲和力
Pub Date : 2022-06-30 DOI: 10.5121/mlaij.2022.9201
Lyu Zhijian, Shaohua Jiang, Yonghao Tan
The research on affinity between drugs and targets (DTA) aims to effectively narrow the target search space for drug repurposing. Therefore, reasonable prediction of drug and target affinities can minimize the waste of resources such as human and material resources. In this work, a novel graph-based model called DSAGLSTM-DTA was proposed for DTA prediction. The proposed model is unlike previous graph-based drug-target affinity model, which incorporated self-attention mechanisms in the feature extraction process of drug molecular graphs to fully extract its effective feature representations. The features of each atom in the 2D molecular graph were weighted based on attention score before being aggregated as molecule representation and two distinct pooling architectures, namely centralized and distributed architectures were implemented and compared on benchmark datasets. In addition, in the course of processing protein sequences, inspired by the approach of protein feature extraction in GDGRU-DTA, we continue to interpret protein sequences as time series and extract their features using Bidirectional Long Short-Term Memory (BiLSTM) networks, since the context-dependence of long amino acid sequences. Similarly, DSAGLSTM-DTA also utilized a self-attention mechanism in the process of protein feature extraction to obtain comprehensive representations of proteins, in which the final hidden states for element in the batch were weighted with the each unit output of LSTM, and the results were represented as the final feature of proteins. Eventually, representations of drug and protein were concatenated and fed into prediction block for final prediction. The proposed model was evaluated on different regression datasets and binary classification datasets, and the results demonstrated that DSAGLSTM-DTA was superior to some state-ofthe-art DTA models and exhibited good generalization ability.
药物与靶点亲和力(DTA)的研究旨在有效缩小药物再利用的靶点搜索空间。因此,合理预测药物和靶标亲和力,可以最大限度地减少人力、物力等资源的浪费。在这项工作中,提出了一种新的基于图的DTA预测模型DSAGLSTM-DTA。该模型不同于以往基于图的药物-靶点亲和力模型,在药物分子图的特征提取过程中引入自注意机制,充分提取其有效特征表征。将二维分子图中每个原子的特征根据注意力得分进行加权,然后聚合为分子表示,并在基准数据集上实现集中式和分布式两种不同的池化架构进行比较。此外,在处理蛋白质序列的过程中,受GDGRU-DTA中蛋白质特征提取方法的启发,我们继续将蛋白质序列解释为时间序列,并利用双向长短期记忆(BiLSTM)网络提取其特征,因为长氨基酸序列具有上下文依赖性。同样,DSAGLSTM-DTA在蛋白质特征提取过程中也利用了自注意机制来获得蛋白质的综合表征,将批中元素的最终隐藏状态与LSTM的每个单元输出进行加权,并将结果表示为蛋白质的最终特征。最后,将药物和蛋白质的表示连接到预测块中进行最终预测。在不同的回归数据集和二值分类数据集上对该模型进行了评价,结果表明,DSAGLSTM-DTA模型优于一些现有的DTA模型,具有良好的泛化能力。
{"title":"DSAGLSTM-DTA: Prediction of Drug-Target Affinity using Dual Self-Attention and LSTM","authors":"Lyu Zhijian, Shaohua Jiang, Yonghao Tan","doi":"10.5121/mlaij.2022.9201","DOIUrl":"https://doi.org/10.5121/mlaij.2022.9201","url":null,"abstract":"The research on affinity between drugs and targets (DTA) aims to effectively narrow the target search space for drug repurposing. Therefore, reasonable prediction of drug and target affinities can minimize the waste of resources such as human and material resources. In this work, a novel graph-based model called DSAGLSTM-DTA was proposed for DTA prediction. The proposed model is unlike previous graph-based drug-target affinity model, which incorporated self-attention mechanisms in the feature extraction process of drug molecular graphs to fully extract its effective feature representations. The features of each atom in the 2D molecular graph were weighted based on attention score before being aggregated as molecule representation and two distinct pooling architectures, namely centralized and distributed architectures were implemented and compared on benchmark datasets. In addition, in the course of processing protein sequences, inspired by the approach of protein feature extraction in GDGRU-DTA, we continue to interpret protein sequences as time series and extract their features using Bidirectional Long Short-Term Memory (BiLSTM) networks, since the context-dependence of long amino acid sequences. Similarly, DSAGLSTM-DTA also utilized a self-attention mechanism in the process of protein feature extraction to obtain comprehensive representations of proteins, in which the final hidden states for element in the batch were weighted with the each unit output of LSTM, and the results were represented as the final feature of proteins. Eventually, representations of drug and protein were concatenated and fed into prediction block for final prediction. The proposed model was evaluated on different regression datasets and binary classification datasets, and the results demonstrated that DSAGLSTM-DTA was superior to some state-ofthe-art DTA models and exhibited good generalization ability.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127160823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exoplanets Identification and Clustering with Machine Learning Methods 用机器学习方法识别和聚类系外行星
Pub Date : 2022-03-31 DOI: 10.5121/mlaij.2022.9101
Yucheng Jin, Lanyi Yang, Chia-En Chiang
The discovery of habitable exoplanets has long been a heated topic in astronomy. Traditional methods for exoplanet identification include the wobble method, direct imaging, gravitational microlensing, etc., which not only require a considerable investment of manpower, time, and money, but also are limited by the performance of astronomical telescopes. In this study, we proposed the idea of using machine learning methods to identify exoplanets. We used the Kepler dataset collected by NASA from the Kepler Space Observatory to conduct supervised learning, which predicts the existence of exoplanet candidates as a three-categorical classification task, using decision tree, random forest, naïve Bayes, and neural network; we used another NASA dataset consisted of the confirmed exoplanets data to conduct unsupervised learning, which divides the confirmed exoplanets into different clusters, using k-means clustering. As a result, our models achieved accuracies of 99.06%, 92.11%, 88.50%, and 99.79%, respectively, in the supervised learning task and successfully obtained reasonable clusters in the unsupervised learning task.
长期以来,发现可居住的系外行星一直是天文学的热门话题。传统的系外行星识别方法包括摆动法、直接成像、引力微透镜等,这些方法不仅需要投入大量的人力、时间和金钱,而且受天文望远镜性能的限制。在这项研究中,我们提出了使用机器学习方法来识别系外行星的想法。我们利用NASA从开普勒空间天文台收集的开普勒数据集进行监督学习,利用决策树、随机森林、naïve贝叶斯和神经网络来预测系外行星候选者的存在性,并将其作为一个三分类任务;我们使用另一个由已确认的系外行星数据组成的NASA数据集进行无监督学习,使用k-means聚类将已确认的系外行星划分为不同的集群。结果表明,我们的模型在有监督学习任务中的准确率分别为99.06%、92.11%、88.50%和99.79%,在无监督学习任务中成功获得了合理的聚类。
{"title":"Exoplanets Identification and Clustering with Machine Learning Methods","authors":"Yucheng Jin, Lanyi Yang, Chia-En Chiang","doi":"10.5121/mlaij.2022.9101","DOIUrl":"https://doi.org/10.5121/mlaij.2022.9101","url":null,"abstract":"The discovery of habitable exoplanets has long been a heated topic in astronomy. Traditional methods for exoplanet identification include the wobble method, direct imaging, gravitational microlensing, etc., which not only require a considerable investment of manpower, time, and money, but also are limited by the performance of astronomical telescopes. In this study, we proposed the idea of using machine learning methods to identify exoplanets. We used the Kepler dataset collected by NASA from the Kepler Space Observatory to conduct supervised learning, which predicts the existence of exoplanet candidates as a three-categorical classification task, using decision tree, random forest, naïve Bayes, and neural network; we used another NASA dataset consisted of the confirmed exoplanets data to conduct unsupervised learning, which divides the confirmed exoplanets into different clusters, using k-means clustering. As a result, our models achieved accuracies of 99.06%, 92.11%, 88.50%, and 99.79%, respectively, in the supervised learning task and successfully obtained reasonable clusters in the unsupervised learning task.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125861261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Survey of Neural Network Hardware Accelerators in Machine Learning 机器学习中的神经网络硬件加速器综述
Pub Date : 2021-12-31 DOI: 10.5121/mlaij.2021.8402
F. Jasem, Manar AlSaraf
The use of Machine Learning in Artificial Intelligence is the inspiration that shaped technology as it is today. Machine Learning has the power to greatly simplify our lives. Improvement in speech recognition and language understanding help the community interact more naturally with technology. The popularity of machine learning opens up the opportunities for optimizing the design of computing platforms using welldefined hardware accelerators. In the upcoming few years, cameras will be utilised as sensors for several applications. For ease of use and privacy restrictions, the requested image processing should be limited to a local embedded computer platform and with a high accuracy. Furthermore, less energy should be consumed. Dedicated acceleration of Convolutional Neural Networks can achieve these targets with high flexibility to perform multiple vision tasks. However, due to the exponential growth in technology constraints (especially in terms of energy) which could lead to heterogeneous multicores, and increasing number of defects, the strategy of defect-tolerant accelerators for heterogeneous multi-cores may become a main micro-architecture research issue. The up to date accelerators used still face some performance issues such as memory limitations, bandwidth, speed etc. This literature summarizes (in terms of a survey) recent work of accelerators including their advantages and disadvantages to make it easier for developers with neural network interests to further improve what has already been established.
机器学习在人工智能中的应用是塑造今天技术的灵感来源。机器学习有能力大大简化我们的生活。语音识别和语言理解的提高有助于社区更自然地与技术互动。机器学习的普及为使用定义良好的硬件加速器优化计算平台的设计提供了机会。在接下来的几年里,摄像头将被用作传感器用于多种应用。为了方便使用和隐私限制,所请求的图像处理应限制在本地嵌入式计算机平台上,并具有较高的精度。此外,应该减少能源消耗。卷积神经网络的专用加速可以实现这些目标,并且具有高度的灵活性来执行多个视觉任务。然而,由于技术约束的指数级增长(特别是在能量方面)可能导致异构多核,以及缺陷数量的增加,异构多核的容错加速器策略可能成为微架构研究的主要问题。使用的最新加速器仍然面临一些性能问题,如内存限制、带宽、速度等。这篇文献总结了加速器的最新工作,包括它们的优点和缺点,使对神经网络感兴趣的开发人员更容易进一步改进已经建立的东西。
{"title":"A Survey of Neural Network Hardware Accelerators in Machine Learning","authors":"F. Jasem, Manar AlSaraf","doi":"10.5121/mlaij.2021.8402","DOIUrl":"https://doi.org/10.5121/mlaij.2021.8402","url":null,"abstract":"The use of Machine Learning in Artificial Intelligence is the inspiration that shaped technology as it is today. Machine Learning has the power to greatly simplify our lives. Improvement in speech recognition and language understanding help the community interact more naturally with technology. The popularity of machine learning opens up the opportunities for optimizing the design of computing platforms using welldefined hardware accelerators. In the upcoming few years, cameras will be utilised as sensors for several applications. For ease of use and privacy restrictions, the requested image processing should be limited to a local embedded computer platform and with a high accuracy. Furthermore, less energy should be consumed. Dedicated acceleration of Convolutional Neural Networks can achieve these targets with high flexibility to perform multiple vision tasks. However, due to the exponential growth in technology constraints (especially in terms of energy) which could lead to heterogeneous multicores, and increasing number of defects, the strategy of defect-tolerant accelerators for heterogeneous multi-cores may become a main micro-architecture research issue. The up to date accelerators used still face some performance issues such as memory limitations, bandwidth, speed etc. This literature summarizes (in terms of a survey) recent work of accelerators including their advantages and disadvantages to make it easier for developers with neural network interests to further improve what has already been established.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133693090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybridization of DBN with SVM and its Impact on Performance in Multi-Document Summarization DBN与SVM的杂交及其对多文档摘要性能的影响
Pub Date : 2021-09-30 DOI: 10.5121/mlaij.2021.8304
Data available from web based sources has grown tremendously with growth of the internet. Users interested in information from such sources often use a search engine to obtain the data which they edit for presentation to their audience. This process can be tedious especially when it involves the generation of a summary. One way to ease the process is by automation of the summary generation process. Efforts by researchers towards automatic summarization have yielded several approaches among them machine learning. Thus, recommendations have been made on combining the algorithms with different strengths, also called hybridization, in order to enhance their performance. Therefore, this research sought to establish the impact of hybridization of Deep Belief Network (DBN) with Support Vector Machine (SVM) on precision, recall, accuracy and F-measure when used in the case of query oriented multi-document summarization. The experiments were carried out using data from National Institute of Standards and Technology (NIST), Document Understanding Conference (DUC) 2006. The data was split into training and test data and used appropriately in DBN, SVM, SVM-DBN hybrid and DBN-SVM hybrid. Results indicated that the hybridized algorithm has better precision, accuracy and F-measure as compared to DBN. Pre-classification hybridization of DBN with SVM (SVM-DBN) gives the best results. This research implies that use of DBN and SVM hybrid algorithms would enhance query oriented multi-document summarization.
随着互联网的发展,可以从基于网络的来源获得的数据也急剧增长。对这些来源的信息感兴趣的用户通常使用搜索引擎来获取他们编辑的数据,以便向听众展示。这个过程可能很乏味,特别是当它涉及到生成摘要时。简化该过程的一种方法是使摘要生成过程自动化。研究人员对自动摘要的努力已经产生了几种方法,其中包括机器学习。因此,建议将不同优势的算法结合起来,也称为杂交,以提高其性能。因此,本研究试图建立深度信念网络(DBN)与支持向量机(SVM)混合在面向查询的多文档摘要中对精密度、查全率、正确率和f测度的影响。实验使用的数据来自美国国家标准与技术研究所(NIST), 2006年文件理解会议(DUC)。将数据分为训练数据和测试数据,并在DBN、SVM、SVM-DBN混合和DBN-SVM混合中适当使用。结果表明,与DBN相比,混合算法具有更高的精度、精度和F-measure。DBN与支持向量机(SVM-DBN)的预分类杂交(SVM-DBN)效果最好。本研究表明,使用DBN和SVM混合算法可以增强面向查询的多文档摘要。
{"title":"Hybridization of DBN with SVM and its Impact on Performance in Multi-Document Summarization","authors":"","doi":"10.5121/mlaij.2021.8304","DOIUrl":"https://doi.org/10.5121/mlaij.2021.8304","url":null,"abstract":"Data available from web based sources has grown tremendously with growth of the internet. Users interested in information from such sources often use a search engine to obtain the data which they edit for presentation to their audience. This process can be tedious especially when it involves the generation of a summary. One way to ease the process is by automation of the summary generation process. Efforts by researchers towards automatic summarization have yielded several approaches among them machine learning. Thus, recommendations have been made on combining the algorithms with different strengths, also called hybridization, in order to enhance their performance. Therefore, this research sought to establish the impact of hybridization of Deep Belief Network (DBN) with Support Vector Machine (SVM) on precision, recall, accuracy and F-measure when used in the case of query oriented multi-document summarization. The experiments were carried out using data from National Institute of Standards and Technology (NIST), Document Understanding Conference (DUC) 2006. The data was split into training and test data and used appropriately in DBN, SVM, SVM-DBN hybrid and DBN-SVM hybrid. Results indicated that the hybridized algorithm has better precision, accuracy and F-measure as compared to DBN. Pre-classification hybridization of DBN with SVM (SVM-DBN) gives the best results. This research implies that use of DBN and SVM hybrid algorithms would enhance query oriented multi-document summarization.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123051072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Development Framework for a Conversational Agent to Explore Machine Learning Concepts 探索机器学习概念的对话代理开发框架
Pub Date : 2021-09-30 DOI: 10.5121/mlaij.2021.8301
Ayse Kok Arslan
This study aims to introduce a discussion platform and curriculum designed to help people understand how machines learn. Research shows how to train an agent through dialogue and understand how information is represented using visualization. This paper starts by providing a comprehensive definition of AI literacy based on existing research and integrates a wide range of different subject documents into a set of key AI literacy skills to develop a user-centered AI. This functionality and structural considerations are organized into a conceptual framework based on the literature. Contributions to this paper can be used to initiate discussion and guide future research on AI learning within the computer science community.
本研究旨在介绍一个讨论平台和课程,旨在帮助人们了解机器如何学习。研究展示了如何通过对话来训练智能体,以及如何使用可视化来理解信息是如何表示的。本文首先在现有研究的基础上提供了人工智能素养的全面定义,并将广泛的不同主题文档集成到一套关键的人工智能素养技能中,以开发以用户为中心的人工智能。这些功能和结构上的考虑被组织成一个基于文献的概念框架。对本文的贡献可以用来在计算机科学社区内发起讨论和指导未来的人工智能学习研究。
{"title":"A Development Framework for a Conversational Agent to Explore Machine Learning Concepts","authors":"Ayse Kok Arslan","doi":"10.5121/mlaij.2021.8301","DOIUrl":"https://doi.org/10.5121/mlaij.2021.8301","url":null,"abstract":"This study aims to introduce a discussion platform and curriculum designed to help people understand how machines learn. Research shows how to train an agent through dialogue and understand how information is represented using visualization. This paper starts by providing a comprehensive definition of AI literacy based on existing research and integrates a wide range of different subject documents into a set of key AI literacy skills to develop a user-centered AI. This functionality and structural considerations are organized into a conceptual framework based on the literature. Contributions to this paper can be used to initiate discussion and guide future research on AI learning within the computer science community.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127883930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Enhancement for the Consistent Depth Estimation of Monocular Videos using Lightweight Network 基于轻量级网络的单目视频一致深度估计增强
Pub Date : 2021-09-30 DOI: 10.5121/mlaij.2021.8302
Mohamed N. Sweilam, N. Tolstokulakov
Depth estimation has made great progress in the last few years due to its applications in robotics science and computer vision. Various methods have been implemented and enhanced to estimate the depth without flickers and missing holes. Despite this progress, it is still one of the main challenges for researchers, especially for the video applications which have more complexity of the neural network which af ects the run time. Moreover to use such input like monocular video for depth estimation is considered an attractive idea, particularly for hand-held devices such as mobile phones, they are very popular for capturing pictures and videos, in addition to having a limited amount of RAM. Here in this work, we focus on enhancing the existing consistent depth estimation for monocular videos approach to be with less usage of RAM and with using less number of parameters without having a significant reduction in the quality of the depth estimation.
深度估计由于其在机器人科学和计算机视觉中的应用,在过去的几年里取得了很大的进展。已经实现和改进了各种方法来估计无闪烁和漏孔的深度。尽管取得了这些进展,但它仍然是研究人员面临的主要挑战之一,特别是在视频应用中,神经网络的复杂性较大,影响了运行时间。此外,使用像单目视频这样的输入来进行深度估计被认为是一个有吸引力的想法,特别是对于像手机这样的手持设备,它们非常受欢迎,用于捕捉图片和视频,此外还有有限的RAM。在这项工作中,我们专注于增强现有的单目视频一致深度估计方法,以减少RAM的使用和使用更少的参数,而不会显著降低深度估计的质量。
{"title":"An Enhancement for the Consistent Depth Estimation of Monocular Videos using Lightweight Network","authors":"Mohamed N. Sweilam, N. Tolstokulakov","doi":"10.5121/mlaij.2021.8302","DOIUrl":"https://doi.org/10.5121/mlaij.2021.8302","url":null,"abstract":"Depth estimation has made great progress in the last few years due to its applications in robotics science and computer vision. Various methods have been implemented and enhanced to estimate the depth without flickers and missing holes. Despite this progress, it is still one of the main challenges for researchers, especially for the video applications which have more complexity of the neural network which af ects the run time. Moreover to use such input like monocular video for depth estimation is considered an attractive idea, particularly for hand-held devices such as mobile phones, they are very popular for capturing pictures and videos, in addition to having a limited amount of RAM. Here in this work, we focus on enhancing the existing consistent depth estimation for monocular videos approach to be with less usage of RAM and with using less number of parameters without having a significant reduction in the quality of the depth estimation.","PeriodicalId":347528,"journal":{"name":"Machine Learning and Applications: An International Journal","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131078505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Machine Learning and Applications: An International Journal
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1