首页 > 最新文献

2022 25th International Conference on Computer and Information Technology (ICCIT)最新文献

英文 中文
Stock Price Prediction: A Time Series Analysis 股票价格预测:一个时间序列分析
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10056009
Fatima Juairiah, Mostafa Mahatabe, Hasan Bin Jamal, Aysha Shiddika, Tanvir Rouf Shawon, Nibir Chandra Mandal
Predicting future stock volatility has always been a demanding chore for research studies. Individuals around the world have long regarded the stock market as a substantial profit. A stock data set contains numerous precise terms that are difficult for an individual to comprehend when considering stock market expenditures. An essential manifestation of a stock’s performance on the stock market is its closing price, but it is challenging to estimate the stock market’s price movements. This study aims to provide a future market scenario supported by statistical data. We used the Microsoft Corporation Stock dataset from 1986 to 2022. To foresee stock market volatility, we used time series analysis with the Long Short-Term Memory (LSTM), Bidirectional Long-short Term Memory (Bi-LSTM), Autoregressive Integrated Moving Average (ARIMA), Hidden Markov Model (HMM), and Multi-Head Attention. We have achieved 0.153, 0.202, 6.674, 14.760, and 21.493 for Transformer, HMM, ARIMA, BiLSTM, and LSTM respectively.
预测未来股票波动对研究来说一直是一项艰巨的任务。长期以来,世界各地的个人都认为股票市场是一笔丰厚的利润。股票数据集包含许多精确的术语,当考虑股票市场支出时,个人很难理解。股票在股票市场上表现的一个基本表现是它的收盘价,但估计股票市场的价格变动是具有挑战性的。本研究的目的是提供一个未来的市场情景支持的统计数据。我们使用了1986年至2022年的微软公司股票数据集。为了预测股票市场的波动,我们使用了长短期记忆(LSTM)、双向长短期记忆(Bi-LSTM)、自回归综合移动平均(ARIMA)、隐马尔可夫模型(HMM)和多头注意的时间序列分析。我们对Transformer、HMM、ARIMA、BiLSTM和LSTM分别实现了0.153、0.202、6.674、14.760和21.493。
{"title":"Stock Price Prediction: A Time Series Analysis","authors":"Fatima Juairiah, Mostafa Mahatabe, Hasan Bin Jamal, Aysha Shiddika, Tanvir Rouf Shawon, Nibir Chandra Mandal","doi":"10.1109/ICCIT57492.2022.10056009","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10056009","url":null,"abstract":"Predicting future stock volatility has always been a demanding chore for research studies. Individuals around the world have long regarded the stock market as a substantial profit. A stock data set contains numerous precise terms that are difficult for an individual to comprehend when considering stock market expenditures. An essential manifestation of a stock’s performance on the stock market is its closing price, but it is challenging to estimate the stock market’s price movements. This study aims to provide a future market scenario supported by statistical data. We used the Microsoft Corporation Stock dataset from 1986 to 2022. To foresee stock market volatility, we used time series analysis with the Long Short-Term Memory (LSTM), Bidirectional Long-short Term Memory (Bi-LSTM), Autoregressive Integrated Moving Average (ARIMA), Hidden Markov Model (HMM), and Multi-Head Attention. We have achieved 0.153, 0.202, 6.674, 14.760, and 21.493 for Transformer, HMM, ARIMA, BiLSTM, and LSTM respectively.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127954092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An efficient Dual-Band CNN for Motor Imagery EEG Signal Classification 一种高效的双频CNN运动图像脑电信号分类方法
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055792
M. Rana, Shaikh Anowarul
Brain-computer interface (BCI) systems have recently gained much attention due to their ability to interpret human thought processes from electroencephalogram (EEG)-based motor imagery (MI) signals. Among various machine learning based methods, instead of using hand-crafted features, end-to-end deep learning (DL) based methods are getting popularity due to their very satisfactory performance. Even in case of non-stationary and subject specific MI-EEG signals, DL methods offer high accuracy in classifying MI tasks. In this paper, a dual band convolutional neural network (DBCNN) based on MI-EEG two band limited efficient signals is proposed. Here, beta wave signals are used to improve classification performance in motor imagery tasks. Beta waves are associated with motor imagery tasks. On the other hand our second signal is the pre-processed bandpass signal. The DBCNN model uses temporal CNN and spatial CNN to explore temporal and spatial features. This model also uses a spatial convolutional network for single channels from multiple channels. Finally the data are classified for MI-EEG signal classification accuracy. An extensive experiment on MI-based BCI IV 2a EEG dataset on nine subjects shows that very satisfactory classification performance is achieved. The proposed model offers higher classification accuracy that is obtained by some state-of-the-art methods.
脑机接口(BCI)系统由于能够从基于脑电图(EEG)的运动图像(MI)信号中解释人类的思维过程,最近受到了广泛的关注。在各种基于机器学习的方法中,基于端到端深度学习(DL)的方法由于其非常令人满意的性能而越来越受欢迎,而不是使用手工制作的特征。即使在非平稳和受试者特定的MI- eeg信号情况下,深度学习方法对MI任务的分类也具有很高的准确性。本文提出了一种基于MI-EEG两波段有限有效信号的双波段卷积神经网络(DBCNN)。在这里,β波信号被用来提高运动想象任务的分类性能。β波与运动想象任务有关。另一方面,我们的第二个信号是预处理带通信号。DBCNN模型使用时间CNN和空间CNN来探索时空特征。该模型还使用空间卷积网络对多个通道中的单个通道进行处理。最后对数据进行分类,提高脑电信号的分类精度。在基于mi的BCI IV 2a脑电数据集上对9个被试进行了大量实验,结果表明该方法取得了令人满意的分类效果。所提出的模型具有比现有方法更高的分类精度。
{"title":"An efficient Dual-Band CNN for Motor Imagery EEG Signal Classification","authors":"M. Rana, Shaikh Anowarul","doi":"10.1109/ICCIT57492.2022.10055792","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055792","url":null,"abstract":"Brain-computer interface (BCI) systems have recently gained much attention due to their ability to interpret human thought processes from electroencephalogram (EEG)-based motor imagery (MI) signals. Among various machine learning based methods, instead of using hand-crafted features, end-to-end deep learning (DL) based methods are getting popularity due to their very satisfactory performance. Even in case of non-stationary and subject specific MI-EEG signals, DL methods offer high accuracy in classifying MI tasks. In this paper, a dual band convolutional neural network (DBCNN) based on MI-EEG two band limited efficient signals is proposed. Here, beta wave signals are used to improve classification performance in motor imagery tasks. Beta waves are associated with motor imagery tasks. On the other hand our second signal is the pre-processed bandpass signal. The DBCNN model uses temporal CNN and spatial CNN to explore temporal and spatial features. This model also uses a spatial convolutional network for single channels from multiple channels. Finally the data are classified for MI-EEG signal classification accuracy. An extensive experiment on MI-based BCI IV 2a EEG dataset on nine subjects shows that very satisfactory classification performance is achieved. The proposed model offers higher classification accuracy that is obtained by some state-of-the-art methods.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129885918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Video Captioning in Bengali With Visual Attention 视频字幕用孟加拉语加上视觉注意
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055190
Suvom Shaha, F. Shah, Amir Hossain Raj, Ashek Seum, Saiful Islam, Sifat Ahmed
Generating automatic video captions is one of the most challenging Artificial Intelligence tasks as it combines Computer Vision and Natural Language Processing research areas. The task is more difficult for a complex language like Bengali as there is a general lack of video captioning datasets in the Bengali language. To overcome this challenge, we introduce a fully human-annotated dataset of Bengali captions in this research for the videos of the MSVD dataset. We have proposed a novel end-to-end architecture with an attention-based decoder to generate meaningful video captions in the Bengali language. First, spatial and temporal features of videos are combined using Bidirectional Gated Recurrent Units (Bi-GRU) that generate the input feature, which is later fed to the attention layer along with embedded caption features. This attention mechanism explores the interdependence between visual and textual representations. Then, a double-layered GRU takes these combined attention features for generating meaningful sentences. We trained this model on our proposed dataset and achieved 39.35% in BLEU-4, 59.67% in CIDEr, and 65.34% score in ROUGE. This is the state-of-the-art result compared to any other video captioning work available in the Bengali language.
自动生成视频字幕是最具挑战性的人工智能任务之一,因为它结合了计算机视觉和自然语言处理的研究领域。对于像孟加拉语这样的复杂语言来说,这项任务更加困难,因为孟加拉语的视频字幕数据集普遍缺乏。为了克服这一挑战,我们在本研究中为MSVD数据集的视频引入了一个完全人工注释的孟加拉语字幕数据集。我们提出了一种新颖的端到端架构,使用基于注意力的解码器来生成有意义的孟加拉语视频字幕。首先,使用双向门控循环单元(Bi-GRU)将视频的空间和时间特征结合起来,生成输入特征,然后将其与嵌入的字幕特征一起馈送到注意层。这种注意机制探讨了视觉表征和文本表征之间的相互依存关系。然后,一个双层GRU利用这些组合的注意特征来生成有意义的句子。我们在我们提出的数据集上训练该模型,BLEU-4的得分为39.35%,CIDEr的得分为59.67%,ROUGE的得分为65.34%。与任何其他可用的孟加拉语视频字幕工作相比,这是最先进的结果。
{"title":"Video Captioning in Bengali With Visual Attention","authors":"Suvom Shaha, F. Shah, Amir Hossain Raj, Ashek Seum, Saiful Islam, Sifat Ahmed","doi":"10.1109/ICCIT57492.2022.10055190","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055190","url":null,"abstract":"Generating automatic video captions is one of the most challenging Artificial Intelligence tasks as it combines Computer Vision and Natural Language Processing research areas. The task is more difficult for a complex language like Bengali as there is a general lack of video captioning datasets in the Bengali language. To overcome this challenge, we introduce a fully human-annotated dataset of Bengali captions in this research for the videos of the MSVD dataset. We have proposed a novel end-to-end architecture with an attention-based decoder to generate meaningful video captions in the Bengali language. First, spatial and temporal features of videos are combined using Bidirectional Gated Recurrent Units (Bi-GRU) that generate the input feature, which is later fed to the attention layer along with embedded caption features. This attention mechanism explores the interdependence between visual and textual representations. Then, a double-layered GRU takes these combined attention features for generating meaningful sentences. We trained this model on our proposed dataset and achieved 39.35% in BLEU-4, 59.67% in CIDEr, and 65.34% score in ROUGE. This is the state-of-the-art result compared to any other video captioning work available in the Bengali language.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115188309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BanglaSAPM: A Deep Learning Model for Suicidal Attempt Prediction Using Social Media Content in Bangla BanglaSAPM:使用孟加拉社交媒体内容预测自杀企图的深度学习模型
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055237
Sabiha Islam, Md. Shafiul Alam Forhad, Hasan Murad
Nowadays, people are constantly expressing their emotions, thoughts, opinions, and daily activities on social media. Therefore, social media posts have become a powerful tool among psychiatrists for the early detection of suicidal tendencies. However, the automatic detection of suicidal posts has become a challenging problem among researchers in the field of Natural Language Processing (NLP). A significant number of previous works have been found in the literature for the automatic detection of suicidal posts in different languages such as English. However, little effort has been devoted to automatically detecting suicidal posts in low-resource languages like Bangla due to the lack of available datasets. In this study, we have created a noble Bangla suicidal posts dataset named BanglaSPD and compared the performance of various machine learning and deep learning models for suicide attempt prediction by training and evaluating with the dataset. Finally, we have found that a deep learning-based model with CNN+BiLSTM has outperformed with 0.61 F1-Score in Fasttext word embedding methods.
如今,人们不断地在社交媒体上表达自己的情绪、想法、观点和日常活动。因此,社交媒体帖子已经成为精神科医生早期发现自杀倾向的有力工具。然而,自杀帖子的自动检测一直是自然语言处理(NLP)领域研究人员面临的一个难题。在不同语言(如英语)的自杀帖子自动检测的文献中,已经发现了大量先前的工作。然而,由于缺乏可用的数据集,很少有人致力于自动检测孟加拉语等低资源语言的自杀帖子。在本研究中,我们创建了一个名为BanglaSPD的高贵孟加拉语自杀帖子数据集,并通过对数据集进行训练和评估,比较了各种机器学习和深度学习模型在自杀企图预测方面的性能。最后,我们发现基于CNN+BiLSTM的深度学习模型在Fasttext词嵌入方法中的表现优于0.61 F1-Score。
{"title":"BanglaSAPM: A Deep Learning Model for Suicidal Attempt Prediction Using Social Media Content in Bangla","authors":"Sabiha Islam, Md. Shafiul Alam Forhad, Hasan Murad","doi":"10.1109/ICCIT57492.2022.10055237","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055237","url":null,"abstract":"Nowadays, people are constantly expressing their emotions, thoughts, opinions, and daily activities on social media. Therefore, social media posts have become a powerful tool among psychiatrists for the early detection of suicidal tendencies. However, the automatic detection of suicidal posts has become a challenging problem among researchers in the field of Natural Language Processing (NLP). A significant number of previous works have been found in the literature for the automatic detection of suicidal posts in different languages such as English. However, little effort has been devoted to automatically detecting suicidal posts in low-resource languages like Bangla due to the lack of available datasets. In this study, we have created a noble Bangla suicidal posts dataset named BanglaSPD and compared the performance of various machine learning and deep learning models for suicide attempt prediction by training and evaluating with the dataset. Finally, we have found that a deep learning-based model with CNN+BiLSTM has outperformed with 0.61 F1-Score in Fasttext word embedding methods.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115288390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Road Traffic Update and Control Traffic Light using Dynamic Patterns from Video Streaming Data 利用视频流数据的动态模式更新和控制道路交通信号灯
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055520
Md. Shafkat Islam, Md. Maharub Hossain, Mizbatul Jannat Refat
Controlling traffic congestion is one of the most challenging issues in the real world, and an automated traffic monitoring system established on the computer vision and internet of things is an emerging scientific field in research. Traffic Congestion becomes severe in the most densely populated area to move timely from one place to another, passing through massive traffic. Although several research papers have presented their solutions for monitoring traffic congestion, we do not receive a permanent solution. In this paper, We try to find a different but unique way to overcome it. Our proposed system works with real-time video sequential data and detects different vehicle shapes using contour-based learning and a convex hull algorithm. We execute the procedure in a way where we work with per-second frames (fps) of the video, and then we determine the total, running, and motionless stagnant vehicles in each frame. After observing vehicle differentiation, an automated decision will be generated, i.e., JAM or NOT JAM which will be transmitted to the Map for users’ assistance. After that, the decision about the traffic situation of all nodes makes a combination, and then we find two patterns, one will identify the traffic situation of all connected roads, and another will control the traffic signal automatically.
控制交通拥堵是现实世界中最具挑战性的问题之一,建立在计算机视觉和物联网基础上的自动交通监控系统是一个新兴的科学研究领域。在人口最密集的地区,及时从一个地方移动到另一个地方,要经过大量的交通,交通拥堵变得严重。虽然有几篇研究论文提出了监测交通拥堵的解决方案,但我们并没有得到一个永久的解决方案。在本文中,我们试图找到一种不同但独特的方法来克服它。我们提出的系统处理实时视频序列数据,并使用基于轮廓的学习和凸包算法检测不同的车辆形状。我们以每秒帧(fps)视频的方式执行该过程,然后我们确定每帧中的总数,运行和静止的停滞车辆。在观察到车辆的差异后,将自动生成一个决策,即JAM或NOT JAM,并将该决策传输到地图以供用户协助。然后对所有节点的交通状况进行决策组合,得到两种模式,一种模式识别所有连接道路的交通状况,另一种模式自动控制交通信号。
{"title":"Road Traffic Update and Control Traffic Light using Dynamic Patterns from Video Streaming Data","authors":"Md. Shafkat Islam, Md. Maharub Hossain, Mizbatul Jannat Refat","doi":"10.1109/ICCIT57492.2022.10055520","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055520","url":null,"abstract":"Controlling traffic congestion is one of the most challenging issues in the real world, and an automated traffic monitoring system established on the computer vision and internet of things is an emerging scientific field in research. Traffic Congestion becomes severe in the most densely populated area to move timely from one place to another, passing through massive traffic. Although several research papers have presented their solutions for monitoring traffic congestion, we do not receive a permanent solution. In this paper, We try to find a different but unique way to overcome it. Our proposed system works with real-time video sequential data and detects different vehicle shapes using contour-based learning and a convex hull algorithm. We execute the procedure in a way where we work with per-second frames (fps) of the video, and then we determine the total, running, and motionless stagnant vehicles in each frame. After observing vehicle differentiation, an automated decision will be generated, i.e., JAM or NOT JAM which will be transmitted to the Map for users’ assistance. After that, the decision about the traffic situation of all nodes makes a combination, and then we find two patterns, one will identify the traffic situation of all connected roads, and another will control the traffic signal automatically.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132072250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Blockchain Based Secure Framework for User-centric Multi-party Skyline Queries 基于区块链的以用户为中心的多方Skyline查询安全框架
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055315
Md. Motaleb Hossen Manik, Kazi Md. Rokibul Alam, Y. Morimoto
For commercial purposes, business organizations with mutual interests while competing with one another usually intend to keep their datasets to be confidential. Secure multiparty skyline queries enable these organizations to retrieve dominated data objects from their sensitive datasets. Here, a distinction among datasets according to priority, retaining data privacy along with integrity, targeted enquires, etc., are crucial. This paper proposes a secure framework for multi-party skyline queries that incorporates these requirements together. To prioritize datasets it assigns distinct weights to parties’ datasets. To retain adequate privacy it exploits a cryptosystem that engages a single key for encryption but multiple separate keys for decryption. To attain data anonymity it re-encrypts and shuffles the encrypted data. To assure data integrity it deploys a unique block comprising encrypted data designated for every party within the blockchain storage. To enable enlisted users to ask intended queries it accepts their inquiries. This paper is a preliminary report and evaluation of the proposed framework.
出于商业目的,具有共同利益的商业组织在相互竞争的同时通常打算对其数据集保密。安全的多方天际线查询使这些组织能够从其敏感数据集中检索主导数据对象。在这里,根据优先级区分数据集,保留数据隐私以及完整性,有针对性的查询等是至关重要的。本文提出了一个多方天际线查询的安全框架,将这些需求结合在一起。为了优先考虑数据集,它为各方的数据集分配不同的权重。为了保持足够的隐私,它利用了一种密码系统,该系统使用单个密钥进行加密,但使用多个单独的密钥进行解密。为了获得数据匿名性,它对加密后的数据进行重新加密和洗牌。为了确保数据的完整性,它部署了一个独特的块,其中包含为区块链存储中的每一方指定的加密数据。为了使登记的用户能够提出预期的查询,它接受他们的查询。本文是对拟议框架的初步报告和评估。
{"title":"A Blockchain Based Secure Framework for User-centric Multi-party Skyline Queries","authors":"Md. Motaleb Hossen Manik, Kazi Md. Rokibul Alam, Y. Morimoto","doi":"10.1109/ICCIT57492.2022.10055315","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055315","url":null,"abstract":"For commercial purposes, business organizations with mutual interests while competing with one another usually intend to keep their datasets to be confidential. Secure multiparty skyline queries enable these organizations to retrieve dominated data objects from their sensitive datasets. Here, a distinction among datasets according to priority, retaining data privacy along with integrity, targeted enquires, etc., are crucial. This paper proposes a secure framework for multi-party skyline queries that incorporates these requirements together. To prioritize datasets it assigns distinct weights to parties’ datasets. To retain adequate privacy it exploits a cryptosystem that engages a single key for encryption but multiple separate keys for decryption. To attain data anonymity it re-encrypts and shuffles the encrypted data. To assure data integrity it deploys a unique block comprising encrypted data designated for every party within the blockchain storage. To enable enlisted users to ask intended queries it accepts their inquiries. This paper is a preliminary report and evaluation of the proposed framework.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"76 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130620013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transfer learning based Facial Emotion Detection for Animated Characters 基于迁移学习的动画人物面部情绪检测
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10054823
Anonnya Ghosh, Raqeebir Rab, Ashikur Rahman
Since 1906 till today animation has been a popular method of storytelling. Animated characters portray unique stories in a more perceivable way by expressing diverse facial expressions. Detection of these emotions has not yet gained as much acclaim as detection of human facial expressions. This research aims to classify and predict emotions of animated characters using Deep Learning Techniques. Images of seven facial expressions namely Anger, Disgust, Fear, Joy, Neutral, Sadness and Surprise are classified using Residual Network (ResNet) and Transfer Learning. In order to conduct research on emotion identification in animated faces, we generated a new dataset with fewer images than the existing dataset [1]. Features from the faces are extracted using Uniform Local Binary Patterns (LBP) and fed to Convolutional Neural Network (CNN) model for classification. In our proposed models, the Transfer Learning-based ResNet50 and ResNet101,were trained on the ImageNet [10] dataset. Among the models ResNet101 achieved the highest detection accuracy of 94% and ResNet50 showed lowest time complexity.
自1906年至今,动画一直是一种流行的讲故事的方法。动画人物通过表达不同的面部表情,以更直观的方式描绘独特的故事。对这些情绪的检测还没有像对人类面部表情的检测那样受到广泛的赞誉。本研究旨在利用深度学习技术对动画人物的情绪进行分类和预测。利用残差网络(ResNet)和迁移学习对愤怒、厌恶、恐惧、喜悦、中性、悲伤和惊讶七种面部表情图像进行分类。为了对动画人脸的情绪识别进行研究,我们生成了一个比现有数据集图像更少的新数据集[1]。使用统一局部二值模式(LBP)提取人脸特征,并将其输入卷积神经网络(CNN)模型进行分类。在我们提出的模型中,基于迁移学习的ResNet50和ResNet101是在ImageNet[10]数据集上进行训练的。其中,ResNet101的检测准确率最高,达到94%,而ResNet50的时间复杂度最低。
{"title":"Transfer learning based Facial Emotion Detection for Animated Characters","authors":"Anonnya Ghosh, Raqeebir Rab, Ashikur Rahman","doi":"10.1109/ICCIT57492.2022.10054823","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10054823","url":null,"abstract":"Since 1906 till today animation has been a popular method of storytelling. Animated characters portray unique stories in a more perceivable way by expressing diverse facial expressions. Detection of these emotions has not yet gained as much acclaim as detection of human facial expressions. This research aims to classify and predict emotions of animated characters using Deep Learning Techniques. Images of seven facial expressions namely Anger, Disgust, Fear, Joy, Neutral, Sadness and Surprise are classified using Residual Network (ResNet) and Transfer Learning. In order to conduct research on emotion identification in animated faces, we generated a new dataset with fewer images than the existing dataset [1]. Features from the faces are extracted using Uniform Local Binary Patterns (LBP) and fed to Convolutional Neural Network (CNN) model for classification. In our proposed models, the Transfer Learning-based ResNet50 and ResNet101,were trained on the ImageNet [10] dataset. Among the models ResNet101 achieved the highest detection accuracy of 94% and ResNet50 showed lowest time complexity.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"512 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123570202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generative Adversarial Network-based Augmented Rice Leaf Disease Detection using Deep Learning
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055716
Syed Taha Yeasin Ramadan, T. Sakib, Md. Mizba Ul Haque, N. Sharmin, Md. Mahbubur Rahman
Rice is one of the most produced crops in the world and the staple food in many South Asian countries. Rice leaf disease can affect the production of rice vastly, which can be prevented through the early detection of it. Many machine learning techniques have been used in recent years to help in the prevention of one of the most serious concerns, which is disease transmission.But there are limited images available of diseased leaf compared to healthy images which makes life tougher for machine learning models as they need a good amount of data for training. To solve this problem, a Generative adversarial network(GAN) has been used in recent days to create new, synthetic instances of an image that can pass as a real image. Recently, it has been used widely in the field of leaf disease identification. But there is very limited work done on rice diseases. In this paper, SRGAN (Super Resolution-GAN) has been considered as a data augmentation method to balance the dataset. Afterward, DenseNet121, DenseNet169, MobileNetV2, and VGG16 have been applied to classify the diseases. Experiment results show that the newly created augmented dataset produces the best results with both DenseNet169 and moboleNetV2 when compared to other models, with high accuracy of 94.30.
水稻是世界上产量最高的作物之一,也是许多南亚国家的主食。水稻叶病对水稻生产的影响很大,可以通过早期发现来预防。近年来,许多机器学习技术被用于帮助预防最严重的问题之一,即疾病传播。但是,与健康图像相比,患病叶子的图像有限,这使得机器学习模型的生活更加困难,因为它们需要大量的数据进行训练。为了解决这个问题,最近使用了生成对抗网络(GAN)来创建新的合成图像实例,这些实例可以作为真实图像传递。近年来,它在叶片病害鉴定领域得到了广泛的应用。但是对水稻病害的研究非常有限。本文将SRGAN (Super Resolution-GAN)作为一种数据增强方法来平衡数据集。随后应用DenseNet121、DenseNet169、MobileNetV2和VGG16对疾病进行分类。实验结果表明,与其他模型相比,新创建的增强数据集在DenseNet169和moboleNetV2上都能产生最好的结果,准确率高达94.30。
{"title":"Generative Adversarial Network-based Augmented Rice Leaf Disease Detection using Deep Learning","authors":"Syed Taha Yeasin Ramadan, T. Sakib, Md. Mizba Ul Haque, N. Sharmin, Md. Mahbubur Rahman","doi":"10.1109/ICCIT57492.2022.10055716","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055716","url":null,"abstract":"Rice is one of the most produced crops in the world and the staple food in many South Asian countries. Rice leaf disease can affect the production of rice vastly, which can be prevented through the early detection of it. Many machine learning techniques have been used in recent years to help in the prevention of one of the most serious concerns, which is disease transmission.But there are limited images available of diseased leaf compared to healthy images which makes life tougher for machine learning models as they need a good amount of data for training. To solve this problem, a Generative adversarial network(GAN) has been used in recent days to create new, synthetic instances of an image that can pass as a real image. Recently, it has been used widely in the field of leaf disease identification. But there is very limited work done on rice diseases. In this paper, SRGAN (Super Resolution-GAN) has been considered as a data augmentation method to balance the dataset. Afterward, DenseNet121, DenseNet169, MobileNetV2, and VGG16 have been applied to classify the diseases. Experiment results show that the newly created augmented dataset produces the best results with both DenseNet169 and moboleNetV2 when compared to other models, with high accuracy of 94.30.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126728004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Comparative Analysis of Process Scheduling Algorithm using AI models 基于AI模型的进程调度算法比较分析
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055395
Md. Moynul Asik Moni, Maharshi Niloy, Aquibul Haq Chowdhury, Farah Jasmin Khan, Md. Fahmid-Ul-Alam Juboraj, Amitabha Chakrabarty
Process scheduling is an integral part of operating systems. The most widely used scheduling algorithm in operating systems is Round Robin, but the average waiting time in RR is often quite long. The purpose of this study is to propose a new algorithm to minimize waiting time and process starvation by determining the optimal time quantum by predicting CPU burst time. For burst time prediction, we are using the machine learning algorithms like linear regression, decision tree, k-nearest neighbors, and Neural Network Model Multi-Layer Perceptron. Moreover, for 10000 predicted burst time of processes with the same configuration, we have compared the average turnaround time, the average waiting time and the number of context switches of the proposed modified round robin algorithm with Traditional Round Robin, Modified Round Robin, Optimized Round Robin and Self-Adjustment Round Robin. The proposed modified round robin i.e. Absolute Difference Based Time Quantum Round Robin (ADRR) is found to be almost 2 times faster than the other algorithm in terms of process scheduling for the used dataset which contains a huge load of processes.
进程调度是操作系统的一个组成部分。操作系统中使用最广泛的调度算法是轮询调度,但轮询调度的平均等待时间往往很长。本研究的目的是提出一种新的算法,通过预测CPU突发时间来确定最佳的时间量,从而最大限度地减少等待时间和进程饥饿。对于突发时间预测,我们使用机器学习算法,如线性回归、决策树、k近邻和神经网络模型多层感知器。此外,对于相同配置的10000个预测突发时间的进程,我们比较了改进轮询算法与传统轮询、改进轮询、优化轮询和自调整轮询的平均周转时间、平均等待时间和上下文切换次数。对于包含大量进程的数据集,本文提出的改进轮询算法即基于绝对差分的时间量子轮询算法(ADRR)在进程调度方面比其他算法快近2倍。
{"title":"Comparative Analysis of Process Scheduling Algorithm using AI models","authors":"Md. Moynul Asik Moni, Maharshi Niloy, Aquibul Haq Chowdhury, Farah Jasmin Khan, Md. Fahmid-Ul-Alam Juboraj, Amitabha Chakrabarty","doi":"10.1109/ICCIT57492.2022.10055395","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055395","url":null,"abstract":"Process scheduling is an integral part of operating systems. The most widely used scheduling algorithm in operating systems is Round Robin, but the average waiting time in RR is often quite long. The purpose of this study is to propose a new algorithm to minimize waiting time and process starvation by determining the optimal time quantum by predicting CPU burst time. For burst time prediction, we are using the machine learning algorithms like linear regression, decision tree, k-nearest neighbors, and Neural Network Model Multi-Layer Perceptron. Moreover, for 10000 predicted burst time of processes with the same configuration, we have compared the average turnaround time, the average waiting time and the number of context switches of the proposed modified round robin algorithm with Traditional Round Robin, Modified Round Robin, Optimized Round Robin and Self-Adjustment Round Robin. The proposed modified round robin i.e. Absolute Difference Based Time Quantum Round Robin (ADRR) is found to be almost 2 times faster than the other algorithm in terms of process scheduling for the used dataset which contains a huge load of processes.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114505270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Deep Learning-Based Bengali Visual Question Answering System 基于深度学习的孟加拉语视觉问答系统
Pub Date : 2022-12-17 DOI: 10.1109/ICCIT57492.2022.10055205
Mahamudul Hasan Rafi, Shifat Islam, S. M. Hasan Imtiaz Labib, SM Sajid Hasan, F. Shah, Sifat Ahmed
Visual Question Answering (VQA) is a challenging task in Artificial Intelligence (AI), where an AI agent answers questions regarding visual content based on images provided. Therefore, to implement a VQA system, a computer system requires complex reasoning over visual aspects of images and textual parts of the questions to anticipate the correct answer. Although there is a good deal of VQA research in English, Bengali still needs to thoroughly explore this area of artificial intelligence. To address this, we have constructed a Bengali VQA dataset by preparing human-annotated question-answers using a small portion of the images from the VQA v2.0 dataset. To overcome high linguistic priors that hide the importance of precise visual information in visual question answering, we have used real-life scenarios to construct a balanced Bengali VQA dataset. This is the first human-annotated dataset of this kind in Bengali. We have proposed a Top-Down Attention-based approach in this study and conducted several studies to assess our model’s performance.
视觉问答(VQA)在人工智能(AI)中是一项具有挑战性的任务,人工智能代理根据提供的图像回答有关视觉内容的问题。因此,为了实现VQA系统,计算机系统需要对图像的视觉方面和问题的文本部分进行复杂的推理,以预测正确的答案。尽管在英语中有大量的VQA研究,但孟加拉语仍然需要深入探索人工智能的这一领域。为了解决这个问题,我们使用VQA v2.0数据集中的一小部分图像来准备人工注释的问答,从而构建了一个孟加拉语VQA数据集。为了克服隐藏精确视觉信息在视觉问答中的重要性的高语言先验,我们使用现实场景构建了一个平衡的孟加拉语VQA数据集。这是第一个在孟加拉语中人工注释的数据集。我们在本研究中提出了一种自上而下的基于注意力的方法,并进行了几项研究来评估我们的模型的性能。
{"title":"A Deep Learning-Based Bengali Visual Question Answering System","authors":"Mahamudul Hasan Rafi, Shifat Islam, S. M. Hasan Imtiaz Labib, SM Sajid Hasan, F. Shah, Sifat Ahmed","doi":"10.1109/ICCIT57492.2022.10055205","DOIUrl":"https://doi.org/10.1109/ICCIT57492.2022.10055205","url":null,"abstract":"Visual Question Answering (VQA) is a challenging task in Artificial Intelligence (AI), where an AI agent answers questions regarding visual content based on images provided. Therefore, to implement a VQA system, a computer system requires complex reasoning over visual aspects of images and textual parts of the questions to anticipate the correct answer. Although there is a good deal of VQA research in English, Bengali still needs to thoroughly explore this area of artificial intelligence. To address this, we have constructed a Bengali VQA dataset by preparing human-annotated question-answers using a small portion of the images from the VQA v2.0 dataset. To overcome high linguistic priors that hide the importance of precise visual information in visual question answering, we have used real-life scenarios to construct a balanced Bengali VQA dataset. This is the first human-annotated dataset of this kind in Bengali. We have proposed a Top-Down Attention-based approach in this study and conducted several studies to assess our model’s performance.","PeriodicalId":255498,"journal":{"name":"2022 25th International Conference on Computer and Information Technology (ICCIT)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115918267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2022 25th International Conference on Computer and Information Technology (ICCIT)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1