首页 > 最新文献

Journal of Information Systems Engineering and Business Intelligence最新文献

英文 中文
A Fast and Reliable Approach for COVID-19 Detection from CT-Scan Images 一种快速可靠的ct扫描图像COVID-19检测方法
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.288-304
Md. Jawwad Bin Zahir, Muhammad Anwarul Azim, Abu Nowshed Chy, Mohammad Khairul Islam
Background: COVID-19 is a highly contagious respiratory disease with multiple mutant variants, an asymptotic nature in patients, and with potential to stay undetected in common tests, which makes it deadlier, more transmissible, and harder to detect. Regardless of variants, the COVID-19 infection shows several observable anomalies in the computed tomography (CT) scans of the lungs, even in the early stages of infection. A quick and reliable way of detecting COVID-19 is essential to manage the growing transmission of COVID-19 and save lives. Objective: This study focuses on developing a deep learning model that can be used as an auxiliary decision system to detect COVID-19 from chest CT-scan images quickly and effectively. Methods: In this research, we propose a MobileNet-based transfer learning model to detect COVID-19 in CT-scan images. To test the performance of our proposed model, we collect three publicly available COVID-19 CT-scan datasets and prepare another dataset by combining the collected datasets. We also implement a mobile application using the model trained on the combined dataset, which can be used as an auxiliary decision system for COVID-19 screening in real life. Results: Our proposed model achieves a promising accuracy of 96.14% on the combined dataset and accuracy of 98.75%, 98.54%, and 97.84% respectively in detecting COVID-19 samples on the collected datasets. It also outperforms other transfer learning models while having lower memory consumption, ensuring the best performance in both normal and low-powered, resource-constrained devices. Conclusion: We believe, the promising performance of our proposed method will facilitate its use as an auxiliary decision system to detect COVID-19 patients quickly and reliably. This will allow authorities to take immediate measures to limit COVID-19 transmission to prevent further casualties as well as accelerate the screening for COVID-19 while reducing the workload of medical personnel. Keywords: Auxiliary Decision System, COVID-19, CT Scan, Deep Learning, MobileNet, Transfer Learning
背景:COVID-19是一种高度传染性的呼吸系统疾病,具有多种突变变体,在患者中具有渐近性,并且可能在普通检测中未被发现,这使得其更致命,更具传染性,并且更难被发现。无论变异如何,即使在感染的早期阶段,COVID-19感染在肺部计算机断层扫描(CT)中也显示出一些可观察到的异常。快速、可靠的COVID-19检测方法对于控制COVID-19日益增长的传播和拯救生命至关重要。目的:研究开发一种可作为辅助决策系统的深度学习模型,快速有效地从胸部ct扫描图像中检测COVID-19。方法:在本研究中,我们提出了一种基于mobilenet的迁移学习模型来检测ct扫描图像中的COVID-19。为了测试我们提出的模型的性能,我们收集了三个公开可用的COVID-19 ct扫描数据集,并将收集到的数据集组合起来准备另一个数据集。我们还使用组合数据集训练的模型实现了一个移动应用程序,该应用程序可以作为现实生活中COVID-19筛查的辅助决策系统。结果:我们提出的模型在组合数据集上的准确率为96.14%,在收集的数据集上检测COVID-19样本的准确率分别为98.75%,98.54%和97.84%。它还优于其他迁移学习模型,同时具有更低的内存消耗,确保在普通和低功耗,资源受限的设备上都具有最佳性能。结论:我们相信,我们所提出的方法具有良好的性能,将有助于其作为辅助决策系统快速可靠地检测COVID-19患者。这将使当局能够立即采取措施限制COVID-19的传播,以防止进一步的伤亡,并加快COVID-19的筛查,同时减少医务人员的工作量。关键词:辅助决策系统,COVID-19, CT扫描,深度学习,MobileNet,迁移学习
{"title":"A Fast and Reliable Approach for COVID-19 Detection from CT-Scan Images","authors":"Md. Jawwad Bin Zahir, Muhammad Anwarul Azim, Abu Nowshed Chy, Mohammad Khairul Islam","doi":"10.20473/jisebi.9.2.288-304","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.288-304","url":null,"abstract":"Background: COVID-19 is a highly contagious respiratory disease with multiple mutant variants, an asymptotic nature in patients, and with potential to stay undetected in common tests, which makes it deadlier, more transmissible, and harder to detect. Regardless of variants, the COVID-19 infection shows several observable anomalies in the computed tomography (CT) scans of the lungs, even in the early stages of infection. A quick and reliable way of detecting COVID-19 is essential to manage the growing transmission of COVID-19 and save lives. Objective: This study focuses on developing a deep learning model that can be used as an auxiliary decision system to detect COVID-19 from chest CT-scan images quickly and effectively. Methods: In this research, we propose a MobileNet-based transfer learning model to detect COVID-19 in CT-scan images. To test the performance of our proposed model, we collect three publicly available COVID-19 CT-scan datasets and prepare another dataset by combining the collected datasets. We also implement a mobile application using the model trained on the combined dataset, which can be used as an auxiliary decision system for COVID-19 screening in real life. Results: Our proposed model achieves a promising accuracy of 96.14% on the combined dataset and accuracy of 98.75%, 98.54%, and 97.84% respectively in detecting COVID-19 samples on the collected datasets. It also outperforms other transfer learning models while having lower memory consumption, ensuring the best performance in both normal and low-powered, resource-constrained devices. Conclusion: We believe, the promising performance of our proposed method will facilitate its use as an auxiliary decision system to detect COVID-19 patients quickly and reliably. This will allow authorities to take immediate measures to limit COVID-19 transmission to prevent further casualties as well as accelerate the screening for COVID-19 while reducing the workload of medical personnel. Keywords: Auxiliary Decision System, COVID-19, CT Scan, Deep Learning, MobileNet, Transfer Learning","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"100 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transfer Learning based Low Shot Classifier for Software Defect Prediction 基于迁移学习的低概率分类器软件缺陷预测
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.228-238
Vikas Suhag, Sanjay Kumar Dubey, Bhupendra Kumar Sharma
Background: The rapid growth and increasing complexity of software applications are causing challenges in maintaining software quality within constraints of time and resources. This challenge led to the emergence of a new field of study known as Software Defect Prediction (SDP), which focuses on predicting future defect in advance, thereby reducing costs and improving productivity in software industry. Objective: This study aimed to address data distribution disparities when applying transfer learning in multi-project scenarios, and to mitigate performance issues resulting from data scarcity in SDP. Methods: The proposed approach, namely Transfer Learning based Low Shot Classifier (TLLSC), combined transfer learning and low shot learning approaches to create an SDP model. This model was designed for application in both new projects and those with minimal historical defect data. Results: Experiments were conducted using standard datasets from projects within the National Aeronautics and Space Administration (NASA) and Software Research Laboratory (SOFTLAB) repository. TLLSC showed an average increase in F1-Measure of 31.22%, 27.66%, and 27.54% for project AR3, AR4, and AR5, respectively. These results surpassed those from Transfer Component Analysis (TCA+), Canonical Correlation Analysis (CCA+), and Kernel Canonical Correlation Analysis plus (KCCA+). Conclusion: The results of the comparison between TLLSC and state-of-the-art algorithms, namely TCA+, CCA+, and KCCA+ from the existing literature consistently showed that TLLSC performed better in terms of F1-Measure. Keywords: Just-in-time, Defect Prediction, Deep Learning, Transfer Learning, Low Shot Learning
背景:软件应用程序的快速增长和日益增加的复杂性给在时间和资源的限制下保持软件质量带来了挑战。这一挑战导致了一个新的研究领域的出现,即软件缺陷预测(SDP),它关注于提前预测未来的缺陷,从而降低软件行业的成本并提高生产率。目的:本研究旨在解决迁移学习在多项目场景下的数据分布差异,并缓解迁移学习中由于数据稀缺而导致的性能问题。方法:提出基于迁移学习的Low Shot Classifier (TLLSC)方法,将迁移学习和Low Shot学习方法相结合,建立SDP模型。这个模型被设计用于新项目和那些具有最小历史缺陷数据的项目。结果:实验使用来自美国国家航空航天局(NASA)和软件研究实验室(SOFTLAB)存储库项目的标准数据集进行。TLLSC在AR3、AR4和AR5项目中F1-Measure的平均增幅分别为31.22%、27.66%和27.54%。这些结果优于传递成分分析(TCA+)、典型相关分析(CCA+)和核典型相关分析+ (KCCA+)。结论:TLLSC与现有文献中最先进的TCA+、CCA+、KCCA+算法的比较结果一致表明,TLLSC在F1-Measure方面表现更好。关键词:准时制,缺陷预测,深度学习,迁移学习,低概率学习
{"title":"Transfer Learning based Low Shot Classifier for Software Defect Prediction","authors":"Vikas Suhag, Sanjay Kumar Dubey, Bhupendra Kumar Sharma","doi":"10.20473/jisebi.9.2.228-238","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.228-238","url":null,"abstract":"Background: The rapid growth and increasing complexity of software applications are causing challenges in maintaining software quality within constraints of time and resources. This challenge led to the emergence of a new field of study known as Software Defect Prediction (SDP), which focuses on predicting future defect in advance, thereby reducing costs and improving productivity in software industry. Objective: This study aimed to address data distribution disparities when applying transfer learning in multi-project scenarios, and to mitigate performance issues resulting from data scarcity in SDP. Methods: The proposed approach, namely Transfer Learning based Low Shot Classifier (TLLSC), combined transfer learning and low shot learning approaches to create an SDP model. This model was designed for application in both new projects and those with minimal historical defect data. Results: Experiments were conducted using standard datasets from projects within the National Aeronautics and Space Administration (NASA) and Software Research Laboratory (SOFTLAB) repository. TLLSC showed an average increase in F1-Measure of 31.22%, 27.66%, and 27.54% for project AR3, AR4, and AR5, respectively. These results surpassed those from Transfer Component Analysis (TCA+), Canonical Correlation Analysis (CCA+), and Kernel Canonical Correlation Analysis plus (KCCA+). Conclusion: The results of the comparison between TLLSC and state-of-the-art algorithms, namely TCA+, CCA+, and KCCA+ from the existing literature consistently showed that TLLSC performed better in terms of F1-Measure. Keywords: Just-in-time, Defect Prediction, Deep Learning, Transfer Learning, Low Shot Learning","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"99 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fine-Tuning IndoBERT for Indonesian Exam Question Classification Based on Bloom's Taxonomy 基于Bloom分类法的印尼语试题分类微调IndoBERT
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.253-263
Fikri Baharuddin, Mohammad Farid Naufal
Background: The learning assessment of elementary schools has recently incorporated Bloom's Taxonomy, a structure in education that categorizes different levels of cognitive learning and thinking skills, as a fundamental framework. This assessment now includes High Order Thinking Skill (HOTS) questions, with a specific focus on Indonesian topics. The implementation of this system has been observed to require teachers to manually categorize or classify questions, and this process typically requires more time and resources. To address the associated difficulty, automated categorization and classification are required to streamline the process. However, despite various research efforts in questions classification, there is still room for improvement in terms of performance, particularly in precision and accuracy. Numerous investigations have explored the use of Deep Learning Natural Language Processing models such as BERT for classification, and IndoBERT is one such pre-trained model for text analysis. Objective: This research aims to build classification system that is capable of classifying Indonesian exam questions in multiple-choice form based on Bloom's Taxonomy using IndoBERT pre-trained model. Methods: The methodology used includes hyperparameter fine-tuning, which was carried out to identify the optimal model performance. This performance was subsequently evaluated based on accuracy, F1 Score, Precision, Recall, and the time required for the training and validation of the model. Results: The proposed Fine Tuned IndoBERT Model showed that the accuracy rate was 97%, 97% F1 Score, 97% Recall, and 98% Precision with an average training time per epoch of 1.55 seconds and an average validation time per epoch of 0.38 seconds. Conclusion: Fine Tuned IndoBERT model was observed to have a relatively high classification performance, and based on this observation, the system was considered capable of classifying Indonesian exam questions at the elementary school level. Keywords: IndoBERT, Fine Tuning, Indonesian Exam Question, Model Classifier, Natural Language Processing, Bloom’s Taxonomy
背景:最近,小学学习评估纳入了布鲁姆分类法(Bloom’s Taxonomy),这是一种将不同层次的认知学习和思维技能分类的教育结构,作为基本框架。该评估现在包括高阶思维技能(HOTS)问题,特别侧重于印度尼西亚主题。据观察,该系统的实施需要教师手动对问题进行分类或分类,这一过程通常需要更多的时间和资源。为了解决相关的困难,需要自动分类和分类来简化流程。然而,尽管在问题分类方面进行了各种各样的研究,但在性能方面,特别是在精密度和准确性方面,仍有提高的空间。许多研究已经探索了使用深度学习自然语言处理模型(如BERT)进行分类,而IndoBERT就是这样一个用于文本分析的预训练模型。目的:利用IndoBERT预训练模型,构建基于Bloom分类法的印尼语选择题分类系统。方法:采用的方法包括超参数微调,以确定最优的模型性能。该性能随后根据准确性、F1分数、精度、召回率以及模型训练和验证所需的时间进行评估。结果:提出的微调IndoBERT模型准确率为97%,F1 Score为97%,Recall为97%,Precision为98%,平均训练时间为1.55秒,平均验证时间为0.38秒。结论:观察到Fine Tuned IndoBERT模型具有较高的分类性能,基于这一观察,认为该系统能够对小学水平的印尼语试题进行分类。关键词:IndoBERT,微调,印尼语试题,模型分类器,自然语言处理,Bloom分类法
{"title":"Fine-Tuning IndoBERT for Indonesian Exam Question Classification Based on Bloom's Taxonomy","authors":"Fikri Baharuddin, Mohammad Farid Naufal","doi":"10.20473/jisebi.9.2.253-263","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.253-263","url":null,"abstract":"Background: The learning assessment of elementary schools has recently incorporated Bloom's Taxonomy, a structure in education that categorizes different levels of cognitive learning and thinking skills, as a fundamental framework. This assessment now includes High Order Thinking Skill (HOTS) questions, with a specific focus on Indonesian topics. The implementation of this system has been observed to require teachers to manually categorize or classify questions, and this process typically requires more time and resources. To address the associated difficulty, automated categorization and classification are required to streamline the process. However, despite various research efforts in questions classification, there is still room for improvement in terms of performance, particularly in precision and accuracy. Numerous investigations have explored the use of Deep Learning Natural Language Processing models such as BERT for classification, and IndoBERT is one such pre-trained model for text analysis. Objective: This research aims to build classification system that is capable of classifying Indonesian exam questions in multiple-choice form based on Bloom's Taxonomy using IndoBERT pre-trained model. Methods: The methodology used includes hyperparameter fine-tuning, which was carried out to identify the optimal model performance. This performance was subsequently evaluated based on accuracy, F1 Score, Precision, Recall, and the time required for the training and validation of the model. Results: The proposed Fine Tuned IndoBERT Model showed that the accuracy rate was 97%, 97% F1 Score, 97% Recall, and 98% Precision with an average training time per epoch of 1.55 seconds and an average validation time per epoch of 0.38 seconds. Conclusion: Fine Tuned IndoBERT model was observed to have a relatively high classification performance, and based on this observation, the system was considered capable of classifying Indonesian exam questions at the elementary school level. Keywords: IndoBERT, Fine Tuning, Indonesian Exam Question, Model Classifier, Natural Language Processing, Bloom’s Taxonomy","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"100 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancement in Bangla Sentiment Analysis: A Comparative Study of Transformer-Based and Transfer Learning Models for E-commerce Sentiment Classification 孟加拉语情感分析的研究进展:基于变压器和迁移学习的电子商务情感分类模型的比较研究
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.181-194
Zishan Ahmed, Shakib Sadat Shanto, Akinul Islam Jony
Background: As a direct result of the Internet's expansion, the quantity of information shared by Internet users across its numerous platforms has increased. Sentiment analysis functions at a higher level when there are more available perspectives and opinions. However, the lack of labeled data significantly complicates sentiment analysis utilizing Bangla natural language processing (NLP). In recent years, nevertheless, due to the development of more effective deep learning models, Bangla sentiment analysis has improved significantly. Objective: This article presents a curated dataset for Bangla e-commerce sentiment analysis obtained solely from the "Daraz" platform. We aim to conduct sentiment analysis in Bangla for binary and understudied multiclass classification tasks. Methods: Transfer learning (LSTM, GRU) and Transformers (Bangla-BERT) approaches are compared for their effectiveness on our dataset. To enhance the overall performance of the models, we fine-tuned them. Results: The accuracy of Bangla-BERT was highest for both binary and multiclass sentiment classification tasks, with 94.5% accuracy for binary classification and 88.78% accuracy for multiclass sentiment classification. Conclusion: Our proposed method performs noticeably better classifying multiclass sentiments in Bangla than previous deep learning techniques. Keywords: Bangla-BERT, Deep Learning, E-commerce, NLP, Sentiment Analysis
背景:互联网发展的直接结果是,互联网用户在其众多平台上共享的信息量增加了。当有更多可用的观点和意见时,情感分析在更高层次上起作用。然而,缺乏标记数据使使用孟加拉语自然语言处理(NLP)的情感分析变得非常复杂。然而,近年来,由于更有效的深度学习模型的发展,孟加拉语情感分析有了显着改善。目的:本文提供了一个仅从“Daraz”平台获得的用于孟加拉电子商务情感分析的精选数据集。我们的目标是在孟加拉语中对二元和未充分研究的多类分类任务进行情感分析。方法:在我们的数据集上比较迁移学习(LSTM, GRU)和变形金刚(Bangla-BERT)方法的有效性。为了提高模型的整体性能,我们对它们进行了微调。结果:Bangla-BERT在二元和多类情感分类任务中准确率最高,二元分类准确率为94.5%,多类情感分类准确率为88.78%。结论:与之前的深度学习技术相比,我们提出的方法在分类孟加拉语的多类情绪方面表现明显更好。关键词:Bangla-BERT,深度学习,电子商务,NLP,情感分析
{"title":"Advancement in Bangla Sentiment Analysis: A Comparative Study of Transformer-Based and Transfer Learning Models for E-commerce Sentiment Classification","authors":"Zishan Ahmed, Shakib Sadat Shanto, Akinul Islam Jony","doi":"10.20473/jisebi.9.2.181-194","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.181-194","url":null,"abstract":"Background: As a direct result of the Internet's expansion, the quantity of information shared by Internet users across its numerous platforms has increased. Sentiment analysis functions at a higher level when there are more available perspectives and opinions. However, the lack of labeled data significantly complicates sentiment analysis utilizing Bangla natural language processing (NLP). In recent years, nevertheless, due to the development of more effective deep learning models, Bangla sentiment analysis has improved significantly. Objective: This article presents a curated dataset for Bangla e-commerce sentiment analysis obtained solely from the \"Daraz\" platform. We aim to conduct sentiment analysis in Bangla for binary and understudied multiclass classification tasks. Methods: Transfer learning (LSTM, GRU) and Transformers (Bangla-BERT) approaches are compared for their effectiveness on our dataset. To enhance the overall performance of the models, we fine-tuned them. Results: The accuracy of Bangla-BERT was highest for both binary and multiclass sentiment classification tasks, with 94.5% accuracy for binary classification and 88.78% accuracy for multiclass sentiment classification. Conclusion: Our proposed method performs noticeably better classifying multiclass sentiments in Bangla than previous deep learning techniques. Keywords: Bangla-BERT, Deep Learning, E-commerce, NLP, Sentiment Analysis","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"96 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Crypto-sentiment Detection in Malay Text Using Language Models with an Attention Mechanism 基于注意机制的马来语文本隐情感检测
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.147-160
Nur Azmina Mohamad Zamani, Norhaslinda Kamaruddin
Background: Due to the increased interest in cryptocurrencies, opinions on cryptocurrency-related topics are shared on news and social media. The enormous amount of sentiment data that is frequently released makes data processing and analytics on such important issues more challenging. In addition, the present sentiment models in the cryptocurrency domain are primarily focused on English with minimal work on Malay language, further complicating problems. Objective: The performance of the sentiment regression model to forecast sentiment scores for Malay news and tweets is examined in this study. Methods: Malay news headlines and tweets on Bitcoin and Ethereum are used as the input. A hybrid Generalized Autoregressive Pretraining for Language Understanding (XLNet) language model in combination with Bidirectional-Gated Recurrent Unit (Bi-GRU) deep learning model is applied in the proposed sentiment regression implementation. The effectiveness of the proposed sentiment regression model is also investigated using the multi-head self-attention mechanism. Then, a comparison analysis using Bidirectional Encoder Representations from Transformers (BERT) is carried out. Results: The experimental results demonstrate that the number of attention heads is vital in improving the XLNet-GRU sentiment model performance. There are slight improvements of 0.03 in the adjusted R2 values with an average MAE of 0.163 (Malay news) and 0.174 (Malay tweets). In addition, an average RMSE of 0.267 and 0.255 were obtained respectively for Malay news and tweets, which show that the proposed XLNet-GRU sentiment model outperforms the BERT sentiment model with lesser prediction errors. Conclusion: The proposed model contributes to predicting sentiment on cryptocurrency. Moreover, this study also introduced two carefully curated Malay corpora, CryptoSentiNews-Malay and CryptoSentiTweets-Malay, which are extracted from news and tweets, respectively. Further works to enhance Malay news and tweets corpora on cryptocurrency-related issues will be expended with implementing the proposed XLNet Bi-GRU deep learning model for greater financial insight. Keywords: Cryptocurrency, Deep learning model, Malay text, Sentiment analysis, Sentiment regression model
背景:由于对加密货币的兴趣增加,新闻和社交媒体上分享了对加密货币相关话题的看法。频繁发布的大量情绪数据使得这些重要问题的数据处理和分析更具挑战性。此外,目前加密货币领域的情感模型主要集中在英语上,对马来语的研究很少,这使问题进一步复杂化。目的:本研究检验了马来语新闻和推文的情绪回归模型预测情绪得分的性能。方法:使用马来语新闻标题和比特币和以太坊上的推文作为输入。提出了一种结合双向门控循环单元(Bi-GRU)深度学习模型的混合广义自回归语言理解预训练(XLNet)语言模型。采用多头自注意机制考察了情绪回归模型的有效性。然后,利用变压器双向编码器表示(BERT)进行对比分析。结果:实验结果表明,注意头的数量对提高XLNet-GRU情感模型的性能至关重要。调整后的R2值略有改善0.03,平均MAE为0.163(马来新闻)和0.174(马来推文)。此外,马来语新闻和推文的平均RMSE分别为0.267和0.255,这表明所提出的XLNet-GRU情绪模型优于BERT情绪模型,预测误差较小。结论:提出的模型有助于预测加密货币的情绪。此外,本研究还引入了两个精心策划的马来语语料库,CryptoSentiNews-Malay和CryptoSentiTweets-Malay,分别从新闻和推文中提取。通过实施拟议的XLNet Bi-GRU深度学习模型,进一步加强与加密货币相关问题的马来语新闻和推文语料库,以获得更大的金融洞察力。关键词:加密货币,深度学习模型,马来语文本,情感分析,情感回归模型
{"title":"Crypto-sentiment Detection in Malay Text Using Language Models with an Attention Mechanism","authors":"Nur Azmina Mohamad Zamani, Norhaslinda Kamaruddin","doi":"10.20473/jisebi.9.2.147-160","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.147-160","url":null,"abstract":"Background: Due to the increased interest in cryptocurrencies, opinions on cryptocurrency-related topics are shared on news and social media. The enormous amount of sentiment data that is frequently released makes data processing and analytics on such important issues more challenging. In addition, the present sentiment models in the cryptocurrency domain are primarily focused on English with minimal work on Malay language, further complicating problems. Objective: The performance of the sentiment regression model to forecast sentiment scores for Malay news and tweets is examined in this study. Methods: Malay news headlines and tweets on Bitcoin and Ethereum are used as the input. A hybrid Generalized Autoregressive Pretraining for Language Understanding (XLNet) language model in combination with Bidirectional-Gated Recurrent Unit (Bi-GRU) deep learning model is applied in the proposed sentiment regression implementation. The effectiveness of the proposed sentiment regression model is also investigated using the multi-head self-attention mechanism. Then, a comparison analysis using Bidirectional Encoder Representations from Transformers (BERT) is carried out. Results: The experimental results demonstrate that the number of attention heads is vital in improving the XLNet-GRU sentiment model performance. There are slight improvements of 0.03 in the adjusted R2 values with an average MAE of 0.163 (Malay news) and 0.174 (Malay tweets). In addition, an average RMSE of 0.267 and 0.255 were obtained respectively for Malay news and tweets, which show that the proposed XLNet-GRU sentiment model outperforms the BERT sentiment model with lesser prediction errors. Conclusion: The proposed model contributes to predicting sentiment on cryptocurrency. Moreover, this study also introduced two carefully curated Malay corpora, CryptoSentiNews-Malay and CryptoSentiTweets-Malay, which are extracted from news and tweets, respectively. Further works to enhance Malay news and tweets corpora on cryptocurrency-related issues will be expended with implementing the proposed XLNet Bi-GRU deep learning model for greater financial insight. Keywords: Cryptocurrency, Deep learning model, Malay text, Sentiment analysis, Sentiment regression model","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"99 1-2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Use of Machine Learning to Detect Financial Transaction Fraud: Multiple Benford Law Model for Auditors 使用机器学习检测金融交易欺诈:审计师的多重本福德定律模型
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.239-252
Doni Wiryadinata, Aris Sugiharto, Tarno Tarno
Background: Fraud in financial transaction is at the root of corruption issues recorded in organization. Detecting fraud practices has become increasingly complex and challenging. As a result, auditors require precise analytical tools for fraud detection. Grouping financial transaction data using K-Means Clustering algorithm can enhance the efficiency of applying Benford Law for optimal fraud detection. Objective: This study aimed to introduce Multiple Benford Law Model for the analysis of data to show potential concealed fraud in the audited organization financial transaction. The data was categorized into low, medium, and high transaction values using K-Means Clustering algorithm. Subsequently, it was reanalyzed through Multiple Benford Law Model in a specialized fraud analysis tool. Methods: In this study, the experimental procedures of Multiple Benford Law Model designed for public sector organizations were applied. The analysis of suspected fraud generated by the toolkit was compared with the actual conditions reported in audit report. The financial transaction dataset was prepared and grouped into three distinct clusters using the Euclidean distance equation. Data in these clusters was analyzed using Benford Law, comparing the frequency of the first digit’s occurrence to the expected frequency based on Benford Law. Significant deviations exceeding ±5% were considered potential areas for further scrutiny in audit. Furthermore, the analysis were validated by cross-referencing the result with the findings presented in the authorized audit organization report. Results: Multiple Benford Law Model developed was incorporated into an audit toolkit to automated calculations based on Benford Law. Furthermore, the datasets were categorized using K-Means Clustering algorithm into three clusters representing low, medium, and high-value transaction data. Results from the application of Benford Law showed a 40.00% potential for fraud detection. However, when using Multiple Benford Law Model and dividing the data into three clusters, fraud detection accuracy increased to 93.33%. The comparative results in audit report indicated a 75.00% consistency with the actual events or facts discovered. Conclusion: The use of Multiple Benford Law Model in audit toolkit substantially improved the accuracy of detecting potential fraud in financial transaction. Validation through audit report showed the conformity between the identified fraud practices and the detected financial transaction. Keywords: Fraud Detection, Benford’s Law, K-Means Clustering, Audit Toolkit, Fraudulent Practices.
背景:金融交易中的欺诈行为是组织腐败问题的根源。检测欺诈行为变得越来越复杂和具有挑战性。因此,审计人员需要精确的分析工具来检测欺诈。利用K-Means聚类算法对金融交易数据进行分组,可以提高应用Benford定律进行最优欺诈检测的效率。目的:本研究旨在引入多重本福德定律模型对数据进行分析,揭示被审计单位财务交易中潜在的隐性欺诈。使用K-Means聚类算法将数据分为低、中、高交易值。随后,在专门的欺诈分析工具中,通过多重本福德定律模型对其进行重新分析。方法:本研究采用针对公共部门组织设计的多重本福德定律模型的实验程序。将工具包生成的可疑欺诈分析与审计报告中报告的实际情况进行比较。利用欧几里得距离方程制备了金融交易数据集,并将其分为三个不同的簇。这些聚类中的数据使用本福德定律进行分析,将第一个数字出现的频率与基于本福德定律的预期频率进行比较。超过±5%的显著偏差被认为是审计中进一步审查的潜在领域。此外,通过将分析结果与授权审计组织报告中提出的调查结果进行交叉对照,验证了分析的有效性。结果:开发的多个本福德定律模型被纳入审计工具包,以基于本福德定律的自动计算。此外,使用K-Means聚类算法将数据集分类为代表低、中、高价值交易数据的三个聚类。应用本福德定律的结果显示,欺诈检测的可能性为40.00%。然而,当使用多重本福德定律模型并将数据分为三类时,欺诈检测准确率提高到93.33%。审计报告的对比结果与发现的实际事件或事实的一致性为75.00%。结论:在审计工具包中使用多重本福德定律模型大大提高了发现金融交易中潜在欺诈的准确性。通过审计报告确认已发现的舞弊行为与已发现的财务交易之间的一致性。关键词:欺诈检测,本福德定律,k均值聚类,审计工具包,欺诈行为
{"title":"The Use of Machine Learning to Detect Financial Transaction Fraud: Multiple Benford Law Model for Auditors","authors":"Doni Wiryadinata, Aris Sugiharto, Tarno Tarno","doi":"10.20473/jisebi.9.2.239-252","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.239-252","url":null,"abstract":"Background: Fraud in financial transaction is at the root of corruption issues recorded in organization. Detecting fraud practices has become increasingly complex and challenging. As a result, auditors require precise analytical tools for fraud detection. Grouping financial transaction data using K-Means Clustering algorithm can enhance the efficiency of applying Benford Law for optimal fraud detection. Objective: This study aimed to introduce Multiple Benford Law Model for the analysis of data to show potential concealed fraud in the audited organization financial transaction. The data was categorized into low, medium, and high transaction values using K-Means Clustering algorithm. Subsequently, it was reanalyzed through Multiple Benford Law Model in a specialized fraud analysis tool. Methods: In this study, the experimental procedures of Multiple Benford Law Model designed for public sector organizations were applied. The analysis of suspected fraud generated by the toolkit was compared with the actual conditions reported in audit report. The financial transaction dataset was prepared and grouped into three distinct clusters using the Euclidean distance equation. Data in these clusters was analyzed using Benford Law, comparing the frequency of the first digit’s occurrence to the expected frequency based on Benford Law. Significant deviations exceeding ±5% were considered potential areas for further scrutiny in audit. Furthermore, the analysis were validated by cross-referencing the result with the findings presented in the authorized audit organization report. Results: Multiple Benford Law Model developed was incorporated into an audit toolkit to automated calculations based on Benford Law. Furthermore, the datasets were categorized using K-Means Clustering algorithm into three clusters representing low, medium, and high-value transaction data. Results from the application of Benford Law showed a 40.00% potential for fraud detection. However, when using Multiple Benford Law Model and dividing the data into three clusters, fraud detection accuracy increased to 93.33%. The comparative results in audit report indicated a 75.00% consistency with the actual events or facts discovered. Conclusion: The use of Multiple Benford Law Model in audit toolkit substantially improved the accuracy of detecting potential fraud in financial transaction. Validation through audit report showed the conformity between the identified fraud practices and the detected financial transaction. Keywords: Fraud Detection, Benford’s Law, K-Means Clustering, Audit Toolkit, Fraudulent Practices.","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"98 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Systematic Literature and Expert Review of Agile Methodology Usage in Business Intelligence Projects 敏捷方法在商业智能项目中的应用的系统文献和专家评论
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.214-227
Hapsari Wulandari, Teguh Raharjo
Background: Agile methodology is known for delivering effective projects with added value within a shorter timeframe, especially in Business Intelligence (BI) system which is a valuable tool for informed decision-making. However, identifying impactful elements for successful BI implementation is complex due to the wide range of Agile attributes. Objective: This research aims to systematically review and analyze the integration of BI within Agile methodology, providing valuable guidance for future projects implementation, enhancing the understanding of effective application, and identifying influential factors. Methods: Based on the Kitchenham method, 19 papers were analyzed from 288 papers, sourced from databases like Scopus, ACM, IEEE, and others published in 2016-2022. Meanwhile the extracted key factors impacting agile BI implementation were validated by qualified expert. Results: Agile was discovered to provide numerous benefits to BI projects by promoting flexibility, collaboration, and rapid iteration for enhanced adaptability, while effectively addressing challenges including those related to technology, management, and skills gaps. In addition, Agile methods, including tasks such as calculating cycle time, measuring defect backlogs, mapping code ownership, and engaging end users, offered practical solutions. The advantages included adaptability, success, value enhancement, cost reduction, shortened timelines, and improved precision. The research additionally considered other critical Agile elements such as BI tools, Agile Practices, Manifesto, and Methods, thereby enhancing insights for successful implementation. Conclusion: In conclusion, the research outlined Agile BI implementation into seven key factor groups, validated by qualified expert, providing guidance for BI integration and practices, and establishing a fundamental baseline for future applications. Keywords: Agile Methodology, Business Intelligence (BI), Expert Judgement, Kitchenham, Systematic Literature Review (SLR)
背景:敏捷方法以在较短的时间框架内交付具有附加价值的有效项目而闻名,特别是在商业智能(BI)系统中,这是一种有价值的明智决策工具。然而,由于敏捷属性的广泛范围,确定成功的BI实现的影响因素是复杂的。目的:本研究旨在系统回顾和分析敏捷方法论中BI的集成,为未来项目实施提供有价值的指导,增强对有效应用的理解,并识别影响因素。方法:基于Kitchenham方法,从2016-2022年发表的288篇论文中选取19篇进行分析,这些论文来源于Scopus、ACM、IEEE等数据库。同时,对提取的影响敏捷BI实施的关键因素进行了专家验证。结果:敏捷被发现通过促进灵活性、协作和快速迭代来增强适应性,同时有效地解决包括与技术、管理和技能差距相关的挑战,为BI项目提供了许多好处。此外,敏捷方法,包括诸如计算周期时间、度量缺陷积压、映射代码所有权和吸引最终用户等任务,提供了实用的解决方案。优点包括适应性、成功、价值提升、成本降低、缩短时间和提高精度。该研究还考虑了其他关键的敏捷元素,如BI工具、敏捷实践、宣言和方法,从而增强了对成功实现的见解。结论:最后,本研究将敏捷BI实现概述为七个关键因素组,并由合格的专家进行验证,为BI集成和实践提供指导,并为未来的应用建立基本基线。关键词:敏捷方法论,商业智能,专家判断,Kitchenham,系统文献综述(SLR)
{"title":"Systematic Literature and Expert Review of Agile Methodology Usage in Business Intelligence Projects","authors":"Hapsari Wulandari, Teguh Raharjo","doi":"10.20473/jisebi.9.2.214-227","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.214-227","url":null,"abstract":"Background: Agile methodology is known for delivering effective projects with added value within a shorter timeframe, especially in Business Intelligence (BI) system which is a valuable tool for informed decision-making. However, identifying impactful elements for successful BI implementation is complex due to the wide range of Agile attributes. Objective: This research aims to systematically review and analyze the integration of BI within Agile methodology, providing valuable guidance for future projects implementation, enhancing the understanding of effective application, and identifying influential factors. Methods: Based on the Kitchenham method, 19 papers were analyzed from 288 papers, sourced from databases like Scopus, ACM, IEEE, and others published in 2016-2022. Meanwhile the extracted key factors impacting agile BI implementation were validated by qualified expert. Results: Agile was discovered to provide numerous benefits to BI projects by promoting flexibility, collaboration, and rapid iteration for enhanced adaptability, while effectively addressing challenges including those related to technology, management, and skills gaps. In addition, Agile methods, including tasks such as calculating cycle time, measuring defect backlogs, mapping code ownership, and engaging end users, offered practical solutions. The advantages included adaptability, success, value enhancement, cost reduction, shortened timelines, and improved precision. The research additionally considered other critical Agile elements such as BI tools, Agile Practices, Manifesto, and Methods, thereby enhancing insights for successful implementation. Conclusion: In conclusion, the research outlined Agile BI implementation into seven key factor groups, validated by qualified expert, providing guidance for BI integration and practices, and establishing a fundamental baseline for future applications. Keywords: Agile Methodology, Business Intelligence (BI), Expert Judgement, Kitchenham, Systematic Literature Review (SLR)","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"100 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Multi-Output Time Series Forecasting with Encoder-Decoder Networks 用编码器-解码器网络增强多输出时间序列预测
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.195-213
Kristoko Dwi Hartomo, Joanito Agili Lopo, Hindriyanto Dwi Purnomo
Background: Multi-output Time series forecasting is a complex problem that requires handling interdependencies and interactions between variables. Traditional statistical approaches and machine learning techniques often struggle to predict such scenarios accurately. Advanced techniques and model reconstruction are necessary to improve forecasting accuracy in complex scenarios. Objective: This study proposed an Encoder-Decoder network to address multi-output time series forecasting challenges by simultaneously predicting each output. This objective is to investigate the capabilities of the Encoder-Decoder architecture in handling multi-output time series forecasting tasks. Methods: This proposed model utilizes a 1-Dimensional Convolution Neural Network with Bidirectional Long Short-Term Memory, specifically in the encoder part. The encoder extracts time series features, incorporating a residual connection to produce a context representation used by the decoder. The decoder employs multiple unidirectional LSTM modules and Linear transformation layers to generate the outputs each time step. Each module is responsible for specific output and shares information and context along the outputs and steps. Results: The result demonstrates that the proposed model achieves lower error rates, as measured by MSE, RMSE, and MAE loss metrics, for all outputs and forecasting horizons. Notably, the 6-hour horizon achieves the highest accuracy across all outputs. Furthermore, the proposed model exhibits robustness in single-output forecast and transfer learning, showing adaptability to different tasks and datasets.   Conclusion: The experiment findings highlight the successful multi-output forecasting capabilities of the proposed model in time series data, with consistently low error rates (MSE, RMSE, MAE). Surprisingly, the model also performs well in single-output forecasts, demonstrating its versatility. Therefore, the proposed model effectively various time series forecasting tasks, showing promise for practical applications. Keywords: Bidirectional Long Short-Term Memory, Convolutional Neural Network, Encoder-Decoder Networks, Multi-output forecasting, Multi-step forecasting, Time-series forecasting
背景:多输出时间序列预测是一个复杂的问题,需要处理变量之间的相互依赖和相互作用。传统的统计方法和机器学习技术往往难以准确预测这种情况。提高复杂情景下的预测精度需要先进的技术和模型重建。目的:提出了一种编码器-解码器网络,通过同时预测每个输出来解决多输出时间序列预测的挑战。这个目标是研究编码器-解码器架构在处理多输出时间序列预测任务中的能力。方法:该模型采用具有双向长短期记忆的一维卷积神经网络,特别是在编码器部分。编码器提取时间序列特征,结合残差连接产生解码器使用的上下文表示。解码器采用多个单向LSTM模块和线性变换层来产生每个时间步长的输出。每个模块负责特定的输出,并沿着输出和步骤共享信息和上下文。结果表明,通过MSE、RMSE和MAE损失度量,对于所有输出和预测范围,所提出的模型实现了较低的错误率。值得注意的是,6小时视界在所有输出中达到了最高的精度。此外,该模型在单输出预测和迁移学习中表现出鲁棒性,显示出对不同任务和数据集的适应性。结论:实验结果突出了该模型在时间序列数据中具有成功的多输出预测能力,并且错误率(MSE, RMSE, MAE)始终较低。令人惊讶的是,该模型在单输出预测中也表现良好,证明了它的多功能性。因此,该模型能有效地完成各种时间序列预测任务,具有实际应用前景。关键词:双向长短期记忆,卷积神经网络,编码器-解码器网络,多输出预测,多步预测,时间序列预测
{"title":"Enhancing Multi-Output Time Series Forecasting with Encoder-Decoder Networks","authors":"Kristoko Dwi Hartomo, Joanito Agili Lopo, Hindriyanto Dwi Purnomo","doi":"10.20473/jisebi.9.2.195-213","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.195-213","url":null,"abstract":"Background: Multi-output Time series forecasting is a complex problem that requires handling interdependencies and interactions between variables. Traditional statistical approaches and machine learning techniques often struggle to predict such scenarios accurately. Advanced techniques and model reconstruction are necessary to improve forecasting accuracy in complex scenarios. Objective: This study proposed an Encoder-Decoder network to address multi-output time series forecasting challenges by simultaneously predicting each output. This objective is to investigate the capabilities of the Encoder-Decoder architecture in handling multi-output time series forecasting tasks. Methods: This proposed model utilizes a 1-Dimensional Convolution Neural Network with Bidirectional Long Short-Term Memory, specifically in the encoder part. The encoder extracts time series features, incorporating a residual connection to produce a context representation used by the decoder. The decoder employs multiple unidirectional LSTM modules and Linear transformation layers to generate the outputs each time step. Each module is responsible for specific output and shares information and context along the outputs and steps. Results: The result demonstrates that the proposed model achieves lower error rates, as measured by MSE, RMSE, and MAE loss metrics, for all outputs and forecasting horizons. Notably, the 6-hour horizon achieves the highest accuracy across all outputs. Furthermore, the proposed model exhibits robustness in single-output forecast and transfer learning, showing adaptability to different tasks and datasets.   Conclusion: The experiment findings highlight the successful multi-output forecasting capabilities of the proposed model in time series data, with consistently low error rates (MSE, RMSE, MAE). Surprisingly, the model also performs well in single-output forecasts, demonstrating its versatility. Therefore, the proposed model effectively various time series forecasting tasks, showing promise for practical applications. Keywords: Bidirectional Long Short-Term Memory, Convolutional Neural Network, Encoder-Decoder Networks, Multi-output forecasting, Multi-step forecasting, Time-series forecasting","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"97 1-4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Information Quality of Business Intelligence Systems: A Maturity-based Assessment 商业智能系统的信息质量:基于成熟度的评估
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.276-287
Abdelhak Ait Touil, Siham Jabraoui
Background: The primary role of a Business Intelligence (BI) system is to provide information to decision-makers within an organization. Moreover, it is crucial to acknowledge that the quality of this information is of greatest significance. Several studies have extensively discussed the importance of information quality in information systems, including BI. However, there is relatively little discussion on the factors influencing 'Information quality”. Objective: This study aimed to address this literature gap by investigating the determinants of BI maturity that impacted information quality. Methods: A maturity model comprising three dimensions was introduced, namely Data quality, BI infrastructure, and Data-driven culture. Data were collected from 84 companies and were analyzed using the SEM-PLS approach. Results: The analysis showed that maturity had a highly positive influence on Information Quality, validating the relevance of the three proposed determinant factors. Conclusion: This study suggested and strongly supported the importance and relevance of Data quality, BI infrastructure, and Data-driven culture as key dimensions of BI maturity. The robust statistical relationship between maturity and information quality showed the effectiveness of approaching the systems from a maturity perspective. This investigation paved the way for exploring additional dimensions that impact Information quality. Keywords: BI infrastructure, BI maturity, Data-driven culture, Data quality, Information quality.
背景:商业智能(BI)系统的主要作用是向组织内的决策者提供信息。此外,至关重要的是要承认这些信息的质量是非常重要的。一些研究广泛地讨论了信息系统(包括BI)中信息质量的重要性。然而,对“信息质量”影响因素的探讨相对较少。目的:本研究旨在通过调查影响信息质量的商业智能成熟度的决定因素来解决这一文献空白。方法:引入了一个包含三个维度的成熟度模型,即数据质量、BI基础设施和数据驱动文化。从84家公司收集数据,并使用SEM-PLS方法进行分析。结果:分析表明,成熟度对信息质量有高度的正向影响,验证了三个决定因素的相关性。结论:本研究建议并强烈支持数据质量、商业智能基础设施和数据驱动文化作为商业智能成熟度的关键维度的重要性和相关性。成熟度和信息质量之间稳健的统计关系显示了从成熟度角度接近系统的有效性。这项调查为探索影响信息质量的其他维度铺平了道路。关键词:BI基础设施,BI成熟度,数据驱动文化,数据质量,信息质量
{"title":"Information Quality of Business Intelligence Systems: A Maturity-based Assessment","authors":"Abdelhak Ait Touil, Siham Jabraoui","doi":"10.20473/jisebi.9.2.276-287","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.276-287","url":null,"abstract":"Background: The primary role of a Business Intelligence (BI) system is to provide information to decision-makers within an organization. Moreover, it is crucial to acknowledge that the quality of this information is of greatest significance. Several studies have extensively discussed the importance of information quality in information systems, including BI. However, there is relatively little discussion on the factors influencing 'Information quality”. Objective: This study aimed to address this literature gap by investigating the determinants of BI maturity that impacted information quality. Methods: A maturity model comprising three dimensions was introduced, namely Data quality, BI infrastructure, and Data-driven culture. Data were collected from 84 companies and were analyzed using the SEM-PLS approach. Results: The analysis showed that maturity had a highly positive influence on Information Quality, validating the relevance of the three proposed determinant factors. Conclusion: This study suggested and strongly supported the importance and relevance of Data quality, BI infrastructure, and Data-driven culture as key dimensions of BI maturity. The robust statistical relationship between maturity and information quality showed the effectiveness of approaching the systems from a maturity perspective. This investigation paved the way for exploring additional dimensions that impact Information quality. Keywords: BI infrastructure, BI maturity, Data-driven culture, Data quality, Information quality.","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"99 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing Cardiovascular Disease Prediction: A Synergistic Approach of Grey Wolf Levenberg Model and Neural Networks 优化心血管疾病预测:灰狼Levenberg模型和神经网络的协同方法
Pub Date : 2023-11-01 DOI: 10.20473/jisebi.9.2.119-135
Sheikh Amir Fayaz Fayaz, Majid Zaman, Sameer Kaul, Waseem Jeelani Bakshi
Background: One of the latest issues in predicting cardiovascular disease is the limited performance of current risk prediction models. Although several models have been developed, they often fail to identify a significant proportion of individuals who go on to develop the disease. This highlights the need for more accurate and personalized prediction models. Objective: This study aims to investigate the effectiveness of the Grey Wolf Levenberg Model and Neural Networks in predicting cardiovascular diseases. The objective is to identify a synergistic approach that can improve the accuracy of predictions. Through this research, the authors seek to contribute to the development of better tools for early detection and prevention of cardiovascular diseases. Methods: The study used a quantitative approach to develop and validate the GWLM_NARX model for predicting cardiovascular disease risk. The approach involved collecting and analyzing a large dataset of clinical and demographic variables. The performance of the model was then evaluated using various metrics such as accuracy, sensitivity, and specificity. Results: the study found that the GWLM_NARX model has shown promising results in predicting cardiovascular disease. The model was found to outperform other conventional methods, with an accuracy of over 90%. The synergistic approach of Grey Wolf Levenberg Model and Neural Networks has proved to be effective in predicting cardiovascular disease with high accuracy. Conclusion: The use of the Grey Wolf Levenberg-Marquardt Neural Network Autoregressive model (GWLM-NARX) in conjunction with traditional learning algorithms, as well as advanced machine learning tools, resulted in a more accurate and effective prediction model for cardiovascular disease. The study demonstrates the potential of machine learning techniques to improve diagnosis and treatment of heart disorders. However, further research is needed to improve the scalability and accuracy of these prediction systems, given the complexity of the data associated with cardiac illness. Keywords: Cardiovascular data, Clinical data., Decision tree, GWLM-NARX, Linear model functions
背景:预测心血管疾病的最新问题之一是当前风险预测模型的有限性能。虽然已经建立了几个模型,但它们往往不能确定很大一部分继续发展为这种疾病的个体。这凸显了对更准确和个性化的预测模型的需求。目的:探讨灰狼Levenberg模型和神经网络在心血管疾病预测中的有效性。目标是确定一种能够提高预测准确性的协同方法。通过这项研究,作者试图为开发更好的工具来早期发现和预防心血管疾病做出贡献。方法:采用定量方法建立并验证GWLM_NARX模型预测心血管疾病风险。该方法包括收集和分析临床和人口变量的大型数据集。然后使用各种指标如准确性、敏感性和特异性来评估模型的性能。结果:研究发现GWLM_NARX模型在预测心血管疾病方面显示出良好的效果。该模型优于其他传统方法,准确率超过90%。灰狼Levenberg模型与神经网络的协同预测方法在心血管疾病预测中具有较高的准确性。结论:将灰狼Levenberg-Marquardt神经网络自回归模型(GWLM-NARX)与传统学习算法以及先进的机器学习工具相结合,可以建立更准确有效的心血管疾病预测模型。这项研究证明了机器学习技术在改善心脏病诊断和治疗方面的潜力。然而,考虑到与心脏病相关的数据的复杂性,需要进一步的研究来提高这些预测系统的可扩展性和准确性。关键词:心血管数据;临床数据;决策树,GWLM-NARX,线性模型函数
{"title":"Optimizing Cardiovascular Disease Prediction: A Synergistic Approach of Grey Wolf Levenberg Model and Neural Networks","authors":"Sheikh Amir Fayaz Fayaz, Majid Zaman, Sameer Kaul, Waseem Jeelani Bakshi","doi":"10.20473/jisebi.9.2.119-135","DOIUrl":"https://doi.org/10.20473/jisebi.9.2.119-135","url":null,"abstract":"Background: One of the latest issues in predicting cardiovascular disease is the limited performance of current risk prediction models. Although several models have been developed, they often fail to identify a significant proportion of individuals who go on to develop the disease. This highlights the need for more accurate and personalized prediction models. Objective: This study aims to investigate the effectiveness of the Grey Wolf Levenberg Model and Neural Networks in predicting cardiovascular diseases. The objective is to identify a synergistic approach that can improve the accuracy of predictions. Through this research, the authors seek to contribute to the development of better tools for early detection and prevention of cardiovascular diseases. Methods: The study used a quantitative approach to develop and validate the GWLM_NARX model for predicting cardiovascular disease risk. The approach involved collecting and analyzing a large dataset of clinical and demographic variables. The performance of the model was then evaluated using various metrics such as accuracy, sensitivity, and specificity. Results: the study found that the GWLM_NARX model has shown promising results in predicting cardiovascular disease. The model was found to outperform other conventional methods, with an accuracy of over 90%. The synergistic approach of Grey Wolf Levenberg Model and Neural Networks has proved to be effective in predicting cardiovascular disease with high accuracy. Conclusion: The use of the Grey Wolf Levenberg-Marquardt Neural Network Autoregressive model (GWLM-NARX) in conjunction with traditional learning algorithms, as well as advanced machine learning tools, resulted in a more accurate and effective prediction model for cardiovascular disease. The study demonstrates the potential of machine learning techniques to improve diagnosis and treatment of heart disorders. However, further research is needed to improve the scalability and accuracy of these prediction systems, given the complexity of the data associated with cardiac illness. Keywords: Cardiovascular data, Clinical data., Decision tree, GWLM-NARX, Linear model functions","PeriodicalId":16185,"journal":{"name":"Journal of Information Systems Engineering and Business Intelligence","volume":"98 3-4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135510376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Information Systems Engineering and Business Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1