首页 > 最新文献

JUITA : Jurnal Informatika最新文献

英文 中文
Bicluster Analysis of Cheng and Church's Algorithm to Identify Patterns of People's Welfare in Indonesia 对 Cheng 和 Church 算法进行双簇分析以确定印度尼西亚人民的福利模式
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17446
Laradea Marifni, Made Sumertajaya, U. Syafitri
Biclustering is a method of grouping numerical data where rows and columns are grouped simultaneously. The Cheng and Church (CC) algorithm is one of the bi-clustering algorithms that try to find the maximum bi-cluster with a high similarity value, called MSR (Mean Square Residue). The association of rows and columns is called a bi-cluster if the MSR is lower than a predetermined threshold value (delta). Detection of people's welfare in Indonesia using Bi-Clustering is essential to get an overview of the characteristics of people's interest in each province in Indonesia. Bi-Clustering using the CC algorithm requires a threshold value (delta) determined by finding the MSR value of the actual data. The threshold value (delta) must be smaller than the MSR of the actual data. This study's threshold values are 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, and 0.8. After evaluating the optimum delta by considering the MSR value and the bi-cluster formed, the optimum delta is obtained as 0.1, with the number of bi-cluster included as 4.
双聚类是一种对数字数据进行分组的方法,其中行和列同时分组。Cheng 和 Church(CC)算法是双聚类算法中的一种,它试图找到具有高相似度值(称为 MSR(均方残差))的最大双聚类。如果 MSR 低于预定的阈值(delta),则行和列的关联称为双簇。使用双聚类法检测印尼人民的福利对于全面了解印尼各省人民的利益特征至关重要。使用 CC 算法进行双聚类分析需要一个阈值(delta),该阈值由实际数据的 MSR 值决定。阈值(delta)必须小于实际数据的 MSR。本研究的阈值为 0.1、0.2、0.3、0.4、0.5、0.6、0.7 和 0.8。通过考虑 MSR 值和所形成的双簇,对最佳 delta 值进行评估后,得出最佳 delta 值为 0.1,双簇数量为 4。
{"title":"Bicluster Analysis of Cheng and Church's Algorithm to Identify Patterns of People's Welfare in Indonesia","authors":"Laradea Marifni, Made Sumertajaya, U. Syafitri","doi":"10.30595/juita.v11i2.17446","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17446","url":null,"abstract":"Biclustering is a method of grouping numerical data where rows and columns are grouped simultaneously. The Cheng and Church (CC) algorithm is one of the bi-clustering algorithms that try to find the maximum bi-cluster with a high similarity value, called MSR (Mean Square Residue). The association of rows and columns is called a bi-cluster if the MSR is lower than a predetermined threshold value (delta). Detection of people's welfare in Indonesia using Bi-Clustering is essential to get an overview of the characteristics of people's interest in each province in Indonesia. Bi-Clustering using the CC algorithm requires a threshold value (delta) determined by finding the MSR value of the actual data. The threshold value (delta) must be smaller than the MSR of the actual data. This study's threshold values are 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, and 0.8. After evaluating the optimum delta by considering the MSR value and the bi-cluster formed, the optimum delta is obtained as 0.1, with the number of bi-cluster included as 4.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"13 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139264315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysis and Implementation of the Apriori Algorithm for Strategies to Increase Sales at Sakinah Mart 分析和实施 Apriori 算法,制定提高 Sakinah Mart 销售额的策略
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17341
Karisma Dwi Fernanda, Arifin Puji Widodo, Julianto Lemantara
Sakinah Mart is a retail business that focuses on determining the layout of goods based on perceptions and implementing a discount system for specific items, but without offering bundling packages. This research aims to provide recommendations using the apriori algorithm as a decision-making tool for analyzing the layout of goods and bundling packages. The apriori algorithm is a data mining technique used to discover association rules and analyze customer purchases, specifically identifying the likelihood of customers buying item X along with item Y. The algorithm consists of two main components: support and confidence. The research applies the Cross-Industry Standard Process for Data Mining (CRISP-DM) method, utilizing the apriori algorithm to analyze sales transaction data. The dataset includes 2000 sales transactions with two attributes, resulting in the identification of 2 and 3 itemsets. The findings include 16 rules with a minimum support value of 42% and a minimum confidence of 85% for the layout of goods. For bundling packages, 5 rules with a minimum support value of 40% and a minimum confidence of 90% were generated. These results offer valuable recommendations to the company, using the apriori algorithm for analyzing the layout of goods and bundling packages.
Sakinah Mart 是一家零售企业,其业务重点是根据感知确定商品布局,并针对特定商品实施折扣制度,但不提供捆绑套餐。本研究旨在使用 apriori 算法作为决策工具,为分析商品布局和捆绑套餐提供建议。apriori 算法是一种数据挖掘技术,用于发现关联规则和分析客户购买情况,特别是识别客户在购买商品 Y 的同时购买商品 X 的可能性。研究采用跨行业数据挖掘标准流程(CRISP-DM)方法,利用 apriori 算法分析销售交易数据。数据集包括 2000 个带有两个属性的销售交易,从而识别出 2 个和 3 个项目集。研究结果包括:在商品布局方面,16 条规则的最小支持值为 42%,最小置信度为 85%。对于捆绑包装,产生了 5 条规则,最小支持值为 40%,最小置信度为 90%。这些结果为公司使用 apriori 算法分析货物布局和捆绑包装提供了有价值的建议。
{"title":"Analysis and Implementation of the Apriori Algorithm for Strategies to Increase Sales at Sakinah Mart","authors":"Karisma Dwi Fernanda, Arifin Puji Widodo, Julianto Lemantara","doi":"10.30595/juita.v11i2.17341","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17341","url":null,"abstract":"Sakinah Mart is a retail business that focuses on determining the layout of goods based on perceptions and implementing a discount system for specific items, but without offering bundling packages. This research aims to provide recommendations using the apriori algorithm as a decision-making tool for analyzing the layout of goods and bundling packages. The apriori algorithm is a data mining technique used to discover association rules and analyze customer purchases, specifically identifying the likelihood of customers buying item X along with item Y. The algorithm consists of two main components: support and confidence. The research applies the Cross-Industry Standard Process for Data Mining (CRISP-DM) method, utilizing the apriori algorithm to analyze sales transaction data. The dataset includes 2000 sales transactions with two attributes, resulting in the identification of 2 and 3 itemsets. The findings include 16 rules with a minimum support value of 42% and a minimum confidence of 85% for the layout of goods. For bundling packages, 5 rules with a minimum support value of 40% and a minimum confidence of 90% were generated. These results offer valuable recommendations to the company, using the apriori algorithm for analyzing the layout of goods and bundling packages.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"48 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139264413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sikarju: Expert System of Major Recommendation to Increase the Chances of Being Accepted by University Sikarju:增加被大学录取机会的专业推荐专家系统
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17424
Siti Izati Nabila, Ami Anggraini Samudra, Irsyadunas Irsyadunas
Major is one of the important factors in the world of lectures. Along with the increasing need for knowledge and skills required in the world of work, increasing the number of majors offered by tertiary institutions. The number of considerations from prospective students regarding the selection of majors causes students to be confused in determining the best major they will choose to continue their education. The research aims to design an expert system-based website that will be used to provide major recommendations. The method to be used is the forward chaining method, where this method works by matching data based on predetermined facts, then obtaining results based on matching the data. Based on the black box testing that has been done, the results show that the designed expert system is by the expected functionality. Therefore this expert system can be said to be feasible to use.
专业是授课的重要因素之一。随着工作领域对知识和技能需求的不断增加,高等院校开设的专业也越来越多。准大学生在选择专业时考虑的因素很多,导致学生在确定继续深造的最佳专业时感到困惑。本研究旨在设计一个基于专家系统的网站,用于提供专业建议。将使用的方法是前向链法,这种方法的工作原理是根据预先确定的事实匹配数据,然后根据匹配数据得出结果。根据所做的黑盒测试,结果显示所设计的专家系统达到了预期的功能。因此,可以说这个专家系统是可行的。
{"title":"Sikarju: Expert System of Major Recommendation to Increase the Chances of Being Accepted by University","authors":"Siti Izati Nabila, Ami Anggraini Samudra, Irsyadunas Irsyadunas","doi":"10.30595/juita.v11i2.17424","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17424","url":null,"abstract":"Major is one of the important factors in the world of lectures. Along with the increasing need for knowledge and skills required in the world of work, increasing the number of majors offered by tertiary institutions. The number of considerations from prospective students regarding the selection of majors causes students to be confused in determining the best major they will choose to continue their education. The research aims to design an expert system-based website that will be used to provide major recommendations. The method to be used is the forward chaining method, where this method works by matching data based on predetermined facts, then obtaining results based on matching the data. Based on the black box testing that has been done, the results show that the designed expert system is by the expected functionality. Therefore this expert system can be said to be feasible to use.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"27 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139265320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implementation of Particle Swarm Optimization on Sentiment Analysis of Cyberbullying using Random Forest 利用随机森林对网络欺凌的情感分析进行粒子群优化
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17920
Helma Herlinda, Muhammad Itqan Mazdadi, M. Muliadi, D. Kartini, Irwan Budiman
Social media has exerted a significant influence on the lives of the majority of individuals in the contemporary era. It not only enables communication among people within specific environments but also facilitates user connectivity in the virtual realm. Instagram is a social media platform that plays a pivotal role in the sharing of information and fostering communication among its users through the medium of photos and videos, which can be commented on by other users. The utilization of Instagram is consistently growing each year, thereby potentially yielding both positive and negative consequences. One prevalent negative consequence that frequently arises is cyberbullying. Conducting sentiment analysis on cyberbullying data can provide insights into the effectiveness of the employed methodology. This research was conducted as an experimental research, aiming to compare the performance of Random Forest and Random Forest after applying the Particle Swarm Optimization feature selection technique on three distinct data split compositions, namely 70:30, 80:20, and 90:10. The evaluation results indicate that the highest accuracy scores were achieved in the 90:10 data split configuration. Specifically, the Random Forest model yielded an accuracy of 87.50%, while the Random Forest model, after undergoing feature selection using the Particle Swarm Optimization algorithm, achieved an accuracy of 92.19%. Therefore, the implementation of Particle Swarm Optimization as a feature selection technique demonstrates the potential to enhance the accuracy of the Random Forest method.
社交媒体对当代大多数人的生活产生了重大影响。它不仅实现了特定环境中人与人之间的交流,还促进了虚拟领域中的用户连接。Instagram 是一个社交媒体平台,在用户之间通过照片和视频共享信息和促进交流方面发挥着举足轻重的作用,其他用户可以对这些照片和视频发表评论。Instagram 的使用率每年都在持续增长,因此可能产生积极和消极的后果。经常出现的一个普遍的负面影响就是网络欺凌。对网络欺凌数据进行情感分析可以深入了解所采用方法的有效性。本研究以实验研究的形式进行,旨在比较随机森林和随机森林在三种不同的数据拆分组合(即 70:30、80:20 和 90:10)上应用粒子群优化特征选择技术后的性能。评估结果表明,90:10 数据分割配置的准确率最高。具体来说,随机森林模型的准确率为 87.50%,而使用粒子群优化算法进行特征选择后,随机森林模型的准确率达到了 92.19%。因此,将粒子群优化算法作为特征选择技术的实施表明,它具有提高随机森林方法准确性的潜力。
{"title":"Implementation of Particle Swarm Optimization on Sentiment Analysis of Cyberbullying using Random Forest","authors":"Helma Herlinda, Muhammad Itqan Mazdadi, M. Muliadi, D. Kartini, Irwan Budiman","doi":"10.30595/juita.v11i2.17920","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17920","url":null,"abstract":"Social media has exerted a significant influence on the lives of the majority of individuals in the contemporary era. It not only enables communication among people within specific environments but also facilitates user connectivity in the virtual realm. Instagram is a social media platform that plays a pivotal role in the sharing of information and fostering communication among its users through the medium of photos and videos, which can be commented on by other users. The utilization of Instagram is consistently growing each year, thereby potentially yielding both positive and negative consequences. One prevalent negative consequence that frequently arises is cyberbullying. Conducting sentiment analysis on cyberbullying data can provide insights into the effectiveness of the employed methodology. This research was conducted as an experimental research, aiming to compare the performance of Random Forest and Random Forest after applying the Particle Swarm Optimization feature selection technique on three distinct data split compositions, namely 70:30, 80:20, and 90:10. The evaluation results indicate that the highest accuracy scores were achieved in the 90:10 data split configuration. Specifically, the Random Forest model yielded an accuracy of 87.50%, while the Random Forest model, after undergoing feature selection using the Particle Swarm Optimization algorithm, achieved an accuracy of 92.19%. Therefore, the implementation of Particle Swarm Optimization as a feature selection technique demonstrates the potential to enhance the accuracy of the Random Forest method.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"69 9","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139266455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Voting Scheme Nearest Neighbors by Difference Distance Metrics Measurement 通过差分距离度量法测量近邻投票方案
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.19298
G. Pradipta, Made Liandana, P. Ayu, Dandy Pramana Hostiadi, Putu Sumardika Eka Putra
K-Nearest Neighbor (KNN) is a widely used method for both classification and regression cases. This algorithm, known for its simplicity and effectiveness, relies primarily on the Euclidean formula for distance metrics. Therefore, this study aimed to develop a voting model where observations were made using different distance calculation formulas. The nearest neighbors algorithm was divided based on differences in distance measurements, with each resulting model contributing a vote to determine the final class. Consequently, three methods were proposed, namely k-nearest neighbors (KNN), Local Mean-based KNN, and Distance-Weighted neighbor (DWKNN), with an inclusion of a voting scheme. The robustness of these models was tested using umbilical cord data characterized by imbalance and small dataset size. The results showed that the proposed voting model for nearest neighbors consistently improved performance by an average of 1-2% across accuracy, precision, recall, and F1 score when compared to the conventional non-voting method.
K-Nearest Neighbor (KNN) 是一种广泛用于分类和回归的方法。这种算法以其简单有效而著称,主要依靠欧几里得公式进行距离度量。因此,本研究旨在开发一种投票模型,使用不同的距离计算公式进行观测。近邻算法根据距离测量的差异进行划分,每个结果模型贡献一票,以确定最终类别。因此,提出了三种方法,即 K 最近邻(KNN)、基于局部均值的 KNN 和距离加权邻(DWKNN),并加入了投票方案。利用脐带数据对这些模型的稳健性进行了测试,这些数据具有不平衡和数据集规模小的特点。结果表明,与传统的非投票方法相比,所提出的近邻投票模型在准确度、精确度、召回率和 F1 分数方面的性能平均提高了 1-2%。
{"title":"Voting Scheme Nearest Neighbors by Difference Distance Metrics Measurement","authors":"G. Pradipta, Made Liandana, P. Ayu, Dandy Pramana Hostiadi, Putu Sumardika Eka Putra","doi":"10.30595/juita.v11i2.19298","DOIUrl":"https://doi.org/10.30595/juita.v11i2.19298","url":null,"abstract":"K-Nearest Neighbor (KNN) is a widely used method for both classification and regression cases. This algorithm, known for its simplicity and effectiveness, relies primarily on the Euclidean formula for distance metrics. Therefore, this study aimed to develop a voting model where observations were made using different distance calculation formulas. The nearest neighbors algorithm was divided based on differences in distance measurements, with each resulting model contributing a vote to determine the final class. Consequently, three methods were proposed, namely k-nearest neighbors (KNN), Local Mean-based KNN, and Distance-Weighted neighbor (DWKNN), with an inclusion of a voting scheme. The robustness of these models was tested using umbilical cord data characterized by imbalance and small dataset size. The results showed that the proposed voting model for nearest neighbors consistently improved performance by an average of 1-2% across accuracy, precision, recall, and F1 score when compared to the conventional non-voting method.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"26 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139263721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implementation of RESHOT Method to Create a Good User Experience in an Application 在应用程序中采用 RESHOT 方法创建良好的用户体验
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17643
Ridwan Ahmad Ma'arif, Fauziah Fauziah
In recent years User Experience (UX) has become something that must be implemented in making applications. Every application design practitioner has applied this discipline, but often they wonder how to simplify a user's journey flow by using applications and make it easier, faster, and simpler. This research aims to explain a method that can simplify the User Experience comprehensively. RESHOT is used (Refine the challenge, Remove, Shrink, Hide, Organize, Time). This method will make applications pay more attention to aspects that can increase user satisfaction. As a result, this study contributes to an explanation of simplifying the flow of uploading donor data files using the RESHOT method in the X blood donor data collection application. The results of streamlining the flow of uploading donor data files have been tested on five respondents and have a 100% user success rate in completing tasks with an average processing time of 14.5 seconds. In testing, there is a misclick rate of 10.7%. This is because the user wants to explore the designed application. And this is also a limitation of this study, namely not making the overall application design interaction.
近年来,用户体验(UX)已成为制作应用程序时必须考虑的问题。每个应用程序设计从业人员都会应用这门学科,但他们往往不知道如何简化用户使用应用程序的流程,使其更方便、更快捷、更简单。本研究旨在解释一种可以全面简化用户体验的方法。我们采用了 RESHOT 方法(细化挑战、删除、缩小、隐藏、组织、时间)。这种方法将使应用程序更加关注能够提高用户满意度的方面。因此,本研究有助于解释在 X 献血者数据收集应用程序中使用 RESHOT 方法简化上传献血者数据文件的流程。简化上传献血者数据文件流程的结果已在五个受访者中进行了测试,用户完成任务的成功率为 100%,平均处理时间为 14.5 秒。在测试中,误点击率为 10.7%。这是因为用户希望探索所设计的应用程序。而这也是本研究的局限性,即没有做到整体应用设计的交互性。
{"title":"Implementation of RESHOT Method to Create a Good User Experience in an Application","authors":"Ridwan Ahmad Ma'arif, Fauziah Fauziah","doi":"10.30595/juita.v11i2.17643","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17643","url":null,"abstract":"In recent years User Experience (UX) has become something that must be implemented in making applications. Every application design practitioner has applied this discipline, but often they wonder how to simplify a user's journey flow by using applications and make it easier, faster, and simpler. This research aims to explain a method that can simplify the User Experience comprehensively. RESHOT is used (Refine the challenge, Remove, Shrink, Hide, Organize, Time). This method will make applications pay more attention to aspects that can increase user satisfaction. As a result, this study contributes to an explanation of simplifying the flow of uploading donor data files using the RESHOT method in the X blood donor data collection application. The results of streamlining the flow of uploading donor data files have been tested on five respondents and have a 100% user success rate in completing tasks with an average processing time of 14.5 seconds. In testing, there is a misclick rate of 10.7%. This is because the user wants to explore the designed application. And this is also a limitation of this study, namely not making the overall application design interaction.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"48 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139262534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysis of the Impact of Vectorization Methods on Machine Learning-Based Sentiment Analysis of Tweets Regarding Readiness for Offline Learning 分析矢量化方法对基于机器学习的推文情感分析在离线学习准备方面的影响
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17568
Yesi Novaria Kunang, Widya Putri Mentari
Twitter users use social media to express emotions about something, whether it is criticism or praise. Analyzing the opinions or sentiments in the tweets that Twitter users send can identify their emotions for a particular topic. This study aims to determine the impact of vectorization methods on public sentiment analysis regarding the readiness for offline learning in Indonesia during the Covid-19 pandemic. The authors labeled sentiment using two different approaches: manually and automatically using the NLP TextBlob library. We compared the vectorization method used by employing count vectorization, TF-IDF, and a combination of both. The feature vectors were then classified using three classification methods: naïve Bayes, logistic regression, and k-nearest neighbor, for both manual and automatic labeling. To assess the performance of sentiment analysis models, we used accuracy, precision, recall, and F1-score for performance metrics. The best results showed that the Logistic regression classifier with the feature extraction technique that combines count vectorization and TF-IDF provided the best performance for both data with manual and automatic labeling.
推特用户使用社交媒体表达对某事的情绪,无论是批评还是赞美。分析推特用户发送的推文中的观点或情绪可以识别他们对特定话题的情绪。本研究旨在确定矢量化方法对公众情绪分析的影响,分析印度尼西亚在 Covid-19 大流行期间线下学习的准备情况。作者使用两种不同的方法对情感进行了标注:人工标注和使用 NLP TextBlob 库自动标注。我们比较了采用计数矢量化、TF-IDF 以及两者结合的矢量化方法。然后,我们使用三种分类方法对特征向量进行了分类:奈夫贝叶斯、逻辑回归和 k-最近邻,并同时进行了手动和自动标注。为了评估情感分析模型的性能,我们使用了准确率、精确度、召回率和 F1 分数作为性能指标。最佳结果显示,逻辑回归分类器结合了计数矢量化和 TF-IDF 的特征提取技术,在手动和自动标注的数据中均表现最佳。
{"title":"Analysis of the Impact of Vectorization Methods on Machine Learning-Based Sentiment Analysis of Tweets Regarding Readiness for Offline Learning","authors":"Yesi Novaria Kunang, Widya Putri Mentari","doi":"10.30595/juita.v11i2.17568","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17568","url":null,"abstract":"Twitter users use social media to express emotions about something, whether it is criticism or praise. Analyzing the opinions or sentiments in the tweets that Twitter users send can identify their emotions for a particular topic. This study aims to determine the impact of vectorization methods on public sentiment analysis regarding the readiness for offline learning in Indonesia during the Covid-19 pandemic. The authors labeled sentiment using two different approaches: manually and automatically using the NLP TextBlob library. We compared the vectorization method used by employing count vectorization, TF-IDF, and a combination of both. The feature vectors were then classified using three classification methods: naïve Bayes, logistic regression, and k-nearest neighbor, for both manual and automatic labeling. To assess the performance of sentiment analysis models, we used accuracy, precision, recall, and F1-score for performance metrics. The best results showed that the Logistic regression classifier with the feature extraction technique that combines count vectorization and TF-IDF provided the best performance for both data with manual and automatic labeling.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"38 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139262866","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modified of Single Deepest Vertical Detection (SDVD) Algorithm for Amniotic Fluid Volume Classification 改进用于羊水体积分类的单一最深垂直检测(SDVD)算法
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.18435
Putu Desiana, Wulaning Ayu, G. Pradipta, Roy Rudolf Huizen, Kadek Eka, Sapta, G. Artana
Amniotic fluid a crucial role in ensuring the well-being of the fetus during pregnancy and is contained within the amnion cavity, which is surrounded by a membrane. Several studies have shown that volume of amniotic fluid can vary throughout pregnancy and is closely linked to the health and safety of the fetus. This indicates that it is essential to perform accurate measurement and identification of its volume. Obstetric specialist often use a manual method to identify amniotic fluid by visually determining the longest straight vertical line between the upper and lower boundaries. Therefore, this study aims to develop detection model, known as modified Single Deepest Vertical Detection (SDVD) algorithm to automatically measure the longest vertical line by following medical rules and regulations. SDVD algorithm was designed to measure the depth of amniotic fluid vertically by searching the column of pixels that comprised the image sample, excluding any intersection with the fetal body. Performance testing was carried out using 130 images by comparing the manual measurement results obtained by obstetric specialists and the proposed model. Based on the experimental results using modified SDVD, the average accuracy, precision, and recall achieved for amniotic fluid classification were 92.63%, 85.23%, and 95.6%, respectively.
羊水在怀孕期间对确保胎儿的健康起着至关重要的作用,羊水包含在羊膜腔内,羊膜腔由一层薄膜包围。多项研究表明,羊水量在整个孕期都会发生变化,并且与胎儿的健康和安全密切相关。这表明,准确测量和识别羊水量至关重要。产科专家通常使用人工方法,通过目测上下边界之间最长的垂直直线来识别羊水。因此,本研究旨在开发检测模型,即改进的单最深垂直检测(SDVD)算法,以遵循医学规则和条例自动测量最长垂直线。SDVD 算法通过搜索图像样本中的像素列,排除与胎儿身体的任何交点,垂直测量羊水深度。通过比较产科专家的人工测量结果和所提出的模型,使用 130 幅图像进行了性能测试。根据使用修改后的 SDVD 的实验结果,羊水分类的平均准确率、精确率和召回率分别为 92.63%、85.23% 和 95.6%。
{"title":"Modified of Single Deepest Vertical Detection (SDVD) Algorithm for Amniotic Fluid Volume Classification","authors":"Putu Desiana, Wulaning Ayu, G. Pradipta, Roy Rudolf Huizen, Kadek Eka, Sapta, G. Artana","doi":"10.30595/juita.v11i2.18435","DOIUrl":"https://doi.org/10.30595/juita.v11i2.18435","url":null,"abstract":"Amniotic fluid a crucial role in ensuring the well-being of the fetus during pregnancy and is contained within the amnion cavity, which is surrounded by a membrane. Several studies have shown that volume of amniotic fluid can vary throughout pregnancy and is closely linked to the health and safety of the fetus. This indicates that it is essential to perform accurate measurement and identification of its volume. Obstetric specialist often use a manual method to identify amniotic fluid by visually determining the longest straight vertical line between the upper and lower boundaries. Therefore, this study aims to develop detection model, known as modified Single Deepest Vertical Detection (SDVD) algorithm to automatically measure the longest vertical line by following medical rules and regulations. SDVD algorithm was designed to measure the depth of amniotic fluid vertically by searching the column of pixels that comprised the image sample, excluding any intersection with the fetal body. Performance testing was carried out using 130 images by comparing the manual measurement results obtained by obstetric specialists and the proposed model. Based on the experimental results using modified SDVD, the average accuracy, precision, and recall achieved for amniotic fluid classification were 92.63%, 85.23%, and 95.6%, respectively.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"17 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139265990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Facebook Prophet Model with Bayesian Optimization for USD Index Prediction 利用贝叶斯优化法预测美元指数的 Facebook 先知模型
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17880
Ahmad Fitra Hamdani, Daniel Swanjaya, Risa Helilintar
Accuracy is the primary focus in prediction research. Optimization is conducted to improve the performance of prediction models, thereby enhancing prediction accuracy. This study aims to optimize the Facebook Prophet model by performing hyperparameter tuning using Bayesian Optimization to improve the accuracy of USD Index Value prediction. Evaluation is conducted through multiple prediction experiments using different ranges of historical data. The results of the study demonstrate that performing hyperparameter tuning on the Facebook Prophet model yields better prediction results. Prior to parameter tuning, the MAPE indicator metric is 1.38% for the historical data range of 2014-2023, and it decreases to 1.33% after parameter tuning. Further evaluation shows improved prediction performance using different ranges of historical data. For the historical data range of 2015-2023, the MAPE value decreases from 1.39% to 1.20%. Similarly, for the data range of 2016-2023, the MAPE decreases from 1.12% to 0.80%. Furthermore, for the data range of 2017-2023, there is a decrease from 0.80% to 0.76%. This is followed by the data range of 2018-2023, with a decrease from 0.75% to 0.70%. Lastly, for the data range of 2019-2023, there is a decrease from 0.63% to 0.55%. These results demonstrate that performing Hyperparameter Optimization using Bayesian Optimization consistently improves prediction accuracy in the Facebook Prophet model.
准确性是预测研究的首要重点。进行优化是为了改善预测模型的性能,从而提高预测准确性。本研究旨在通过使用贝叶斯优化法进行超参数调整来优化 Facebook Prophet 模型,从而提高美元指数值预测的准确性。通过使用不同范围的历史数据进行多次预测实验来进行评估。研究结果表明,对 Facebook Prophet 模型进行超参数调整可获得更好的预测结果。参数调整前,2014-2023 年历史数据范围内的 MAPE 指标为 1.38%,参数调整后降至 1.33%。进一步的评估显示,使用不同范围的历史数据,预测性能有所提高。对于 2015-2023 年的历史数据范围,MAPE 值从 1.39% 降至 1.20%。同样,对于 2016-2023 年的数据范围,MAPE 值从 1.12% 降至 0.80%。此外,2017-2023 年的数据范围也从 0.80% 降至 0.76%。其次是 2018-2023 年的数据范围,从 0.75% 下降到 0.70%。最后是 2019-2023 年的数据范围,从 0.63% 下降到 0.55%。这些结果表明,使用贝叶斯优化技术进行超参数优化可以持续提高 Facebook Prophet 模型的预测准确性。
{"title":"Facebook Prophet Model with Bayesian Optimization for USD Index Prediction","authors":"Ahmad Fitra Hamdani, Daniel Swanjaya, Risa Helilintar","doi":"10.30595/juita.v11i2.17880","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17880","url":null,"abstract":"Accuracy is the primary focus in prediction research. Optimization is conducted to improve the performance of prediction models, thereby enhancing prediction accuracy. This study aims to optimize the Facebook Prophet model by performing hyperparameter tuning using Bayesian Optimization to improve the accuracy of USD Index Value prediction. Evaluation is conducted through multiple prediction experiments using different ranges of historical data. The results of the study demonstrate that performing hyperparameter tuning on the Facebook Prophet model yields better prediction results. Prior to parameter tuning, the MAPE indicator metric is 1.38% for the historical data range of 2014-2023, and it decreases to 1.33% after parameter tuning. Further evaluation shows improved prediction performance using different ranges of historical data. For the historical data range of 2015-2023, the MAPE value decreases from 1.39% to 1.20%. Similarly, for the data range of 2016-2023, the MAPE decreases from 1.12% to 0.80%. Furthermore, for the data range of 2017-2023, there is a decrease from 0.80% to 0.76%. This is followed by the data range of 2018-2023, with a decrease from 0.75% to 0.70%. Lastly, for the data range of 2019-2023, there is a decrease from 0.63% to 0.55%. These results demonstrate that performing Hyperparameter Optimization using Bayesian Optimization consistently improves prediction accuracy in the Facebook Prophet model.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"39 2","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139264659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implementation of Convolutional Neural Network Method in Identifying Fashion Image 卷积神经网络方法在时尚图像识别中的应用
Pub Date : 2023-11-17 DOI: 10.30595/juita.v11i2.17372
Christian Sri Kusuma Aditya, Vinna Rahmayanti Setyaning Nastiti, Qori Raditya Damayanti, Gian Bagus Sadewa
The fashion industry has changed a lot over the years, which makes it hard for people to compare different kinds of fashion. To make it easier, different styles of clothing are tried out to find the exact and precise look desired. So, we opted to employ the Convolutional Neural Network (CNN) method for fashion classification. This approach represents one of the methodologies employed to utilize computers for the purpose of recognizing and categorizing items. The goal of this research is to see how well the Convolutional Neural Network method classifies the Fashion-MNIST dataset compared to other methods, models, and classification processes used in previous research. The information in this dataset is about different types of clothes and accessories. These items are divided into 10 categories, which include ankle boots, bags, coats, dresses, pullovers, sandals, shirts, sneakers, t-shirts, and trousers. The new classification method worked better than before on the test dataset. It had an accuracy value of 95. 92%, which is higher than in previous research. This research also uses a method called image data generator to make the Fashion MNIST image better. This method helps prevent too much focus on certain details and makes the results more accurate.
多年来,时装业发生了很大变化,这使得人们很难对不同种类的时装进行比较。为了方便起见,人们会尝试不同风格的服装,以找到所需的精确造型。因此,我们选择使用卷积神经网络(CNN)方法进行时装分类。这种方法是利用计算机对物品进行识别和分类的方法之一。本研究的目标是,与以往研究中使用的其他方法、模型和分类过程相比,看看卷积神经网络方法对时尚-MNIST 数据集的分类效果如何。该数据集中的信息涉及不同类型的服装和配饰。这些物品分为 10 个类别,包括踝靴、包、外套、连衣裙、套头衫、凉鞋、衬衫、运动鞋、T恤和裤子。新的分类方法在测试数据集上的效果比以前更好。它的准确率达到了 95.92%,高于之前的研究。这项研究还使用了一种名为图像数据生成器的方法,使时尚 MNIST 图像变得更好。这种方法有助于避免过于关注某些细节,使结果更加准确。
{"title":"Implementation of Convolutional Neural Network Method in Identifying Fashion Image","authors":"Christian Sri Kusuma Aditya, Vinna Rahmayanti Setyaning Nastiti, Qori Raditya Damayanti, Gian Bagus Sadewa","doi":"10.30595/juita.v11i2.17372","DOIUrl":"https://doi.org/10.30595/juita.v11i2.17372","url":null,"abstract":"The fashion industry has changed a lot over the years, which makes it hard for people to compare different kinds of fashion. To make it easier, different styles of clothing are tried out to find the exact and precise look desired. So, we opted to employ the Convolutional Neural Network (CNN) method for fashion classification. This approach represents one of the methodologies employed to utilize computers for the purpose of recognizing and categorizing items. The goal of this research is to see how well the Convolutional Neural Network method classifies the Fashion-MNIST dataset compared to other methods, models, and classification processes used in previous research. The information in this dataset is about different types of clothes and accessories. These items are divided into 10 categories, which include ankle boots, bags, coats, dresses, pullovers, sandals, shirts, sneakers, t-shirts, and trousers. The new classification method worked better than before on the test dataset. It had an accuracy value of 95. 92%, which is higher than in previous research. This research also uses a method called image data generator to make the Fashion MNIST image better. This method helps prevent too much focus on certain details and makes the results more accurate.","PeriodicalId":151254,"journal":{"name":"JUITA : Jurnal Informatika","volume":"44 8","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139264268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
JUITA : Jurnal Informatika
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1