首页 > 最新文献

Big Data and Cognitive Computing最新文献

英文 中文
A Survey of Incremental Deep Learning for Defect Detection in Manufacturing 用于制造业缺陷检测的增量式深度学习调查
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-05 DOI: 10.3390/bdcc8010007
R. Mohandas, Mark Southern, Eoin O’Connell, Martin J Hayes
Deep learning based visual cognition has greatly improved the accuracy of defect detection, reducing processing times and increasing product throughput across a variety of manufacturing use cases. There is however a continuing need for rigorous procedures to dynamically update model-based detection methods that use sequential streaming during the training phase. This paper reviews how new process, training or validation information is rigorously incorporated in real time when detection exceptions arise during inspection. In particular, consideration is given to how new tasks, classes or decision pathways are added to existing models or datasets in a controlled fashion. An analysis of studies from the incremental learning literature is presented, where the emphasis is on the mitigation of process complexity challenges such as, catastrophic forgetting. Further, practical implementation issues that are known to affect the complexity of deep learning model architecture, including memory allocation for incoming sequential data or incremental learning accuracy, is considered. The paper highlights case study results and methods that have been used to successfully mitigate such real-time manufacturing challenges.
基于深度学习的视觉认知技术大大提高了缺陷检测的准确性,缩短了处理时间,提高了各种制造用例的产品吞吐量。然而,对于在训练阶段使用顺序流的基于模型的检测方法,仍然需要严格的程序来动态更新。本文回顾了在检测过程中出现检测异常时,如何实时严格地纳入新的流程、培训或验证信息。特别是考虑了如何以受控方式将新任务、类别或决策路径添加到现有模型或数据集中。本文对增量学习文献中的研究进行了分析,重点是减轻过程复杂性挑战,如灾难性遗忘。此外,还考虑了已知会影响深度学习模型架构复杂性的实际实施问题,包括传入连续数据的内存分配或增量学习的准确性。本文重点介绍了成功缓解此类实时制造挑战的案例研究结果和方法。
{"title":"A Survey of Incremental Deep Learning for Defect Detection in Manufacturing","authors":"R. Mohandas, Mark Southern, Eoin O’Connell, Martin J Hayes","doi":"10.3390/bdcc8010007","DOIUrl":"https://doi.org/10.3390/bdcc8010007","url":null,"abstract":"Deep learning based visual cognition has greatly improved the accuracy of defect detection, reducing processing times and increasing product throughput across a variety of manufacturing use cases. There is however a continuing need for rigorous procedures to dynamically update model-based detection methods that use sequential streaming during the training phase. This paper reviews how new process, training or validation information is rigorously incorporated in real time when detection exceptions arise during inspection. In particular, consideration is given to how new tasks, classes or decision pathways are added to existing models or datasets in a controlled fashion. An analysis of studies from the incremental learning literature is presented, where the emphasis is on the mitigation of process complexity challenges such as, catastrophic forgetting. Further, practical implementation issues that are known to affect the complexity of deep learning model architecture, including memory allocation for incoming sequential data or incremental learning accuracy, is considered. The paper highlights case study results and methods that have been used to successfully mitigate such real-time manufacturing challenges.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"78 22","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139381411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic Similarity of Common Verbal Expressions in Older Adults through a Pre-Trained Model 通过预训练模型识别老年人常见口头表达的语义相似性
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-29 DOI: 10.3390/bdcc8010003
Zuchao Li, Min Peng, Marcos Orellana, Patricio Santiago García, Guillermo Daniel Ramon, Jorge Luis Zambrano-Martinez, Andrés Patiño-León, María Verónica Serrano, Priscila Cedillo
Health problems in older adults lead to situations where communication with peers, family and caregivers becomes challenging for seniors; therefore, it is necessary to use alternative methods to facilitate communication. In this context, Augmentative and Alternative Communication (AAC) methods are widely used to support this population segment. Moreover, with Artificial Intelligence (AI), and specifically, machine learning algorithms, AAC can be improved. Although there have been several studies in this field, it is interesting to analyze common phrases used by seniors, depending on their context (i.e., slang and everyday expressions typical of their age). This paper proposes a semantic analysis of the common phrases of older adults and their corresponding meanings through Natural Language Processing (NLP) techniques and a pre-trained language model using semantic textual similarity to represent the older adults’ phrases with their corresponding graphic images (pictograms). The results show good scores achieved in the semantic similarity between the phrases of the older adults and the definitions, so the relationship between the phrase and the pictogram has a high degree of probability.
老年人的健康问题导致他们与同龄人、家人和照顾者之间的沟通变得困难重重,因此有必要使用替代方法来促进沟通。在这种情况下,辅助和替代性交流(AAC)方法被广泛用于为这部分人群提供支持。此外,借助人工智能(AI),特别是机器学习算法,AAC 可以得到改善。尽管在这一领域已有多项研究,但根据语境(即俚语和他们这个年龄的典型日常表达)分析老年人使用的常用短语还是很有意义的。本文提出通过自然语言处理(NLP)技术和使用语义文本相似性预先训练的语言模型,对老年人的常用短语及其相应含义进行语义分析,从而用相应的图形图像(象形图)来表示老年人的短语。结果表明,老年人的短语和定义之间的语义相似性得分很高,因此短语和象形图之间的关系具有很高的可能性。
{"title":"Semantic Similarity of Common Verbal Expressions in Older Adults through a Pre-Trained Model","authors":"Zuchao Li, Min Peng, Marcos Orellana, Patricio Santiago García, Guillermo Daniel Ramon, Jorge Luis Zambrano-Martinez, Andrés Patiño-León, María Verónica Serrano, Priscila Cedillo","doi":"10.3390/bdcc8010003","DOIUrl":"https://doi.org/10.3390/bdcc8010003","url":null,"abstract":"Health problems in older adults lead to situations where communication with peers, family and caregivers becomes challenging for seniors; therefore, it is necessary to use alternative methods to facilitate communication. In this context, Augmentative and Alternative Communication (AAC) methods are widely used to support this population segment. Moreover, with Artificial Intelligence (AI), and specifically, machine learning algorithms, AAC can be improved. Although there have been several studies in this field, it is interesting to analyze common phrases used by seniors, depending on their context (i.e., slang and everyday expressions typical of their age). This paper proposes a semantic analysis of the common phrases of older adults and their corresponding meanings through Natural Language Processing (NLP) techniques and a pre-trained language model using semantic textual similarity to represent the older adults’ phrases with their corresponding graphic images (pictograms). The results show good scores achieved in the semantic similarity between the phrases of the older adults and the definitions, so the relationship between the phrase and the pictogram has a high degree of probability.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"47 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139147192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BNMI-DINA: A Bayesian Cognitive Diagnosis Model for Enhanced Personalized Learning BNMI-DINA:用于增强个性化学习的贝叶斯认知诊断模型
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-29 DOI: 10.3390/bdcc8010004
Yiming Chen, Shuang Liang
In the field of education, cognitive diagnosis is crucial for achieving personalized learning. The widely adopted DINA (Deterministic Inputs, Noisy And gate) model uncovers students’ mastery of essential skills necessary to answer questions correctly. However, existing DINA-based approaches overlook the dependency between knowledge points, and their model training process is computationally inefficient for large datasets. In this paper, we propose a new cognitive diagnosis model called BNMI-DINA, which stands for Bayesian Network-based Multiprocess Incremental DINA. Our proposed model aims to enhance personalized learning by providing accurate and detailed assessments of students’ cognitive abilities. By incorporating a Bayesian network, BNMI-DINA establishes the dependency relationship between knowledge points, enabling more accurate evaluations of students’ mastery levels. To enhance model convergence speed, key steps of our proposed algorithm are parallelized. We also provide theoretical proof of the convergence of BNMI-DINA. Extensive experiments demonstrate that our approach effectively enhances model accuracy and reduces computational time compared to state-of-the-art cognitive diagnosis models.
在教育领域,认知诊断对于实现个性化学习至关重要。被广泛采用的 DINA(确定性输入、噪声和门)模型能发现学生对正确回答问题所需的基本技能的掌握情况。然而,现有的基于 DINA 的方法忽略了知识点之间的依赖关系,而且其模型训练过程对于大型数据集而言计算效率低下。在本文中,我们提出了一种新的认知诊断模型,称为 BNMI-DINA,即基于贝叶斯网络的多进程增量式 DINA。我们提出的模型旨在通过对学生的认知能力进行准确而详细的评估,来提高个性化学习的效果。通过结合贝叶斯网络,BNMI-DINA 建立了知识点之间的依赖关系,从而能够更准确地评估学生的掌握程度。为了提高模型的收敛速度,我们对算法的关键步骤进行了并行化处理。我们还提供了 BNMI-DINA 收敛性的理论证明。广泛的实验证明,与最先进的认知诊断模型相比,我们的方法有效地提高了模型的准确性,并缩短了计算时间。
{"title":"BNMI-DINA: A Bayesian Cognitive Diagnosis Model for Enhanced Personalized Learning","authors":"Yiming Chen, Shuang Liang","doi":"10.3390/bdcc8010004","DOIUrl":"https://doi.org/10.3390/bdcc8010004","url":null,"abstract":"In the field of education, cognitive diagnosis is crucial for achieving personalized learning. The widely adopted DINA (Deterministic Inputs, Noisy And gate) model uncovers students’ mastery of essential skills necessary to answer questions correctly. However, existing DINA-based approaches overlook the dependency between knowledge points, and their model training process is computationally inefficient for large datasets. In this paper, we propose a new cognitive diagnosis model called BNMI-DINA, which stands for Bayesian Network-based Multiprocess Incremental DINA. Our proposed model aims to enhance personalized learning by providing accurate and detailed assessments of students’ cognitive abilities. By incorporating a Bayesian network, BNMI-DINA establishes the dependency relationship between knowledge points, enabling more accurate evaluations of students’ mastery levels. To enhance model convergence speed, key steps of our proposed algorithm are parallelized. We also provide theoretical proof of the convergence of BNMI-DINA. Extensive experiments demonstrate that our approach effectively enhances model accuracy and reduces computational time compared to state-of-the-art cognitive diagnosis models.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":" 3","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139142073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Knowledge-Based and Generative-AI-Driven Pedagogical Conversational Agents: A Comparative Study of Grice’s Cooperative Principles and Trust 基于知识和生成式人工智能驱动的教学对话代理:格莱斯合作原则与信任的比较研究
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-26 DOI: 10.3390/bdcc8010002
Matthias Wölfel, Mehrnoush Barani Shirzad, Andreas Reich, Katharina Anderer
The emergence of generative language models (GLMs), such as OpenAI’s ChatGPT, is changing the way we communicate with computers and has a major impact on the educational landscape. While GLMs have great potential to support education, their use is not unproblematic, as they suffer from hallucinations and misinformation. In this paper, we investigate how a very limited amount of domain-specific data, from lecture slides and transcripts, can be used to build knowledge-based and generative educational chatbots. We found that knowledge-based chatbots allow full control over the system’s response but lack the verbosity and flexibility of GLMs. The answers provided by GLMs are more trustworthy and offer greater flexibility, but their correctness cannot be guaranteed. Adapting GLMs to domain-specific data trades flexibility for correctness.
生成式语言模型(GLM)(如 OpenAI 的 ChatGPT)的出现正在改变我们与计算机交流的方式,并对教育领域产生了重大影响。虽然 GLM 在支持教育方面有着巨大的潜力,但其使用也并非没有问题,因为它们会产生幻觉和错误信息。在本文中,我们研究了如何利用来自讲座幻灯片和记录稿的非常有限的特定领域数据来构建基于知识的生成式教育聊天机器人。我们发现,基于知识的聊天机器人可以对系统的反应进行完全控制,但却缺乏 GLMs 的滔滔不绝和灵活性。GLMs 提供的答案更可信、更灵活,但其正确性无法保证。根据特定领域的数据调整 GLM,可以用灵活性换取正确性。
{"title":"Knowledge-Based and Generative-AI-Driven Pedagogical Conversational Agents: A Comparative Study of Grice’s Cooperative Principles and Trust","authors":"Matthias Wölfel, Mehrnoush Barani Shirzad, Andreas Reich, Katharina Anderer","doi":"10.3390/bdcc8010002","DOIUrl":"https://doi.org/10.3390/bdcc8010002","url":null,"abstract":"The emergence of generative language models (GLMs), such as OpenAI’s ChatGPT, is changing the way we communicate with computers and has a major impact on the educational landscape. While GLMs have great potential to support education, their use is not unproblematic, as they suffer from hallucinations and misinformation. In this paper, we investigate how a very limited amount of domain-specific data, from lecture slides and transcripts, can be used to build knowledge-based and generative educational chatbots. We found that knowledge-based chatbots allow full control over the system’s response but lack the verbosity and flexibility of GLMs. The answers provided by GLMs are more trustworthy and offer greater flexibility, but their correctness cannot be guaranteed. Adapting GLMs to domain-specific data trades flexibility for correctness.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"102 4","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139157165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed Bayesian Inference for Large-Scale IoT Systems 大规模物联网系统的分布式贝叶斯推理
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-19 DOI: 10.3390/bdcc8010001
Eleni Vlachou, Aristeidis Karras, Christos N. Karras, Leonidas Theodorakopoulos, C. Halkiopoulos, S. Sioutas
In this work, we present a Distributed Bayesian Inference Classifier for Large-Scale Systems, where we assess its performance and scalability on distributed environments such as PySpark. The presented classifier consistently showcases efficient inference time, irrespective of the variations in the size of the test set, implying a robust ability to handle escalating data sizes without a proportional increase in computational demands. Notably, throughout the experiments, there is an observed increase in memory usage with growing test set sizes, this increment is sublinear, demonstrating the proficiency of the classifier in memory resource management. This behavior is consistent with the typical tendencies of PySpark tasks, which witness increasing memory consumption due to data partitioning and various data operations as datasets expand. CPU resource utilization, which is another crucial factor, also remains stable, emphasizing the capability of the classifier to manage larger computational workloads without significant resource strain. From a classification perspective, the Bayesian Logistic Regression Spark Classifier consistently achieves reliable performance metrics, with a particular focus on high specificity, indicating its aptness for applications where pinpointing true negatives is crucial. In summary, based on all experiments conducted under various data sizes, our classifier emerges as a top contender for scalability-driven applications in IoT systems, highlighting its dependable performance, adept resource management, and consistent prediction accuracy.
在这项工作中,我们介绍了一种适用于大规模系统的分布式贝叶斯推理分类器,并评估了它在 PySpark 等分布式环境中的性能和可扩展性。无论测试集的规模如何变化,所提出的分类器都能始终如一地显示出高效的推理时间,这意味着该分类器具有强大的能力来处理不断升级的数据规模,而不会相应增加计算需求。值得注意的是,在整个实验过程中,观察到内存使用量随着测试集大小的增加而增加,但这种增加是亚线性的,这表明分类器在内存资源管理方面非常熟练。这种行为与 PySpark 任务的典型趋势一致,即随着数据集的扩大,数据分区和各种数据操作会导致内存消耗增加。作为另一个关键因素的 CPU 资源利用率也保持稳定,这突出表明分类器有能力管理更大的计算工作量,而不会造成明显的资源压力。从分类的角度来看,贝叶斯逻辑回归 Spark 分类器始终保持着可靠的性能指标,尤其是在高特异性方面,这表明它非常适合于精确定位真阴性的应用。总之,基于在各种数据规模下进行的所有实验,我们的分类器成为物联网系统中可扩展性驱动型应用的最佳竞争者,突出了其可靠的性能、出色的资源管理和一致的预测准确性。
{"title":"Distributed Bayesian Inference for Large-Scale IoT Systems","authors":"Eleni Vlachou, Aristeidis Karras, Christos N. Karras, Leonidas Theodorakopoulos, C. Halkiopoulos, S. Sioutas","doi":"10.3390/bdcc8010001","DOIUrl":"https://doi.org/10.3390/bdcc8010001","url":null,"abstract":"In this work, we present a Distributed Bayesian Inference Classifier for Large-Scale Systems, where we assess its performance and scalability on distributed environments such as PySpark. The presented classifier consistently showcases efficient inference time, irrespective of the variations in the size of the test set, implying a robust ability to handle escalating data sizes without a proportional increase in computational demands. Notably, throughout the experiments, there is an observed increase in memory usage with growing test set sizes, this increment is sublinear, demonstrating the proficiency of the classifier in memory resource management. This behavior is consistent with the typical tendencies of PySpark tasks, which witness increasing memory consumption due to data partitioning and various data operations as datasets expand. CPU resource utilization, which is another crucial factor, also remains stable, emphasizing the capability of the classifier to manage larger computational workloads without significant resource strain. From a classification perspective, the Bayesian Logistic Regression Spark Classifier consistently achieves reliable performance metrics, with a particular focus on high specificity, indicating its aptness for applications where pinpointing true negatives is crucial. In summary, based on all experiments conducted under various data sizes, our classifier emerges as a top contender for scalability-driven applications in IoT systems, highlighting its dependable performance, adept resource management, and consistent prediction accuracy.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":" 7","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138962058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extraction of Significant Features by Fixed-Weight Layer of Processing Elements for the Development of an Efficient Spiking Neural Network Classifier 通过固定权重层处理元件提取重要特征以开发高效尖峰神经网络分类器
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-18 DOI: 10.3390/bdcc7040184
A. Sboev, R. Rybka, Dmitry Kunitsyn, A. Serenko, Vyacheslav Ilyin, Vadim V Putrolaynen
In this paper, we demonstrate that fixed-weight layers generated from random distribution or logistic functions can effectively extract significant features from input data, resulting in high accuracy on a variety of tasks, including Fisher’s Iris, Wisconsin Breast Cancer, and MNIST datasets. We have observed that logistic functions yield high accuracy with less dispersion in results. We have also assessed the precision of our approach under conditions of minimizing the number of spikes generated in the network. It is practically useful for reducing energy consumption in spiking neural networks. Our findings reveal that the proposed method demonstrates the highest accuracy on Fisher’s iris and MNIST datasets with decoding using logistic regression. Furthermore, they surpass the accuracy of the conventional (non-spiking) approach using only logistic regression in the case of Wisconsin Breast Cancer. We have also investigated the impact of non-stochastic spike generation on accuracy.
在本文中,我们证明了由随机分布或逻辑函数生成的固定权重层可以有效地从输入数据中提取重要特征,从而在费雪虹膜、威斯康星乳腺癌和 MNIST 数据集等各种任务中获得高准确率。我们观察到,逻辑函数能产生较高的准确率,而且结果的离散性较小。我们还评估了在尽量减少网络中产生的尖峰数量的条件下,我们的方法的精确度。这种方法对于降低尖峰神经网络的能耗非常实用。我们的研究结果表明,在使用逻辑回归解码的费舍尔虹膜和 MNIST 数据集上,我们提出的方法具有最高的精确度。此外,在威斯康星乳腺癌的案例中,它们的准确率超过了仅使用逻辑回归的传统(非尖峰)方法。我们还研究了非随机尖峰生成对准确性的影响。
{"title":"Extraction of Significant Features by Fixed-Weight Layer of Processing Elements for the Development of an Efficient Spiking Neural Network Classifier","authors":"A. Sboev, R. Rybka, Dmitry Kunitsyn, A. Serenko, Vyacheslav Ilyin, Vadim V Putrolaynen","doi":"10.3390/bdcc7040184","DOIUrl":"https://doi.org/10.3390/bdcc7040184","url":null,"abstract":"In this paper, we demonstrate that fixed-weight layers generated from random distribution or logistic functions can effectively extract significant features from input data, resulting in high accuracy on a variety of tasks, including Fisher’s Iris, Wisconsin Breast Cancer, and MNIST datasets. We have observed that logistic functions yield high accuracy with less dispersion in results. We have also assessed the precision of our approach under conditions of minimizing the number of spikes generated in the network. It is practically useful for reducing energy consumption in spiking neural networks. Our findings reveal that the proposed method demonstrates the highest accuracy on Fisher’s iris and MNIST datasets with decoding using logistic regression. Furthermore, they surpass the accuracy of the conventional (non-spiking) approach using only logistic regression in the case of Wisconsin Breast Cancer. We have also investigated the impact of non-stochastic spike generation on accuracy.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"93 3","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138995075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Artificial-Intelligence-Driven Spanish Poetry Classification Framework 人工智能驱动的西班牙语诗歌分类框架
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-14 DOI: 10.3390/bdcc7040183
Shutian Deng, Gang Wang, Hongjun Wang, Fuliang Chang
Spain possesses a vast number of poems. Most have features that mean they present significantly different styles. A superficial reading of these poems may confuse readers due to their complexity. Therefore, it is of vital importance to classify the style of the poems in advance. Currently, poetry classification studies are mostly carried out manually, which creates extremely high requirements for the professional quality of classifiers and consumes a large amount of time. Furthermore, the objectivity of the classification cannot be guaranteed because of the influence of the classifier’s subjectivity. To solve these problems, a Spanish poetry classification framework was designed using artificial intelligence technology, which improves the accuracy, efficiency, and objectivity of classification. First, an artificial-intelligence-driven Spanish poetry classification framework is described in detail, and is illustrated by a framework diagram to clearly represent each step in the process. The framework includes many algorithms and models, such as the Term Frequency–Inverse Document Frequency (TF_IDF), Bagging, Support Vector Machines (SVMs), Adaptive Boosting (AdaBoost), logistic regression (LR), Gradient Boosting Decision Trees (GBDT), LightGBM (LGB), eXtreme Gradient Boosting (XGBoost), and Random Forest (RF). The roles of each algorithm in the framework are clearly defined. Finally, experiments were performed for model selection, comparing the results of these algorithms.The Bagging model stood out for its high accuracy, and the experimental results showed that the proposed framework can help researchers carry out poetry research work more efficiently, accurately, and objectively.
西班牙拥有大量的诗歌。大多数诗歌的特点是风格迥异。由于这些诗歌的复杂性,肤浅的阅读可能会使读者感到困惑。因此,提前对诗歌风格进行分类至关重要。目前,诗歌分类研究多以人工方式进行,这对分类者的专业素质提出了极高的要求,也耗费了大量时间。此外,由于分类器的主观性影响,分类的客观性也无法保证。为了解决这些问题,利用人工智能技术设计了西班牙诗歌分类框架,提高了分类的准确性、效率和客观性。首先,详细介绍了人工智能驱动的西班牙语诗歌分类框架,并通过框架图清晰地表示了分类过程中的每一个步骤。该框架包括多种算法和模型,如词频-反向文档频率(TF_IDF)、袋式分类、支持向量机(SVM)、自适应提升(AdaBoost)、逻辑回归(LR)、梯度提升决策树(GBDT)、LightGBM(LGB)、极端梯度提升(XGBoost)和随机森林(RF)。每种算法在框架中的作用都有明确定义。最后,对模型选择进行了实验,比较了这些算法的结果。Bagging 模型因其高精度而脱颖而出,实验结果表明,所提出的框架可以帮助研究人员更高效、准确、客观地开展诗歌研究工作。
{"title":"An Artificial-Intelligence-Driven Spanish Poetry Classification Framework","authors":"Shutian Deng, Gang Wang, Hongjun Wang, Fuliang Chang","doi":"10.3390/bdcc7040183","DOIUrl":"https://doi.org/10.3390/bdcc7040183","url":null,"abstract":"Spain possesses a vast number of poems. Most have features that mean they present significantly different styles. A superficial reading of these poems may confuse readers due to their complexity. Therefore, it is of vital importance to classify the style of the poems in advance. Currently, poetry classification studies are mostly carried out manually, which creates extremely high requirements for the professional quality of classifiers and consumes a large amount of time. Furthermore, the objectivity of the classification cannot be guaranteed because of the influence of the classifier’s subjectivity. To solve these problems, a Spanish poetry classification framework was designed using artificial intelligence technology, which improves the accuracy, efficiency, and objectivity of classification. First, an artificial-intelligence-driven Spanish poetry classification framework is described in detail, and is illustrated by a framework diagram to clearly represent each step in the process. The framework includes many algorithms and models, such as the Term Frequency–Inverse Document Frequency (TF_IDF), Bagging, Support Vector Machines (SVMs), Adaptive Boosting (AdaBoost), logistic regression (LR), Gradient Boosting Decision Trees (GBDT), LightGBM (LGB), eXtreme Gradient Boosting (XGBoost), and Random Forest (RF). The roles of each algorithm in the framework are clearly defined. Finally, experiments were performed for model selection, comparing the results of these algorithms.The Bagging model stood out for its high accuracy, and the experimental results showed that the proposed framework can help researchers carry out poetry research work more efficiently, accurately, and objectively.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"59 4","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139003374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Computers’ Interpretations of Knowledge Representation Using Pre-Conceptual Schemas: An Approach Based on the BERT and Llama 2-Chat Models 计算机使用前概念模式解释知识表示:基于 BERT 和 Llama 2-Chat 模型的方法
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-14 DOI: 10.3390/bdcc7040182
Jesus Insuasti, Felipe Roa, C. M. Zapata-Jaramillo
Pre-conceptual schemas are a straightforward way to represent knowledge using controlled language regardless of context. Despite the benefits of using pre-conceptual schemas by humans, they present challenges when interpreted by computers. We propose an approach to making computers able to interpret the basic pre-conceptual schemas made by humans. To do that, the construction of a linguistic corpus is required to work with large language models—LLM. The linguistic corpus was mainly fed using Master’s and doctoral theses from the digital repository of the University of Nariño to produce a training dataset for re-training the BERT model; in addition, we complement this by explaining the elicited sentences in triads from the pre-conceptual schemas using one of the cutting-edge large language models in natural language processing: Llama 2-Chat by Meta AI. The diverse topics covered in these theses allowed us to expand the spectrum of linguistic use in the BERT model and empower the generative capabilities using the fine-tuned Llama 2-Chat model and the proposed solution. As a result, the first version of a computational solution was built to consume the language models based on BERT and Llama 2-Chat and thus automatically interpret pre-conceptual schemas by computers via natural language processing, adding, at the same time, generative capabilities. The validation of the computational solution was performed in two phases: the first one for detecting sentences and interacting with pre-conceptual schemas with students in the Formal Languages and Automata Theory course—the seventh semester of the systems engineering undergraduate program at the University of Nariño’s Tumaco campus. The second phase was for exploring the generative capabilities based on pre-conceptual schemas; this second phase was performed with students in the Object-oriented Design course—the second semester of the systems engineering undergraduate program at the University of Nariño’s Tumaco campus. This validation yielded favorable results in implementing natural language processing using the BERT and Llama 2-Chat models. In this way, some bases were laid for future developments related to this research topic.
前概念图式是使用受控语言表示知识的一种直接方法,不受上下文的影响。尽管人类使用前概念图式有很多好处,但在由计算机解释时却面临着挑战。我们提出了一种方法,让计算机能够解释人类的基本前概念图式。为此,需要构建一个语言语料库,以便与大型语言模型--LLM 协同工作。该语言语料库主要使用纳里尼奥大学数字资料库中的硕士和博士论文,为重新训练 BERT 模型提供训练数据集;此外,我们还使用自然语言处理领域最先进的大型语言模型之一,对前概念图式中的三元组句子进行解释,以此作为补充:此外,我们还利用自然语言处理领域最前沿的大型语言模型之一:Meta AI 公司的 Llama 2-Chat,来解释前概念图式中的三元组句子。这些论文所涉及的不同主题使我们能够扩大 BERT 模型的语言使用范围,并利用经过微调的 Llama 2-Chat 模型和建议的解决方案增强生成能力。因此,我们建立了第一个版本的计算解决方案,以使用基于 BERT 和 Llama 2-Chat 的语言模型,从而由计算机通过自然语言处理自动解释预概念模式,同时增加生成能力。计算解决方案的验证分两个阶段进行:第一阶段是检测句子,并与形式语言和自动机理论课程(纳里尼奥大学图马科校区系统工程本科课程的第七学期)的学生就前概念模式进行互动。第二阶段是探索基于前概念模式的生成能力;第二阶段是与面向对象设计课程(纳里尼奥大学图马科校区系统工程本科课程的第二学期)的学生一起进行。在使用 BERT 和 Llama 2-Chat 模型进行自然语言处理的过程中,这一验证取得了良好的结果。通过这种方式,为今后与该研究课题相关的发展奠定了一些基础。
{"title":"Computers’ Interpretations of Knowledge Representation Using Pre-Conceptual Schemas: An Approach Based on the BERT and Llama 2-Chat Models","authors":"Jesus Insuasti, Felipe Roa, C. M. Zapata-Jaramillo","doi":"10.3390/bdcc7040182","DOIUrl":"https://doi.org/10.3390/bdcc7040182","url":null,"abstract":"Pre-conceptual schemas are a straightforward way to represent knowledge using controlled language regardless of context. Despite the benefits of using pre-conceptual schemas by humans, they present challenges when interpreted by computers. We propose an approach to making computers able to interpret the basic pre-conceptual schemas made by humans. To do that, the construction of a linguistic corpus is required to work with large language models—LLM. The linguistic corpus was mainly fed using Master’s and doctoral theses from the digital repository of the University of Nariño to produce a training dataset for re-training the BERT model; in addition, we complement this by explaining the elicited sentences in triads from the pre-conceptual schemas using one of the cutting-edge large language models in natural language processing: Llama 2-Chat by Meta AI. The diverse topics covered in these theses allowed us to expand the spectrum of linguistic use in the BERT model and empower the generative capabilities using the fine-tuned Llama 2-Chat model and the proposed solution. As a result, the first version of a computational solution was built to consume the language models based on BERT and Llama 2-Chat and thus automatically interpret pre-conceptual schemas by computers via natural language processing, adding, at the same time, generative capabilities. The validation of the computational solution was performed in two phases: the first one for detecting sentences and interacting with pre-conceptual schemas with students in the Formal Languages and Automata Theory course—the seventh semester of the systems engineering undergraduate program at the University of Nariño’s Tumaco campus. The second phase was for exploring the generative capabilities based on pre-conceptual schemas; this second phase was performed with students in the Object-oriented Design course—the second semester of the systems engineering undergraduate program at the University of Nariño’s Tumaco campus. This validation yielded favorable results in implementing natural language processing using the BERT and Llama 2-Chat models. In this way, some bases were laid for future developments related to this research topic.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"1 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138973149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Text Classification Based on the Heterogeneous Graph Considering the Relationships between Documents 基于考虑文档间关系的异构图的文本分类
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-13 DOI: 10.3390/bdcc7040181
Hiromu Nakajima, Minoru Sasaki
Text classification is the task of estimating the genre of a document based on information such as word co-occurrence and frequency of occurrence. Text classification has been studied by various approaches. In this study, we focused on text classification using graph structure data. Conventional graph-based methods express relationships between words and relationships between words and documents as weights between nodes. Then, a graph neural network is used for learning. However, there is a problem that conventional methods are not able to represent the relationship between documents on the graph. In this paper, we propose a graph structure that considers the relationships between documents. In the proposed method, the cosine similarity of document vectors is set as weights between document nodes. This completes a graph that considers the relationship between documents. The graph is then input into a graph convolutional neural network for training. Therefore, the aim of this study is to improve the text classification performance of conventional methods by using this graph that considers the relationships between document nodes. In this study, we conducted evaluation experiments using five different corpora of English documents. The results showed that the proposed method outperformed the performance of the conventional method by up to 1.19%, indicating that the use of relationships between documents is effective. In addition, the proposed method was shown to be particularly effective in classifying long documents.
文本分类是根据单词共现和出现频率等信息估算文档流派的任务。已有多种方法对文本分类进行了研究。在本研究中,我们主要研究使用图结构数据进行文本分类。传统的基于图的方法将词与词之间的关系以及词与文档之间的关系表示为节点之间的权重。然后,使用图神经网络进行学习。然而,传统方法无法在图上表示文档之间的关系。在本文中,我们提出了一种考虑到文档之间关系的图结构。在所提出的方法中,文档向量的余弦相似度被设定为文档节点之间的权重。这样就完成了一个考虑了文档之间关系的图。然后将该图输入图卷积神经网络进行训练。因此,本研究的目的是通过使用这种考虑了文档节点之间关系的图来提高传统方法的文本分类性能。在这项研究中,我们使用五个不同的英语文档语料库进行了评估实验。结果表明,所提出的方法比传统方法的性能高出 1.19%,这表明使用文档之间的关系是有效的。此外,建议的方法在长文档分类方面也特别有效。
{"title":"Text Classification Based on the Heterogeneous Graph Considering the Relationships between Documents","authors":"Hiromu Nakajima, Minoru Sasaki","doi":"10.3390/bdcc7040181","DOIUrl":"https://doi.org/10.3390/bdcc7040181","url":null,"abstract":"Text classification is the task of estimating the genre of a document based on information such as word co-occurrence and frequency of occurrence. Text classification has been studied by various approaches. In this study, we focused on text classification using graph structure data. Conventional graph-based methods express relationships between words and relationships between words and documents as weights between nodes. Then, a graph neural network is used for learning. However, there is a problem that conventional methods are not able to represent the relationship between documents on the graph. In this paper, we propose a graph structure that considers the relationships between documents. In the proposed method, the cosine similarity of document vectors is set as weights between document nodes. This completes a graph that considers the relationship between documents. The graph is then input into a graph convolutional neural network for training. Therefore, the aim of this study is to improve the text classification performance of conventional methods by using this graph that considers the relationships between document nodes. In this study, we conducted evaluation experiments using five different corpora of English documents. The results showed that the proposed method outperformed the performance of the conventional method by up to 1.19%, indicating that the use of relationships between documents is effective. In addition, the proposed method was shown to be particularly effective in classifying long documents.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"115 4","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139003584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding the Influence of Genre-Specific Music Using Network Analysis and Machine Learning Algorithms 利用网络分析和机器学习算法了解特定音乐流派的影响
IF 3.7 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-04 DOI: 10.3390/bdcc7040180
Bishal Lamichhane, Aniket Kumar Singh, Sumana Devkota, Uttam Dhakal, Subham Singh, Chandra Dhakal
This study analyzes a network of musical influence using machine learning and network analysis techniques. A directed network model is used to represent the influence relations between artists as nodes and edges. Network properties and centrality measures are analyzed to identify influential patterns. In addition, influence within and outside the genre is quantified using in-genre and out-genre weights. Regression analysis is performed to determine the impact of musical attributes on influence. We find that speechiness, acousticness, and valence are the top features of the most influential artists. We also introduce the IRDI, an algorithm that provides an innovative approach to quantify an artist’s influence by capturing the degree of dominance among their followers. This approach underscores influential artists who drive the evolution of music, setting trends and significantly inspiring a new generation of artists. The independent cascade model is further employed to open up the temporal dynamics of influence propagation across the entire musical network, highlighting how initial seeds of influence can contagiously spread through the network. This multidisciplinary approach provides a nuanced understanding of musical influence that refines existing methods and sheds light on influential trends and dynamics.
本研究使用机器学习和网络分析技术分析了音乐影响网络。使用有向网络模型将艺术家之间的影响关系表示为节点和边。分析了网络特性和中心性度量,以确定影响模式。此外,使用类型内和类型外权重来量化类型内和类型外的影响。进行回归分析以确定音乐属性对影响的影响。我们发现,最具影响力的艺术家的最主要特征是言语性、声学性和价性。我们还介绍了IRDI,这是一种算法,它提供了一种创新的方法,通过捕捉艺术家在追随者中的主导程度来量化艺术家的影响力。这种方法强调了有影响力的艺术家,他们推动了音乐的发展,引领了潮流,并极大地激励了新一代艺术家。独立级联模型进一步揭示了整个音乐网络中影响传播的时间动态,突出了影响的初始种子如何通过网络传染传播。这种多学科的方法提供了对音乐影响的细致理解,改进了现有的方法,并揭示了有影响力的趋势和动态。
{"title":"Understanding the Influence of Genre-Specific Music Using Network Analysis and Machine Learning Algorithms","authors":"Bishal Lamichhane, Aniket Kumar Singh, Sumana Devkota, Uttam Dhakal, Subham Singh, Chandra Dhakal","doi":"10.3390/bdcc7040180","DOIUrl":"https://doi.org/10.3390/bdcc7040180","url":null,"abstract":"This study analyzes a network of musical influence using machine learning and network analysis techniques. A directed network model is used to represent the influence relations between artists as nodes and edges. Network properties and centrality measures are analyzed to identify influential patterns. In addition, influence within and outside the genre is quantified using in-genre and out-genre weights. Regression analysis is performed to determine the impact of musical attributes on influence. We find that speechiness, acousticness, and valence are the top features of the most influential artists. We also introduce the IRDI, an algorithm that provides an innovative approach to quantify an artist’s influence by capturing the degree of dominance among their followers. This approach underscores influential artists who drive the evolution of music, setting trends and significantly inspiring a new generation of artists. The independent cascade model is further employed to open up the temporal dynamics of influence propagation across the entire musical network, highlighting how initial seeds of influence can contagiously spread through the network. This multidisciplinary approach provides a nuanced understanding of musical influence that refines existing methods and sheds light on influential trends and dynamics.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"13 12","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138603177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Big Data and Cognitive Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1