首页 > 最新文献

ACM Transactions on Asian and Low-Resource Language Information Processing最新文献

英文 中文
Fast Recurrent Neural Network with Bi-LSTM for Handwritten Tamil text segmentation in NLP 快速循环神经网络与 Bi-LSTM 在 NLP 中用于泰米尔语手写文本分割
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-02-07 DOI: 10.1145/3643808
C. Vinotheni, Lakshmana Pandian S.

Tamil text segmentation is a long-standing test in language comprehension that entails separating a record into adjacent pieces based on its semantic design. Each segment is important in its own way. The segments are organised according to the purpose of the content examination as text groups, sentences, phrases, words, characters or any other data unit. That process has been portioned using rapid tangled neural organisation in this research, which presents content segmentation methods based on deep learning in natural language processing (NLP). This study proposes a bidirectional long short-term memory (Bi-LSTM) neural network prototype in which fast recurrent neural network (FRNN) are used to learn Tamil text group embedding and phrases are fragmented using text-oriented data. As a result, this prototype is capable of handling variable measured setting data and gives a vast new dataset for naturally segmenting text in Tamil. In addition, we develop a segmentation prototype and show how well it sums up to unnoticeable regular content using this dataset as a base. With Bi-LSTM, the segmentation precision of FRNN is superior to that of other segmentation approaches; however, it is still inferior to that of certain other techniques. Every content is scaled to the required size in the proposed framework, which is immediately accessible for the preparation. This means, each word in a scaled Tamil text is employed to prepare neural organisation as fragmented content. The results reveal that the proposed framework produces high rates of segmentation for manually authored material that are nearly equivalent to segmentation-based plans.

泰米尔语文本分段是语言理解中一项历史悠久的测试,需要根据语义设计将记录分成相邻的片段。每个片段都有其自身的重要性。这些片段根据内容检查的目的组织成文本组、句子、短语、单词、字符或任何其他数据单元。本研究利用快速纠缠神经组织对这一过程进行了分段,提出了基于自然语言处理(NLP)中深度学习的内容分段方法。本研究提出了一种双向长短期记忆(Bi-LSTM)神经网络原型,其中使用了快速循环神经网络(FRNN)来学习泰米尔语文本组嵌入,并使用面向文本的数据对短语进行分割。因此,该原型能够处理可变的测量设置数据,并为自然分割泰米尔语文本提供了一个庞大的新数据集。此外,我们还开发了一个分段原型,并以该数据集为基础,展示了它对不易察觉的常规内容的总结效果。在使用 Bi-LSTM 的情况下,FRNN 的分割精度优于其他分割方法,但仍低于某些其他技术。在所提出的框架中,每个内容都被缩放为所需的大小,可立即用于准备工作。这意味着,缩放的泰米尔语文本中的每个单词都会被用作神经组织的片段内容。结果表明,对于人工撰写的材料,建议的框架能产生很高的分割率,几乎等同于基于分割的计划。
{"title":"Fast Recurrent Neural Network with Bi-LSTM for Handwritten Tamil text segmentation in NLP","authors":"C. Vinotheni, Lakshmana Pandian S.","doi":"10.1145/3643808","DOIUrl":"https://doi.org/10.1145/3643808","url":null,"abstract":"<p>Tamil text segmentation is a long-standing test in language comprehension that entails separating a record into adjacent pieces based on its semantic design. Each segment is important in its own way. The segments are organised according to the purpose of the content examination as text groups, sentences, phrases, words, characters or any other data unit. That process has been portioned using rapid tangled neural organisation in this research, which presents content segmentation methods based on deep learning in natural language processing (NLP). This study proposes a bidirectional long short-term memory (Bi-LSTM) neural network prototype in which fast recurrent neural network (FRNN) are used to learn Tamil text group embedding and phrases are fragmented using text-oriented data. As a result, this prototype is capable of handling variable measured setting data and gives a vast new dataset for naturally segmenting text in Tamil. In addition, we develop a segmentation prototype and show how well it sums up to unnoticeable regular content using this dataset as a base. With Bi-LSTM, the segmentation precision of FRNN is superior to that of other segmentation approaches; however, it is still inferior to that of certain other techniques. Every content is scaled to the required size in the proposed framework, which is immediately accessible for the preparation. This means, each word in a scaled Tamil text is employed to prepare neural organisation as fragmented content. The results reveal that the proposed framework produces high rates of segmentation for manually authored material that are nearly equivalent to segmentation-based plans.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"16 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139752555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Seq2Set2Seq: A Two-stage Disentangled Method for Reply Keyword Generation in Social Media via Multi-label Prediction and Determinantal Point Processes Seq2Set2Seq:通过多标签预测和确定性点过程在社交媒体中生成回复关键词的两阶段分离法
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-02-05 DOI: 10.1145/3644074
Jie Liu, Yaguang Li, Shizhu He, Shun Wu, Kang Liu, Shenping Liu, Jiong Wang, Qing Zhang

Social media produces large amounts of contents every day. How to predict the potential influences of the contents from a social reply feedback perspective is a key issue that has not been explored. Thus, we propose a novel task named reply keyword prediction in social media, which aims to predict the keywords in the potential replies as many aspects as possible. One prerequisite challenge is that the accessible social media datasets labeling such keywords remain absent. To solve this issue, we propose a new dataset, to study the reply keyword prediction in Social Media. This task could be seen as a single-turn dialogue keyword prediction for open-domain dialogue system. However, existing methods for dialogue keyword prediction cannot be adopted directly, which have two main drawbacks. First, they do not provide an explicit mechanism to model topic complementarity between keywords which is crucial in social media to controllably model all aspects of replies. Second, the collocations of keywords are not explicitly modeled, which also makes it less controllable to optimize for fine-grained prediction since the context information is much less than that in dialogue. To address these issues, we propose a two-stage disentangled framework, which can optimize the complementarity and collocation explicitly in a disentangled fashion. In the first stage, we use a sequence-to-set paradigm via multi-label prediction and determinantal point processes, to generate a set of keyword seeds satisfying the complementarity. In the second stage, we adopt a set-to-sequence paradigm via seq2seq model with the keyword seeds guidance from the set, to generate the more-fine-grained keywords with collocation. Experiments show that this method can generate not only a more diverse set of keywords but also more relevant and consistent keywords. Furthermore, the keywords obtained based on this method can achieve better reply generation results in the retrieval-based system than others.

社交媒体每天都会产生大量内容。如何从社交回复反馈的角度预测这些内容的潜在影响是一个尚未探索的关键问题。因此,我们提出了一项名为 "社交媒体中回复关键词预测 "的新任务,旨在尽可能多地预测潜在回复中的关键词。一个先决挑战是,标注此类关键词的可访问社交媒体数据集仍然缺乏。为了解决这个问题,我们提出了一个新的数据集来研究社交媒体中的回复关键词预测。这项任务可视为开放域对话系统的单轮对话关键词预测。然而,现有的对话关键词预测方法不能直接采用,它们有两个主要缺点。首先,它们没有提供明确的机制来模拟关键词之间的话题互补性,而这在社交媒体中对于可控地模拟回复的各个方面至关重要。其次,关键词的搭配没有明确建模,这也使得优化细粒度预测的可控性降低,因为上下文信息比对话中的信息要少得多。为了解决这些问题,我们提出了一个两阶段分解框架,可以分解的方式明确优化互补性和搭配。在第一阶段,我们使用序列到集合范式,通过多标签预测和行列式点过程,生成一组满足互补性的关键词种子。在第二阶段,我们通过 seq2seq 模型,采用集合到序列的范式,以集合中的关键字种子为导向,生成具有搭配性的更细粒度关键字。实验表明,这种方法不仅能生成更多样化的关键词集,还能生成更相关、更一致的关键词。此外,在基于检索的系统中,基于该方法生成的关键词能获得比其他方法更好的回复生成结果。
{"title":"Seq2Set2Seq: A Two-stage Disentangled Method for Reply Keyword Generation in Social Media via Multi-label Prediction and Determinantal Point Processes","authors":"Jie Liu, Yaguang Li, Shizhu He, Shun Wu, Kang Liu, Shenping Liu, Jiong Wang, Qing Zhang","doi":"10.1145/3644074","DOIUrl":"https://doi.org/10.1145/3644074","url":null,"abstract":"<p>Social media produces large amounts of contents every day. How to predict the potential influences of the contents from a social reply feedback perspective is a key issue that has not been explored. Thus, we propose a novel task named reply keyword prediction in social media, which aims to predict the keywords in the potential replies as many aspects as possible. One prerequisite challenge is that the accessible social media datasets labeling such keywords remain absent. To solve this issue, we propose a new dataset, to study the reply keyword prediction in Social Media. This task could be seen as a single-turn dialogue keyword prediction for open-domain dialogue system. However, existing methods for dialogue keyword prediction cannot be adopted directly, which have two main drawbacks. First, they do not provide an explicit mechanism to model topic complementarity between keywords which is crucial in social media to controllably model all aspects of replies. Second, the collocations of keywords are not explicitly modeled, which also makes it less controllable to optimize for fine-grained prediction since the context information is much less than that in dialogue. To address these issues, we propose a two-stage disentangled framework, which can optimize the complementarity and collocation explicitly in a disentangled fashion. In the first stage, we use a sequence-to-set paradigm via multi-label prediction and determinantal point processes, to generate a set of keyword seeds satisfying the complementarity. In the second stage, we adopt a set-to-sequence paradigm via seq2seq model with the keyword seeds guidance from the set, to generate the more-fine-grained keywords with collocation. Experiments show that this method can generate not only a more diverse set of keywords but also more relevant and consistent keywords. Furthermore, the keywords obtained based on this method can achieve better reply generation results in the retrieval-based system than others.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"1 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139752535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved BIO-based Chinese Automatic Abstract-generation Model 基于 BIO 的改进型中文自动摘要生成模型
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-02-05 DOI: 10.1145/3643695
Qing Li, Weibin Wan, Yuming Zhao, Xiaoyan Jiang

With its unique information-filtering function, text summarization technology has become a significant aspect of search engines and question-and-answer systems. However, existing models that include the copy mechanism often lack the ability to extract important fragments, resulting in generated content that suffers from thematic deviation and insufficient generalization. Specifically, Chinese automatic summarization using traditional generation methods often loses semantics because of its reliance on word lists. To address these issues, we proposed the novel BioCopy mechanism for the summarization task. By training the tags of predictive words and reducing the probability distribution range on the glossary, we enhanced the ability to generate continuous segments, which effectively solves the above problems. Additionally, we applied reinforced canonicality to the inputs to obtain better model results, making the model share the sub-network weight parameters and sparsing the model output to reduce the search space for model prediction. To further improve the model’s performance, we calculated the bilingual evaluation understudy (BLEU) score on the English dataset CNN/DailyMail to filter the thresholds and reduce the difficulty of word separation and the dependence of the output on the word list. We fully fine-tuned the model using the LCSTS dataset for the Chinese summarization task and conducted small-sample experiments using the CSL dataset. We also conducted ablation experiments on the Chinese dataset. The experimental results demonstrate that the optimized model can learn the semantic representation of the original text better than other models and performs well with small sample sizes.

凭借其独特的信息过滤功能,文本摘要技术已成为搜索引擎和问答系统的重要组成部分。然而,现有的包含复制机制的模型往往缺乏提取重要片段的能力,导致生成的内容存在主题偏离和概括性不足的问题。具体来说,使用传统生成方法进行中文自动摘要时,由于依赖词表,往往会丢失语义。为了解决这些问题,我们针对摘要任务提出了新颖的 BioCopy 机制。通过训练预测词的标签和缩小词汇表的概率分布范围,我们增强了生成连续词段的能力,从而有效地解决了上述问题。此外,为了获得更好的模型效果,我们还对输入进行了强化规范性处理,使模型共享子网络权重参数,并对模型输出进行稀疏化处理,以减少模型预测的搜索空间。为了进一步提高模型的性能,我们在英文数据集 CNN/DailyMail 上计算了双语评估劣度(BLEU)得分,以过滤阈值,降低分词难度和输出对词表的依赖性。我们使用 LCSTS 数据集对中文摘要任务的模型进行了全面微调,并使用 CSL 数据集进行了小样本实验。我们还在中文数据集上进行了消减实验。实验结果表明,优化后的模型能比其他模型更好地学习原文的语义表征,并且在样本量较小的情况下表现良好。
{"title":"Improved BIO-based Chinese Automatic Abstract-generation Model","authors":"Qing Li, Weibin Wan, Yuming Zhao, Xiaoyan Jiang","doi":"10.1145/3643695","DOIUrl":"https://doi.org/10.1145/3643695","url":null,"abstract":"<p>With its unique information-filtering function, text summarization technology has become a significant aspect of search engines and question-and-answer systems. However, existing models that include the copy mechanism often lack the ability to extract important fragments, resulting in generated content that suffers from thematic deviation and insufficient generalization. Specifically, Chinese automatic summarization using traditional generation methods often loses semantics because of its reliance on word lists. To address these issues, we proposed the novel BioCopy mechanism for the summarization task. By training the tags of predictive words and reducing the probability distribution range on the glossary, we enhanced the ability to generate continuous segments, which effectively solves the above problems. Additionally, we applied reinforced canonicality to the inputs to obtain better model results, making the model share the sub-network weight parameters and sparsing the model output to reduce the search space for model prediction. To further improve the model’s performance, we calculated the bilingual evaluation understudy (BLEU) score on the English dataset CNN/DailyMail to filter the thresholds and reduce the difficulty of word separation and the dependence of the output on the word list. We fully fine-tuned the model using the LCSTS dataset for the Chinese summarization task and conducted small-sample experiments using the CSL dataset. We also conducted ablation experiments on the Chinese dataset. The experimental results demonstrate that the optimized model can learn the semantic representation of the original text better than other models and performs well with small sample sizes.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"304 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139752620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Expert System for Indian Sign Language Recognition using Spatial Attention based Feature and Temporal Feature 利用空间注意力特征和时间特征识别印度手语的专家系统
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-02-03 DOI: 10.1145/3643824
Soumen Das, Saroj Kr. Biswas, Biswajit Purkayastha

Sign Language (SL) is the only means of communication for the hearing-impaired people. Normal people have difficulty understanding SL, resulting in a communication barrier between hearing impaired people and hearing community. However, the Sign Language Recognition System (SLRS) has helped to bridge the communication gap. Many SLRs are proposed for recognizing SL; however, a limited number of works are reported for Indian Sign Language (ISL). Most of the existing SLRS focus on global features other than the Region of Interest (ROI). Focusing more on the hand region and extracting local features from the ROI improves system accuracy. The attention mechanism is a widely used technique for emphasizing the ROI. However, only a few SLRS used the attention method. They employed the Convolution Block Attention Module (CBAM) and temporal attention but Spatial Attention (SA) is not utilized in previous SLRS. Therefore, a novel SA based SLRS named Spatial Attention-based Sign Language Recognition Module (SASLRM) is proposed to recognize ISL words for emergency situations. SASLRM recognizes ISL words by combining convolution features from a pretrained VGG-19 model and attention features from a SA module. The proposed model accomplished an average accuracy of 95.627% on the ISL dataset. The proposed SASLRM is further validated on LSA64, WLASL and Cambridge Hand Gesture Recognition (HGR) datasets where, the proposed model reached an accuracy of 97.84 %, 98.86% and 98.22’% respectively. The results indicate the effectiveness of the proposed SLRS in comparison with the existing SLRS.

手语是听障人士唯一的交流方式。正常人很难理解手语,导致听障人士与健听群体之间存在沟通障碍。然而,手语识别系统(SLRS)有助于消除这一沟通障碍。许多手语识别系统都是为识别手语而提出的,但针对印度手语(ISL)的报告数量有限。大多数现有的手语识别系统都侧重于兴趣区域(ROI)以外的全局特征。更多地关注手部区域并从感兴趣区域中提取局部特征可提高系统的准确性。注意力机制是一种广泛使用的强调 ROI 的技术。然而,只有少数 SLRS 使用了注意力方法。他们使用了卷积块注意力模块(CBAM)和时间注意力,但空间注意力(SA)在以前的 SLRS 中并没有使用。因此,我们提出了一种基于空间注意力的新型手语识别系统,名为基于空间注意力的手语识别模块(SASLRM),用于识别紧急情况下的 ISL 词语。SASLRM 通过结合来自预训练 VGG-19 模型的卷积特征和来自 SA 模块的注意力特征来识别 ISL 单词。所提出的模型在 ISL 数据集上的平均准确率达到 95.627%。提议的 SASLRM 在 LSA64、WLASL 和剑桥手势识别(HGR)数据集上得到进一步验证,准确率分别达到 97.84%、98.86% 和 98.22'%。这些结果表明,与现有的 SLRS 相比,所提出的 SLRS 非常有效。
{"title":"An Expert System for Indian Sign Language Recognition using Spatial Attention based Feature and Temporal Feature","authors":"Soumen Das, Saroj Kr. Biswas, Biswajit Purkayastha","doi":"10.1145/3643824","DOIUrl":"https://doi.org/10.1145/3643824","url":null,"abstract":"<p>Sign Language (SL) is the only means of communication for the hearing-impaired people. Normal people have difficulty understanding SL, resulting in a communication barrier between hearing impaired people and hearing community. However, the Sign Language Recognition System (SLRS) has helped to bridge the communication gap. Many SLRs are proposed for recognizing SL; however, a limited number of works are reported for Indian Sign Language (ISL). Most of the existing SLRS focus on global features other than the Region of Interest (ROI). Focusing more on the hand region and extracting local features from the ROI improves system accuracy. The attention mechanism is a widely used technique for emphasizing the ROI. However, only a few SLRS used the attention method. They employed the Convolution Block Attention Module (CBAM) and temporal attention but Spatial Attention (SA) is not utilized in previous SLRS. Therefore, a novel SA based SLRS named Spatial Attention-based Sign Language Recognition Module (SASLRM) is proposed to recognize ISL words for emergency situations. SASLRM recognizes ISL words by combining convolution features from a pretrained VGG-19 model and attention features from a SA module. The proposed model accomplished an average accuracy of 95.627% on the ISL dataset. The proposed SASLRM is further validated on LSA64, WLASL and Cambridge Hand Gesture Recognition (HGR) datasets where, the proposed model reached an accuracy of 97.84 %, 98.86% and 98.22’% respectively. The results indicate the effectiveness of the proposed SLRS in comparison with the existing SLRS.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"13 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139678930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Construction of Interval-Valued Fuzzy Hindi WordNet using Lexico-Syntactic Patterns and Word Embeddings 利用词典句法模式和词语嵌入自动构建区间值模糊印地语词网
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-02-02 DOI: 10.1145/3643132
Minni Jain, Rajni Jindal, Amita Jain

A computational lexicon is the backbone of any language processing system. It helps computers to understand the language complexity as a human does by inculcating words and their semantic associations. Manually constructed famous Hindi WordNet (HWN) consists of various classical semantic relations (crisp relations). To handle uncertainty and represent Hindi WordNet more semantically, Type- 1 fuzzy graphs are applied to relations of Hindi WordNet. But uncertainty in the crisp membership degree is not considered in Type 1 fuzzy set (T1FS). Also collecting billions (5,55,69,51,753 relations in HWN) of membership values from experts (humans) is not feasible. This paper applied the concept of Interval-Valued Fuzzy graphs and proposed Interval- Valued Fuzzy Hindi WordNet (IVFHWN). IVFHWN automatically identifies Interval- Valued Fuzzy relations between words and their degree of membership using word embeddings and lexico-syntactic patterns. The experimental results for the word sense disambiguation problem show better outcomes when IVFHWN is being used in place of Type 1 Fuzzy Hindi WordNet and classical Hindi WordNet.

计算词典是任何语言处理系统的支柱。它通过灌输单词及其语义关联,帮助计算机像人类一样理解语言的复杂性。人工构建的著名印地语词网(HWN)由各种经典语义关系(清晰关系)组成。为了处理不确定性并更语义化地表示印地语 WordNet,对印地语 WordNet 的关系应用了 1 类模糊图。但在 1 类模糊集(T1FS)中没有考虑清晰成员度的不确定性。此外,从专家(人类)那里收集数十亿(HWN 中的 5,55,69,51,753 个关系)的成员值也不可行。本文应用了区间值模糊图的概念,并提出了区间值模糊印地语词网(IVFHWN)。IVFHWN 利用词嵌入和词义句法模式自动识别词与词之间的区间值模糊关系及其成员度。词义消歧问题的实验结果表明,用 IVFHWN 代替第一类模糊印地语词网和经典印地语词网时,效果更好。
{"title":"Automatic Construction of Interval-Valued Fuzzy Hindi WordNet using Lexico-Syntactic Patterns and Word Embeddings","authors":"Minni Jain, Rajni Jindal, Amita Jain","doi":"10.1145/3643132","DOIUrl":"https://doi.org/10.1145/3643132","url":null,"abstract":"<p>A computational lexicon is the backbone of any language processing system. It helps computers to understand the language complexity as a human does by inculcating words and their semantic associations. Manually constructed famous Hindi WordNet (HWN) consists of various classical semantic relations (crisp relations). To handle uncertainty and represent Hindi WordNet more semantically, Type- 1 fuzzy graphs are applied to relations of Hindi WordNet. But uncertainty in the crisp membership degree is not considered in Type 1 fuzzy set (T1FS). Also collecting billions (5,55,69,51,753 relations in HWN) of membership values from experts (humans) is not feasible. This paper applied the concept of Interval-Valued Fuzzy graphs and proposed Interval- Valued Fuzzy Hindi WordNet (IVFHWN). IVFHWN automatically identifies Interval- Valued Fuzzy relations between words and their degree of membership using word embeddings and lexico-syntactic patterns. The experimental results for the word sense disambiguation problem show better outcomes when IVFHWN is being used in place of Type 1 Fuzzy Hindi WordNet and classical Hindi WordNet.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"42 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139667272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi View Image Fusion Using Ensemble Deep Learning Algorithm for Mri and CT Images 利用集合深度学习算法实现 Mri 和 CT 图像的多视图图像融合
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-31 DOI: 10.1145/3640811
N. Thenmoezhi, B. Perumal, A. Lakshmi

Medical image fusions are crucial elements in image based health care diagnostics or therapies, and generically applications of computer visions. However, majority of existing methods suffer from noise distortion that affect the overall output. When pictures are distorted by noises, classical fusion techniques perform badly. Hence, fusion techniques that properly maintain information comprehensively from multiple faulty pictures need to be created. This work presents ESLOs (Enhanced Lion Swarm Optimizations) with EDL (Ensemble Deep Learning) to address the aforementioned issues. The primary steps in this study include image fusions, segmentation, noise reduction, feature extraction, picture classification, and feature selection.AMFs (Adaptive Median Filters) are first used for noise removal in sequence to enhance image quality by eliminating noises. The MRIs and CTS images are then segmented using the RKMC algorithm to separate the images into their component regions or objects. Images in black and white are divided into image. In the white image, the RKMC algorithm successfully considered the earlier tumour probability. The next step is feature extraction, which is accomplished by using the MPCA (Modified Principal Component Analysis) to draw out the most informative aspects of the images. Then, ELSOs algorithm is applied for optimal feature selection which is computed by best fitness values. After that, multi view image fusions of multi modal images derive lower, middle and higher level images contents. It is done by using DCNNs (Deep Convolution Neural Networks) and TAcGANs (Tissue-Aware conditional Generative Adversarial Networks) algorithm which fuses the multi view features and relevant image features and it is used for real time applications. The results of this study implies that proposed ELSO+EDL algorithm results in better performances in terms of higher values of accuracies, PSNR and lower RMSE, MAPE with faster executions when compared to other existing algorithms.

医学图像融合是基于图像的保健诊断或治疗以及计算机视觉应用的关键要素。然而,现有的大多数方法都存在噪声失真问题,影响了整体输出效果。当图片被噪声扭曲时,传统的融合技术就会表现不佳。因此,需要创建能从多张有问题的图片中全面适当地保留信息的融合技术。本作品提出了 ESLOs(增强型狮群优化)与 EDL(集合深度学习)来解决上述问题。本研究的主要步骤包括图像融合、分割、降噪、特征提取、图片分类和特征选择。首先使用 AMF(自适应中值滤波器)依次去除噪声,通过消除噪声来提高图像质量。然后使用 RKMC 算法对 MRI 和 CTS 图像进行分割,将图像分成不同的区域或对象。图像分为黑白两色。在白色图像中,RKMC 算法成功地考虑了早期肿瘤的概率。下一步是特征提取,通过使用 MPCA(修正主成分分析)来提取图像中信息量最大的部分。然后,应用 ELSOs 算法进行最佳特征选择,该算法由最佳适配值计算得出。然后,对多模态图像进行多视图图像融合,得出低层、中层和高层图像内容。这是通过使用 DCNNs(深度卷积神经网络)和 TAcGANs(组织感知条件生成对抗网络)算法完成的,该算法将多视图特征和相关图像特征融合在一起,并用于实时应用。这项研究的结果表明,与其他现有算法相比,拟议的 ELSO+EDL 算法在更高的精确度、PSNR 值和更低的 RMSE、MAPE 方面性能更好,执行速度更快。
{"title":"Multi View Image Fusion Using Ensemble Deep Learning Algorithm for Mri and CT Images","authors":"N. Thenmoezhi, B. Perumal, A. Lakshmi","doi":"10.1145/3640811","DOIUrl":"https://doi.org/10.1145/3640811","url":null,"abstract":"<p>Medical image fusions are crucial elements in image based health care diagnostics or therapies, and generically applications of computer visions. However, majority of existing methods suffer from noise distortion that affect the overall output. When pictures are distorted by noises, classical fusion techniques perform badly. Hence, fusion techniques that properly maintain information comprehensively from multiple faulty pictures need to be created. This work presents ESLOs (Enhanced Lion Swarm Optimizations) with EDL (Ensemble Deep Learning) to address the aforementioned issues. The primary steps in this study include image fusions, segmentation, noise reduction, feature extraction, picture classification, and feature selection.AMFs (Adaptive Median Filters) are first used for noise removal in sequence to enhance image quality by eliminating noises. The MRIs and CTS images are then segmented using the RKMC algorithm to separate the images into their component regions or objects. Images in black and white are divided into image. In the white image, the RKMC algorithm successfully considered the earlier tumour probability. The next step is feature extraction, which is accomplished by using the MPCA (Modified Principal Component Analysis) to draw out the most informative aspects of the images. Then, ELSOs algorithm is applied for optimal feature selection which is computed by best fitness values. After that, multi view image fusions of multi modal images derive lower, middle and higher level images contents. It is done by using DCNNs (Deep Convolution Neural Networks) and TAcGANs (Tissue-Aware conditional Generative Adversarial Networks) algorithm which fuses the multi view features and relevant image features and it is used for real time applications. The results of this study implies that proposed ELSO+EDL algorithm results in better performances in terms of higher values of accuracies, PSNR and lower RMSE, MAPE with faster executions when compared to other existing algorithms.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"283 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139646778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding the performance of AI algorithms in Text-Based Emotion Detection for Conversational Agents 了解人工智能算法在对话式代理基于文本的情感检测中的表现
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-31 DOI: 10.1145/3643133
Sheetal D. Kusal, Shruti G. Patil, Jyoti Choudrie, Ketan V. Kotecha

Current industry trends demand automation in every aspect, where machines could replace humans. Recent advancements in conversational agents have grabbed a lot of attention from industries, markets, and businesses. Building conversational agents that exhibit human communication characteristics is a need in today's marketplace. Thus, by accumulating emotions, we can build emotionally-aware conversational agents. Emotion detection in text-based dialogues has turned into a pivotal component of conversational agents, enhancing their ability to understand and respond to users' emotional states. This paper extensively compares various AI - techniques adapted to text-based emotion detection for conversational agents. This study covers a wide range of methods ranging from machine learning models to cutting-edge pre-trained models as well as deep learning models. The authors evaluate the performance of these techniques on the benchmark unbalanced topical chat and empathetic dialogue, balanced datasets. This paper offers an overview of the practical implications of emotion detection techniques in conversational systems and their impact on user response. The outcomes of this paper contribute to the ongoing development of empathetic conversational agents, emphasizing natural human-machine interactions.

当前的行业趋势要求各方面都实现自动化,机器可以取代人类。对话式代理的最新进展引起了各行业、市场和企业的广泛关注。当今的市场需要建立能展现人类交流特征的对话代理。因此,通过积累情感,我们可以建立具有情感意识的对话代理。基于文本的对话中的情感检测已成为会话代理的一个关键组成部分,它能增强会话代理理解和响应用户情感状态的能力。本文广泛比较了适用于对话代理基于文本的情感检测的各种人工智能技术。这项研究涵盖了从机器学习模型到前沿预训练模型以及深度学习模型等多种方法。作者评估了这些技术在基准非平衡主题聊天和移情对话平衡数据集上的性能。本文概述了情感检测技术在对话系统中的实际意义及其对用户响应的影响。本文的研究成果有助于当前开发移情对话代理,强调自然的人机交互。
{"title":"Understanding the performance of AI algorithms in Text-Based Emotion Detection for Conversational Agents","authors":"Sheetal D. Kusal, Shruti G. Patil, Jyoti Choudrie, Ketan V. Kotecha","doi":"10.1145/3643133","DOIUrl":"https://doi.org/10.1145/3643133","url":null,"abstract":"<p>Current industry trends demand automation in every aspect, where machines could replace humans. Recent advancements in conversational agents have grabbed a lot of attention from industries, markets, and businesses. Building conversational agents that exhibit human communication characteristics is a need in today's marketplace. Thus, by accumulating emotions, we can build emotionally-aware conversational agents. Emotion detection in text-based dialogues has turned into a pivotal component of conversational agents, enhancing their ability to understand and respond to users' emotional states. This paper extensively compares various AI - techniques adapted to text-based emotion detection for conversational agents. This study covers a wide range of methods ranging from machine learning models to cutting-edge pre-trained models as well as deep learning models. The authors evaluate the performance of these techniques on the benchmark unbalanced topical chat and empathetic dialogue, balanced datasets. This paper offers an overview of the practical implications of emotion detection techniques in conversational systems and their impact on user response. The outcomes of this paper contribute to the ongoing development of empathetic conversational agents, emphasizing natural human-machine interactions.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"27 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139646764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Human Emotion Recognition Based on Machine Learning Algorithms with low Resource Environment 低资源环境下基于机器学习算法的人类情感识别
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-26 DOI: 10.1145/3640340
Asha P., Hemamalini V., Poongodaia., Swapna N., Soujanya K. L. S., Vaishali Gaikwad (Mohite)

It is difficult to discover significant audio elements and conduct systematic comparison analyses when trying to automatically detect emotions in speech. In situations when it is desirable to reduce memory and processing constraints, this research deals with emotion recognition. One way to achieve this is by reducing the amount of features. In this study, propose "Active Feature Selection" (AFS) method and compares it against different state-of-the-art techniques. According to the results, smaller subsets of features than the complete feature set can produce accuracy that is comparable to or better than the full feature set. The memory and processing requirements of an emotion identification system will be reduced, which can minimise the hurdles to using health monitoring technology. The results show by using 696 characteristics, the AFS technique for emobase yields a Unweighted average recall (UAR) of 75.8%.

在试图自动检测语音中的情绪时,很难发现重要的音频元素并进行系统的对比分析。在希望减少记忆和处理限制的情况下,这项研究涉及情感识别。实现这一目标的方法之一是减少特征数量。本研究提出了 "主动特征选择"(AFS)方法,并将其与不同的先进技术进行了比较。结果表明,比完整特征集更小的特征子集所产生的准确率可与完整特征集相媲美,甚至更好。情绪识别系统对内存和处理的要求也会降低,这可以最大限度地减少使用健康监测技术的障碍。结果表明,通过使用 696 个特征,针对 emobase 的 AFS 技术的非加权平均召回率(UAR)为 75.8%。
{"title":"Human Emotion Recognition Based on Machine Learning Algorithms with low Resource Environment","authors":"Asha P., Hemamalini V., Poongodaia., Swapna N., Soujanya K. L. S., Vaishali Gaikwad (Mohite)","doi":"10.1145/3640340","DOIUrl":"https://doi.org/10.1145/3640340","url":null,"abstract":"<p>It is difficult to discover significant audio elements and conduct systematic comparison analyses when trying to automatically detect emotions in speech. In situations when it is desirable to reduce memory and processing constraints, this research deals with emotion recognition. One way to achieve this is by reducing the amount of features. In this study, propose \"Active Feature Selection\" (AFS) method and compares it against different state-of-the-art techniques. According to the results, smaller subsets of features than the complete feature set can produce accuracy that is comparable to or better than the full feature set. The memory and processing requirements of an emotion identification system will be reduced, which can minimise the hurdles to using health monitoring technology. The results show by using 696 characteristics, the AFS technique for emobase yields a Unweighted average recall (UAR) of 75.8%.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"64 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139580337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fuzzified Deep Learning based Forgery Detection of Signatures in the Healthcare Mission Records 基于模糊化深度学习的医疗任务记录签名伪造检测
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-24 DOI: 10.1145/3641818
Ishu Priya, Nisha Chaurasia, Ashutosh Kumar Singh, Nakul Mehta, Abhishek Singh Kilak, Ahmed Alkhayyat

In an era subjected to digital solutions, handwritten signatures continue playing a crucial role in identity verification and document authentication. These signatures, a form of bio-metric verification, are unique to every individual, serving as a primitive method for confirming identity and ensuring security of an individual. Signatures, apart from being a means of personal authentication, are often considered a cornerstone in the validation of critical documents and processes, especially within the healthcare sector. In healthcare missions, particularly in the regions that are underdeveloped, hand-written records persist as the primary mode of documentation. The credibility of these handwritten documents hinges on the authenticity of the accompanying signatures, making signature verification a paramount safeguard for the integrity and security of medical information. Nonetheless, traditional offline methods of signature identification can be time-consuming and inefficient, particularly while dealing with a massive volume of documents. This arises the evident need for automated signature verification systems. Our research introduces an innovative signature verification system which synthesizes the strengths of fuzzy logic and CNN (Convolutional Neural Networks) to deliver precise and efficient signature verification. Leveraging the capabilities of Fuzzy Logic for feature representation and CNNs for discriminative learning, our proposed hybrid model offers a compelling solution. Through rigorous training, spanning a mere 28 epochs, our hybrid model exhibits remarkable performance by attaining a training accuracy of 91.29% and a test accuracy of 88.47%, underscoring its robust generalization capacity. In an era of evolving security requirements and the persistent relevance of handwritten signatures, our research links the disparity between tradition and modernity.

在数字解决方案大行其道的时代,手写签名仍在身份验证和文件认证中发挥着至关重要的作用。这些签名是生物计量验证的一种形式,对每个人来说都是独一无二的,是确认身份和确保个人安全的原始方法。签名除了是个人身份认证的一种手段外,还经常被视为验证重要文件和流程的基石,尤其是在医疗保健领域。在医疗保健任务中,尤其是在欠发达地区,手写记录一直是主要的文件记录方式。这些手写文件的可信度取决于所附签名的真实性,因此签名验证是医疗信息完整性和安全性的重要保障。然而,传统的离线签名识别方法耗时且效率低下,尤其是在处理大量文件时。因此,对自动签名验证系统的需求显而易见。我们的研究引入了一种创新的签名验证系统,它综合了模糊逻辑和卷积神经网络(CNN)的优势,可提供精确高效的签名验证。利用模糊逻辑在特征表示方面的能力和 CNN 在判别学习方面的能力,我们提出的混合模型提供了令人信服的解决方案。我们的混合模型经过仅 28 个历时的严格训练,取得了 91.29% 的训练准确率和 88.47% 的测试准确率,表现出了卓越的性能,凸显了其强大的泛化能力。在不断发展的安全要求和手写签名的持续相关性的时代,我们的研究将传统与现代之间的差距联系起来。
{"title":"Fuzzified Deep Learning based Forgery Detection of Signatures in the Healthcare Mission Records","authors":"Ishu Priya, Nisha Chaurasia, Ashutosh Kumar Singh, Nakul Mehta, Abhishek Singh Kilak, Ahmed Alkhayyat","doi":"10.1145/3641818","DOIUrl":"https://doi.org/10.1145/3641818","url":null,"abstract":"<p>In an era subjected to digital solutions, handwritten signatures continue playing a crucial role in identity verification and document authentication. These signatures, a form of bio-metric verification, are unique to every individual, serving as a primitive method for confirming identity and ensuring security of an individual. Signatures, apart from being a means of personal authentication, are often considered a cornerstone in the validation of critical documents and processes, especially within the healthcare sector. In healthcare missions, particularly in the regions that are underdeveloped, hand-written records persist as the primary mode of documentation. The credibility of these handwritten documents hinges on the authenticity of the accompanying signatures, making signature verification a paramount safeguard for the integrity and security of medical information. Nonetheless, traditional offline methods of signature identification can be time-consuming and inefficient, particularly while dealing with a massive volume of documents. This arises the evident need for automated signature verification systems. Our research introduces an innovative signature verification system which synthesizes the strengths of fuzzy logic and CNN (Convolutional Neural Networks) to deliver precise and efficient signature verification. Leveraging the capabilities of Fuzzy Logic for feature representation and CNNs for discriminative learning, our proposed hybrid model offers a compelling solution. Through rigorous training, spanning a mere 28 epochs, our hybrid model exhibits remarkable performance by attaining a training accuracy of 91.29% and a test accuracy of 88.47%, underscoring its robust generalization capacity. In an era of evolving security requirements and the persistent relevance of handwritten signatures, our research links the disparity between tradition and modernity.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"39 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139554871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MODELLING A NOVEL APPROACH FOR EMOTION RECOGNITION USING LEARNING AND NATURAL LANGUAGE PROCESSING 利用学习和自然语言处理建立情感识别新方法模型
IF 2 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-22 DOI: 10.1145/3641851
Lakshmi Lalitha V., Dinesh Kumar Anguraj

Various facts, including politics, entertainment, industry, and research fields, are connected to analysing the audience's emotional. Syntactic Analysis (SA) is a Natural Language Processing (NLP) concept that uses statistical and lexical forms as well as learning techniques to forecast how different types of content in social media will express the audience's neutral, positive, and negative emotions. The lack of an adequate tool to quantify the characteristics and independent text for assessing the primary audience emotion from the available online social media dataset. The focus of this research is on modeling a cutting-edge method for decoding the connectivity among social media texts and assessing audience emotions. Here, a novel dense layer graph model (DLG-TF) for textual feature analysis is used to analyze the relevant connectedness inside the complex media environment to forecast emotions. The information from the social media dataset is extracted using some popular convolution network models, and the predictions are made by examining the textual properties. The experimental results show that, when compared to different standard emotions, the proposed DLG-TF model accurately predicts a greater number of possible emotions. The macro-average of baseline is 58%, the affective is 55%, the crawl is 55% and the ultra-dense is 59% respectively. The feature analysis comparison of baseline, affective, crawl, ultra-dense and DLG-TF using the unsupervised model based on EmoTweet gives the precision, recall and F1-score of the anticipated model are explained. The micro and macro average based on these parameters are compared and analyzed. The macro-average of baseline is 47%, the affective is 46%, the crawl is 50% and the ultra-dense is 85% respectively. It makes precise predictions using the social media dataset that is readily available. A few criteria, including accuracy, recall, precision, and F-measure, are assessed and contrasted with alternative methods.

包括政治、娱乐、工业和研究领域在内的各种事实都与分析受众情绪有关。句法分析(SA)是一个自然语言处理(NLP)概念,它利用统计和词汇形式以及学习技术来预测社交媒体中不同类型的内容将如何表达受众的中性、积极和消极情绪。由于缺乏适当的工具来量化现有在线社交媒体数据集中用于评估主要受众情绪的特征和独立文本,因此,本研究的重点是对社交媒体中的受众情绪进行建模。本研究的重点是建立一个前沿方法模型,用于解码社交媒体文本之间的连接性并评估受众情绪。在这里,一种用于文本特征分析的新型密集层图模型(DLG-TF)被用来分析复杂媒体环境中的相关连通性,从而预测情感。使用一些流行的卷积网络模型从社交媒体数据集中提取信息,并通过研究文本属性进行预测。实验结果表明,与不同的标准情绪相比,所提出的 DLG-TF 模型能准确预测更多可能的情绪。基线的宏观平均值为 58%,情感的宏观平均值为 55%,爬行的宏观平均值为 55%,超密集的宏观平均值为 59%。使用基于 EmoTweet 的无监督模型对基线、情感、爬行、超密度和 DLG-TF 进行的特征分析比较给出了预期模型的精确度、召回率和 F1 分数。对基于这些参数的微观和宏观平均值进行了比较和分析。基线宏观平均值为 47%,情感平均值为 46%,爬行平均值为 50%,超密集平均值为 85%。它利用现成的社交媒体数据集进行精确预测。评估了准确率、召回率、精确度和 F 测量等几项标准,并与其他方法进行了对比。
{"title":"MODELLING A NOVEL APPROACH FOR EMOTION RECOGNITION USING LEARNING AND NATURAL LANGUAGE PROCESSING","authors":"Lakshmi Lalitha V., Dinesh Kumar Anguraj","doi":"10.1145/3641851","DOIUrl":"https://doi.org/10.1145/3641851","url":null,"abstract":"<p>Various facts, including politics, entertainment, industry, and research fields, are connected to analysing the audience's emotional. Syntactic Analysis (SA) is a Natural Language Processing (NLP) concept that uses statistical and lexical forms as well as learning techniques to forecast how different types of content in social media will express the audience's neutral, positive, and negative emotions. The lack of an adequate tool to quantify the characteristics and independent text for assessing the primary audience emotion from the available online social media dataset. The focus of this research is on modeling a cutting-edge method for decoding the connectivity among social media texts and assessing audience emotions. Here, a novel dense layer graph model (DLG-TF) for textual feature analysis is used to analyze the relevant connectedness inside the complex media environment to forecast emotions. The information from the social media dataset is extracted using some popular convolution network models, and the predictions are made by examining the textual properties. The experimental results show that, when compared to different standard emotions, the proposed DLG-TF model accurately predicts a greater number of possible emotions. The macro-average of baseline is 58%, the affective is 55%, the crawl is 55% and the ultra-dense is 59% respectively. The feature analysis comparison of baseline, affective, crawl, ultra-dense and DLG-TF using the unsupervised model based on EmoTweet gives the precision, recall and F1-score of the anticipated model are explained. The micro and macro average based on these parameters are compared and analyzed. The macro-average of baseline is 47%, the affective is 46%, the crawl is 50% and the ultra-dense is 85% respectively. It makes precise predictions using the social media dataset that is readily available. A few criteria, including accuracy, recall, precision, and F-measure, are assessed and contrasted with alternative methods.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"3 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139554971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ACM Transactions on Asian and Low-Resource Language Information Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1