首页 > 最新文献

ACM Transactions on Asian and Low-Resource Language Information Processing最新文献

英文 中文
Performance of Binarization Algorithms on Tamizhi Inscription Images: An Analysis Tamizhi 铭文图像的二值化算法性能:分析
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-08 DOI: 10.1145/3656583
Monisha Munivel, V S Felix Enigo

Binarization of Tamizhi (Tamil-Brahmi) inscription images are highly challenging as it is captured from very old stone inscriptions that exists around 3rd century BCE in India. The difficulty is due to the degradation of these inscriptions by environmental factors and human negligence over ages. Though many works have been carried out in the binarization of inscription images, very few research was performed for inscription images and no work has been reported for binarization of inscriptions inscribed on irregular medium. The findings of the analysis hold true to all writings that are carved in irregular background. This paper reviews the performance of various binarization techniques on Tamizhi inscription images. Since no previous work was performed, we have applied the existing binarization algorithms on Tamizhi inscription images and analyzed the performance of these algorithms with proper reasoning. In future, we believe that this reasoning on the results will help a new researcher, to adapt or combine or devise new binarization techniques.

Tamizhi(泰米尔-婆罗米)碑文图像的二值化具有很高的挑战性,因为它是从印度公元前 3 世纪左右的非常古老的石刻中采集的。困难的原因在于这些碑文因环境因素和人类长期疏忽而退化。虽然已有许多研究对碑文图像进行了二值化处理,但针对碑文图像的研究却寥寥无几,而且还没有关于对刻写在不规则介质上的碑文进行二值化处理的研究报告。分析结果适用于所有刻在不规则背景上的文字。本文回顾了各种二值化技术在 Tamizhi 碑文图像上的表现。由于之前没有相关工作,我们将现有的二值化算法应用于 Tamizhi 碑文图像,并通过适当的推理分析了这些算法的性能。今后,我们相信这种对结果的推理将有助于新的研究人员调整、组合或设计新的二值化技术。
{"title":"Performance of Binarization Algorithms on Tamizhi Inscription Images: An Analysis","authors":"Monisha Munivel, V S Felix Enigo","doi":"10.1145/3656583","DOIUrl":"https://doi.org/10.1145/3656583","url":null,"abstract":"<p>Binarization of Tamizhi (Tamil-Brahmi) inscription images are highly challenging as it is captured from very old stone inscriptions that exists around 3rd century BCE in India. The difficulty is due to the degradation of these inscriptions by environmental factors and human negligence over ages. Though many works have been carried out in the binarization of inscription images, very few research was performed for inscription images and no work has been reported for binarization of inscriptions inscribed on irregular medium. The findings of the analysis hold true to all writings that are carved in irregular background. This paper reviews the performance of various binarization techniques on Tamizhi inscription images. Since no previous work was performed, we have applied the existing binarization algorithms on Tamizhi inscription images and analyzed the performance of these algorithms with proper reasoning. In future, we believe that this reasoning on the results will help a new researcher, to adapt or combine or devise new binarization techniques.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140594746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Extractive Text Summarization using Multiple Linguistic Features 利用多种语言特征自动提取文本摘要
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-08 DOI: 10.1145/3656471
Pooja Gupta, Swati Nigam, Rajiv Singh

Automatic text summarization (ATS) provides a summary of distinct categories of information using natural language processing (NLP). Low-resource languages like Hindi have restricted applications of these techniques. This study proposes a method for automatically generating summaries of Hindi documents using extractive technique. The approach retrieves pertinent sentences from the source documents by employing multiple linguistic features and machine learning (ML) using maximum likelihood estimation (MLE) and maximum entropy (ME). We conducted pre-processing on the input documents, such as eliminating Hindi stop words and stemming. We have obtained 15 linguistic feature scores from each document to identify the phrases with high scores for summary generation. We have performed experiments over BBC News articles, CNN News, DUC 2004, Hindi Text Short Summarization Corpus, Indian Language News Text Summarization Corpus, and Wikipedia Articles for the proposed text summarizer. The Hindi Text Short Summarization Corpus and Indian Language News Text Summarization Corpus datasets are in Hindi, whereas BBC News articles, CNN News, and the DUC 2004 datasets have been translated into Hindi using Google, Microsoft Bing, and Systran translators for experiments. The summarization results have been calculated and shown for Hindi as well as for English to compare the performance of a low and rich-resource language. Multiple ROUGE metrics, along with precision, recall, and F-measure, have been used for the evaluation, which shows the better performance of the proposed method with multiple ROUGE scores. We compare the proposed method with the supervised and unsupervised machine learning methodologies, including support vector machine (SVM), Naive Bayes (NB), decision tree (DT), latent semantic analysis (LSA), latent Dirichlet allocation (LDA), and K-means clustering, and it was found that the proposed method outperforms these methods.

自动文本摘要(ATS)利用自然语言处理(NLP)对不同类别的信息进行摘要。印地语等低资源语言限制了这些技术的应用。本研究提出了一种使用提取技术自动生成印地语文档摘要的方法。该方法通过使用多种语言特征以及最大似然估计(MLE)和最大熵(ME)的机器学习(ML),从源文档中检索相关句子。我们对输入文档进行了预处理,如消除印地语停滞词和词干。我们从每篇文档中获取了 15 个语言特征分数,以识别出分数较高的短语,从而生成摘要。我们对 BBC 新闻报道、CNN 新闻、DUC 2004、印地语文本简短摘要语料库、印度语新闻文本摘要语料库和维基百科文章进行了实验,以验证所提议的文本摘要器。印地语文本简短摘要语料库和印度语新闻文本摘要语料库的数据集是印地语的,而 BBC News 文章、CNN News 和 DUC 2004 数据集已使用 Google、Microsoft Bing 和 Systran 翻译器翻译成印地语进行实验。计算并显示了印地语和英语的摘要结果,以比较低资源语言和丰富资源语言的性能。在评估中使用了多个 ROUGE 指标以及精确度、召回率和 F-measure,结果表明使用多个 ROUGE 分数的拟议方法性能更佳。我们将提出的方法与支持向量机 (SVM)、奈夫贝叶斯 (NB)、决策树 (DT)、潜在语义分析 (LSA)、潜在 Dirichlet 分配 (LDA) 和 K-means 聚类等有监督和无监督机器学习方法进行了比较,发现提出的方法优于这些方法。
{"title":"Automatic Extractive Text Summarization using Multiple Linguistic Features","authors":"Pooja Gupta, Swati Nigam, Rajiv Singh","doi":"10.1145/3656471","DOIUrl":"https://doi.org/10.1145/3656471","url":null,"abstract":"<p>Automatic text summarization (ATS) provides a summary of distinct categories of information using natural language processing (NLP). Low-resource languages like Hindi have restricted applications of these techniques. This study proposes a method for automatically generating summaries of Hindi documents using extractive technique. The approach retrieves pertinent sentences from the source documents by employing multiple linguistic features and machine learning (ML) using maximum likelihood estimation (MLE) and maximum entropy (ME). We conducted pre-processing on the input documents, such as eliminating Hindi stop words and stemming. We have obtained 15 linguistic feature scores from each document to identify the phrases with high scores for summary generation. We have performed experiments over BBC News articles, CNN News, DUC 2004, Hindi Text Short Summarization Corpus, Indian Language News Text Summarization Corpus, and Wikipedia Articles for the proposed text summarizer. The Hindi Text Short Summarization Corpus and Indian Language News Text Summarization Corpus datasets are in Hindi, whereas BBC News articles, CNN News, and the DUC 2004 datasets have been translated into Hindi using Google, Microsoft Bing, and Systran translators for experiments. The summarization results have been calculated and shown for Hindi as well as for English to compare the performance of a low and rich-resource language. Multiple ROUGE metrics, along with precision, recall, and F-measure, have been used for the evaluation, which shows the better performance of the proposed method with multiple ROUGE scores. We compare the proposed method with the supervised and unsupervised machine learning methodologies, including support vector machine (SVM), Naive Bayes (NB), decision tree (DT), latent semantic analysis (LSA), latent Dirichlet allocation (LDA), and K-means clustering, and it was found that the proposed method outperforms these methods.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140594619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SUSTEM: An Improved Rule-Based Sundanese Stemmer SUSTEM:基于规则的改进型巽他语词根生成器
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-05 DOI: 10.1145/3656342
Irwan Setiawan, Hung-Yu Kao

Current Sundanese stemmers either ignore reduplication words or define rules to handle only affixes. There is a significant amount of reduplication words in the Sundanese language. Because of that, it is impossible to achieve superior stemming precision in the Sundanese language without addressing reduplication words. This paper presents an improved stemmer for the Sundanese language, which handles affixed and reduplicated words. With a Sundanese root word list, we use a rules-based stemming technique. In our approach, all stems produced by the affixes removal or normalization processes are added to the stem list. Using a stem list can help increase stemmer accuracy by reducing stemming errors caused by affix removal sequence errors or morphological issues. The current Sundanese language stemmer, RBSS, was used as a comparison. Two datasets with 8218 unique affixed words and reduplication words were evaluated. The results show that our stemmer's strength and accuracy have improved noticeably. The use of stem list and word reduplication rules improved our stemmer's affixed type recognition and allowed us to achieve up to 99.30% accuracy.

目前的巽他语词干生成器要么忽略重合词,要么只定义处理词缀的规则。巽他语中有大量的重迭词。因此,如果不处理重合词,就不可能在巽他语中实现卓越的词干处理精度。本文介绍了一种改进的巽他语词干生成器,它能处理后缀词和重复词。通过巽他语词根列表,我们使用了基于规则的词干处理技术。在我们的方法中,由词缀去除或规范化过程产生的所有词干都被添加到词干列表中。使用词干列表可以减少因词缀去除顺序错误或形态问题造成的词干错误,从而有助于提高词干生成器的准确性。目前的巽他语干词表 RBSS 被用作对比。两个数据集包含 8218 个独特的词缀词和重复词,我们对这两个数据集进行了评估。结果表明,我们的干词器的强度和准确性都有明显提高。使用词干列表和单词重迭规则提高了干译员的词缀类型识别能力,使我们的准确率达到 99.30%。
{"title":"SUSTEM: An Improved Rule-Based Sundanese Stemmer","authors":"Irwan Setiawan, Hung-Yu Kao","doi":"10.1145/3656342","DOIUrl":"https://doi.org/10.1145/3656342","url":null,"abstract":"<p>Current Sundanese stemmers either ignore reduplication words or define rules to handle only affixes. There is a significant amount of reduplication words in the Sundanese language. Because of that, it is impossible to achieve superior stemming precision in the Sundanese language without addressing reduplication words. This paper presents an improved stemmer for the Sundanese language, which handles affixed and reduplicated words. With a Sundanese root word list, we use a rules-based stemming technique. In our approach, all stems produced by the affixes removal or normalization processes are added to the stem list. Using a stem list can help increase stemmer accuracy by reducing stemming errors caused by affix removal sequence errors or morphological issues. The current Sundanese language stemmer, RBSS, was used as a comparison. Two datasets with 8218 unique affixed words and reduplication words were evaluated. The results show that our stemmer's strength and accuracy have improved noticeably. The use of stem list and word reduplication rules improved our stemmer's affixed type recognition and allowed us to achieve up to 99.30% accuracy.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140594747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph4IUR: Incomplete Utterance Rewriting with Semantic Graph Graph4IUR:利用语义图重写不完整语句
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-04 DOI: 10.1145/3653301
Zipeng Gao, Jinke Wang, Tong Xu, Zhefeng Wang, Yu Yang, Jia Su, Enhong Chen

Utterance rewriting aims to identify and supply the omitted information in human conversation, which further enables the downstream task to understand conversations more comprehensively. Recently, sequence edit methods, which leverage the overlap between two sentences, have been widely applied to narrow the search space confronted by the previous linear generation methods. However, these methods ignore the relationship between linguistic elements in the conversation, which reflects how the knowledge and thoughts are organized in human communication. In this case, although most of the content in rewritten sentences can be found in the context, we found that some connecting words expressing relationships are often missing, which results in the out-of-context problem for the previous sentence edit method. To that end, in this paper, we propose a new semantic Graph-based Incomplete Utterance Rewriting (Graph4IUR) framework, which takes the semantic graph to depict the relationship between linguistic elements and captures out-of-context words. Specifically, we adopt the Abstract Meaning Representation (AMR) [4] graph as the basic sentence-to-graph method to depict the dialogue from the graph perspective, which could well represent the high-level semantics relationships of sentences. Along this line, we further adapt the sentence editing models to rewrite without changing the sentence architecture, which brings a restriction to exploring the overlap part of the current and rewritten sentences in the IUR task. Extensive experimental results indicate that our Graph4IUR framework can effectively alleviate the out-of-context problem and improve the performance of the previous edit-based methods in the IUR task.

语句重写的目的是识别和提供人类会话中遗漏的信息,从而进一步帮助下游任务更全面地理解会话。最近,序列编辑方法得到了广泛应用,这种方法利用了两个句子之间的重叠,缩小了以往线性生成方法所面临的搜索空间。然而,这些方法忽略了会话中语言元素之间的关系,而这种关系反映了人类交流中知识和思想的组织方式。在这种情况下,虽然改写句子中的大部分内容都能在上下文中找到,但我们发现一些表达关系的连接词往往缺失,这就造成了以往句子编辑方法的断章取义问题。为此,我们在本文中提出了一种新的基于语义图的不完整语篇重写(Graph4IUR)框架,它利用语义图来描绘语言元素之间的关系,并捕捉断章取义的词语。具体来说,我们采用抽象意义表示(AMR)[4] 图作为句子到图的基本方法,从图的角度来描绘对话,这可以很好地表示句子的高层语义关系。沿着这一思路,我们进一步调整了句子编辑模型,在不改变句子结构的情况下进行重写,这就限制了在 IUR 任务中探索当前句子和重写句子的重叠部分。广泛的实验结果表明,我们的 Graph4IUR 框架可以有效缓解断章取义问题,并提高以往基于编辑的方法在 IUR 任务中的性能。
{"title":"Graph4IUR: Incomplete Utterance Rewriting with Semantic Graph","authors":"Zipeng Gao, Jinke Wang, Tong Xu, Zhefeng Wang, Yu Yang, Jia Su, Enhong Chen","doi":"10.1145/3653301","DOIUrl":"https://doi.org/10.1145/3653301","url":null,"abstract":"<p>Utterance rewriting aims to identify and supply the omitted information in human conversation, which further enables the downstream task to understand conversations more comprehensively. Recently, sequence edit methods, which leverage the overlap between two sentences, have been widely applied to narrow the search space confronted by the previous linear generation methods. However, these methods ignore the relationship between linguistic elements in the conversation, which reflects how the knowledge and thoughts are organized in human communication. In this case, although most of the content in rewritten sentences can be found in the context, we found that some connecting words expressing relationships are often missing, which results in the out-of-context problem for the previous sentence edit method. To that end, in this paper, we propose a new semantic Graph-based Incomplete Utterance Rewriting (Graph4IUR) framework, which takes the semantic graph to depict the relationship between linguistic elements and captures out-of-context words. Specifically, we adopt the Abstract Meaning Representation (AMR) [4] graph as the basic sentence-to-graph method to depict the dialogue from the graph perspective, which could well represent the high-level semantics relationships of sentences. Along this line, we further adapt the sentence editing models to rewrite without changing the sentence architecture, which brings a restriction to exploring the overlap part of the current and rewritten sentences in the IUR task. Extensive experimental results indicate that our Graph4IUR framework can effectively alleviate the out-of-context problem and improve the performance of the previous edit-based methods in the IUR task.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140594621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MIMIC: Misogyny Identification in Multimodal Internet Content in Hindi-English Code-Mixed Language MIMIC:印地语-英语代码混合语言多模态互联网内容中的厌女症识别
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-04 DOI: 10.1145/3656169
Aakash Singh, Deepawali Sharma, Vivek Kumar Singh

Over the years, social media has emerged as one of the most popular platforms where people express their views and share thoughts about various aspects. The social media content now includes a variety of components such as text, images, videos etc. One type of interest is memes, which often combine text and images. It is relevant to mention here that, social media being an unregulated platform, sometimes also has instances of discriminatory, offensive and hateful content being posted. Such content adversely affects the online well-being of the users. Therefore, it is very important to develop computational models to automatically detect such content so that appropriate corrective action can be taken. Accordingly, there have been research efforts on automatic detection of such content focused mainly on the texts. However, the fusion of multimodal data (as in memes) creates various challenges in developing computational models that can handle such data, more so in the case of low-resource languages. Among such challenges, the lack of suitable datasets for developing computational models for handling memes in low-resource languages is a major problem. This work attempts to bridge the research gap by providing a large-sized curated dataset comprising 5,054 memes in Hindi-English code-mixed language, which are manually annotated by three independent annotators. It comprises two subtasks: (i) Subtask-1 (Binary classification involving tagging a meme as misogynous or non-misogynous), and (ii) Subtask-2 (multi-label classification of memes into different categories). The data quality is evaluated by computing Krippendorff's alpha. Different computational models are then applied on the data in three settings: text-only, image-only, and multimodal models using fusion techniques. The results show that the proposed multimodal method using the fusion technique may be the preferred choice for the identification of misogyny in multimodal Internet content and that the dataset is suitable for advancing research and development in the area.

多年来,社交媒体已成为人们表达观点和分享各方面想法的最流行平台之一。现在,社交媒体的内容包括文字、图片、视频等多种形式。人们感兴趣的一种类型是 "备忘录",它通常将文字和图片结合在一起。值得一提的是,社交媒体作为一个不受监管的平台,有时也会出现发布歧视性、攻击性和仇恨性内容的情况。这些内容会对用户的在线福祉产生不利影响。因此,开发自动检测此类内容的计算模型非常重要,以便采取适当的纠正措施。因此,自动检测此类内容的研究工作主要集中在文本方面。然而,多模态数据的融合(如备忘录中的数据)给开发可处理此类数据的计算模型带来了各种挑战,对于低资源语言来说更是如此。在这些挑战中,缺乏合适的数据集来开发处理低资源语言中memes的计算模型是一个主要问题。这项工作试图通过提供一个由 5,054 个印地语-英语混合语代码组成的大型数据集来弥补这一研究空白,这些数据集由三个独立的注释者手动注释。它由两个子任务组成:(i) 子任务-1(二元分类,涉及将备忘录标记为厌恶或非厌恶)和 (ii) 子任务-2(将备忘录分为不同类别的多标签分类)。数据质量通过计算克里彭多夫α进行评估。然后在三种情况下对数据应用不同的计算模型:纯文本模型、纯图像模型和使用融合技术的多模态模型。结果表明,所提出的使用融合技术的多模态方法可能是识别多模态互联网内容中厌女症的首选,而且该数据集适合用于推进该领域的研究和开发。
{"title":"MIMIC: Misogyny Identification in Multimodal Internet Content in Hindi-English Code-Mixed Language","authors":"Aakash Singh, Deepawali Sharma, Vivek Kumar Singh","doi":"10.1145/3656169","DOIUrl":"https://doi.org/10.1145/3656169","url":null,"abstract":"<p>Over the years, social media has emerged as one of the most popular platforms where people express their views and share thoughts about various aspects. The social media content now includes a variety of components such as text, images, videos etc. One type of interest is memes, which often combine text and images. It is relevant to mention here that, social media being an unregulated platform, sometimes also has instances of discriminatory, offensive and hateful content being posted. Such content adversely affects the online well-being of the users. Therefore, it is very important to develop computational models to automatically detect such content so that appropriate corrective action can be taken. Accordingly, there have been research efforts on automatic detection of such content focused mainly on the texts. However, the fusion of multimodal data (as in memes) creates various challenges in developing computational models that can handle such data, more so in the case of low-resource languages. Among such challenges, the lack of suitable datasets for developing computational models for handling memes in low-resource languages is a major problem. This work attempts to bridge the research gap by providing a large-sized curated dataset comprising 5,054 memes in Hindi-English code-mixed language, which are manually annotated by three independent annotators. It comprises two subtasks: (i) Subtask-1 (Binary classification involving tagging a meme as misogynous or non-misogynous), and (ii) Subtask-2 (multi-label classification of memes into different categories). The data quality is evaluated by computing Krippendorff's alpha. Different computational models are then applied on the data in three settings: text-only, image-only, and multimodal models using fusion techniques. The results show that the proposed multimodal method using the fusion technique may be the preferred choice for the identification of misogyny in multimodal Internet content and that the dataset is suitable for advancing research and development in the area.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140594755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Student's Emotion Recognition using Multimodality and Deep Learning 利用多模态和深度学习识别学生情绪
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-01 DOI: 10.1145/3654797
M. Kalaiyarasi, B. V. V. Siva Prasad, Janjhyam Venkata Naga Ramesh, Ravindra Kumar Kushwaha, Ruchi Patel, Balajee J

The goal of emotion detection is to find and recognise emotions in text, speech, gestures, facial expressions, and more. This paper proposes an effective multimodal emotion recognition system based on facial expressions, sentence-level text, and voice. Using public datasets, we examine face expression image classification and feature extraction. The Tri-modal fusion is used to integrate the findings and to provide the final emotion. The proposed method has been verified in classroom students, and the feelings correlate with their performance. This method categorizes students' expressions into seven emotions: happy, surprise, sad, fear, disgust, anger, and contempt. Compared to the unimodal models, the suggested multimodal network design may reach up to 65% accuracy. The proposed method can detect negative feelings such as boredom or loss of interest in the learning environment.

情感检测的目标是发现并识别文本、语音、手势、面部表情等中的情感。本文提出了一种基于面部表情、句子级文本和语音的有效多模态情感识别系统。我们利用公共数据集研究了面部表情图像分类和特征提取。三模态融合用于整合研究结果并提供最终情绪。所提出的方法已在班级学生中得到验证,其情感与学生的表现相关。该方法将学生的表情分为七种情绪:快乐、惊讶、悲伤、恐惧、厌恶、愤怒和蔑视。与单模态模型相比,建议的多模态网络设计的准确率可达 65%。建议的方法可以检测出学习环境中的负面情绪,如无聊或失去兴趣。
{"title":"Student's Emotion Recognition using Multimodality and Deep Learning","authors":"M. Kalaiyarasi, B. V. V. Siva Prasad, Janjhyam Venkata Naga Ramesh, Ravindra Kumar Kushwaha, Ruchi Patel, Balajee J","doi":"10.1145/3654797","DOIUrl":"https://doi.org/10.1145/3654797","url":null,"abstract":"<p>The goal of emotion detection is to find and recognise emotions in text, speech, gestures, facial expressions, and more. This paper proposes an effective multimodal emotion recognition system based on facial expressions, sentence-level text, and voice. Using public datasets, we examine face expression image classification and feature extraction. The Tri-modal fusion is used to integrate the findings and to provide the final emotion. The proposed method has been verified in classroom students, and the feelings correlate with their performance. This method categorizes students' expressions into seven emotions: happy, surprise, sad, fear, disgust, anger, and contempt. Compared to the unimodal models, the suggested multimodal network design may reach up to 65% accuracy. The proposed method can detect negative feelings such as boredom or loss of interest in the learning environment.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140594618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cleansing Jewel: A Neural Spelling Correction Model Built On Google OCR-ed Tibetan Manuscripts 净化宝石:基于谷歌 OCR 编辑的藏文手稿的神经拼写校正模型
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-03-30 DOI: 10.1145/3654811
Queenie Luo, Yung-Sung Chuang

Scholars in the humanities heavily rely on ancient manuscripts to study history, religion, and socio-political structures of the past. Significant efforts have been devoted to digitizing these precious manuscripts using OCR technology. However, most manuscripts have been blemished over the centuries, making it unrealistic for OCR programs to accurately capture faded characters. This work presents the Transformer + Confidence Score mechanism architecture for post-processing Google’s Tibetan OCR-ed outputs. According to the Loss and Character Error Rate metrics, our Transformer + Confidence Score mechanism architecture proves superior to the Transformer, LSTM-to-LSTM, and GRU-to-GRU architectures. Our method can be adapted to any language dealing with post-processing OCR outputs.

人文学科的学者非常依赖古代手稿来研究历史、宗教和过去的社会政治结构。利用 OCR 技术对这些珍贵的手稿进行数字化处理是一项艰巨的任务。然而,大多数手稿在几个世纪的时间里都已褪色,因此 OCR 程序无法准确捕捉褪色的字符。本作品提出了用于谷歌藏文 OCR 后处理的 Transformer + Confidence Score 机制架构。根据损失率和字符错误率指标,我们的变换器+置信分机制架构证明优于变换器、LSTM-to-LSTM 和 GRU-to-GRU 架构。我们的方法可适用于处理 OCR 输出后处理的任何语言。
{"title":"Cleansing Jewel: A Neural Spelling Correction Model Built On Google OCR-ed Tibetan Manuscripts","authors":"Queenie Luo, Yung-Sung Chuang","doi":"10.1145/3654811","DOIUrl":"https://doi.org/10.1145/3654811","url":null,"abstract":"<p>Scholars in the humanities heavily rely on ancient manuscripts to study history, religion, and socio-political structures of the past. Significant efforts have been devoted to digitizing these precious manuscripts using OCR technology. However, most manuscripts have been blemished over the centuries, making it unrealistic for OCR programs to accurately capture faded characters. This work presents the Transformer + Confidence Score mechanism architecture for post-processing Google’s Tibetan OCR-ed outputs. According to the Loss and Character Error Rate metrics, our Transformer + Confidence Score mechanism architecture proves superior to the Transformer, LSTM-to-LSTM, and GRU-to-GRU architectures. Our method can be adapted to any language dealing with post-processing OCR outputs.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140594744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MRMI-TTS: Multi-reference audios and Mutual Information Driven Zero-shot Voice cloning MRMI-TTS:多参考音频和互信息驱动的零镜头语音克隆
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-03-30 DOI: 10.1145/3649501
Yiting Chen, Wanting Li, Buzhou Tang
Voice cloning in text-to-speech (TTS) is the process of replicating the voice of a target speaker with limited data. Among various voice cloning techniques, this paper focuses on zero-shot voice cloning. Although existing TTS models can generate high-quality speech for seen speakers, cloning the voice of an unseen speaker remains a challenging task. The key aspect of zero-shot voice cloning is to obtain a speaker embedding from the target speaker. Previous works have used a speaker encoder to obtain a fixed-size speaker embedding from a single reference audio unsupervised, but they suffer from insufficient speaker information and content information leakage in speaker embedding.To address these issues, this paper proposes MRMI-TTS, a FastSpeech2-based framework that uses speaker embedding as a conditioning variable to provide speaker information. The MRMI-TTS extracts speaker embedding and content embedding from multi-reference audios using a speaker encoder and a content encoder. To obtain sufficient speaker information, multi-reference audios are selected based on sentence similarity. The proposed model applies mutual information minimization on the two embeddings to remove entangled information within each embedding.Experiments on the public English dataset VCTK show that our method can improve synthesized speech in terms of both similarity and naturalness, even for unseen speakers. Compared to state-of-the-art reference embedding learned methods, our method achieves the best performance on the zero-shot voice cloning task. Furthermore, we demonstrate that the proposed method has a better capability of maintaining the speaker embedding in different languages. Sample outputs are available on the demo page.
文本到语音(TTS)中的语音克隆是用有限的数据复制目标说话者语音的过程。在各种语音克隆技术中,本文重点研究零镜头语音克隆技术。虽然现有的 TTS 模型可以为看到的说话者生成高质量的语音,但克隆未看到的说话者的语音仍然是一项具有挑战性的任务。零镜头语音克隆的关键是获得目标说话者的说话者嵌入。以前的工作使用扬声器编码器从单个参考音频中无监督地获取固定大小的扬声器嵌入,但它们存在扬声器信息不足和扬声器嵌入中内容信息泄漏的问题。为了解决这些问题,本文提出了基于 FastSpeech2 的框架 MRMI-TTS,它使用扬声器嵌入作为提供扬声器信息的条件变量。MRMI-TTS 使用扬声器编码器和内容编码器从多参考音频中提取扬声器嵌入和内容嵌入。为了获得足够的说话者信息,多参考音频是根据句子相似度来选择的。在公共英语数据集 VCTK 上进行的实验表明,我们的方法可以提高合成语音的相似度和自然度,即使是未见过的说话人。与最先进的参考嵌入学习方法相比,我们的方法在零镜头语音克隆任务中取得了最佳性能。此外,我们还证明了所提出的方法在不同语言中保持说话人嵌入的能力更强。演示页面提供了输出样本。
{"title":"MRMI-TTS: Multi-reference audios and Mutual Information Driven Zero-shot Voice cloning","authors":"Yiting Chen, Wanting Li, Buzhou Tang","doi":"10.1145/3649501","DOIUrl":"https://doi.org/10.1145/3649501","url":null,"abstract":"Voice cloning in text-to-speech (TTS) is the process of replicating the voice of a target speaker with limited data. Among various voice cloning techniques, this paper focuses on zero-shot voice cloning. Although existing TTS models can generate high-quality speech for seen speakers, cloning the voice of an unseen speaker remains a challenging task. The key aspect of zero-shot voice cloning is to obtain a speaker embedding from the target speaker. Previous works have used a speaker encoder to obtain a fixed-size speaker embedding from a single reference audio unsupervised, but they suffer from insufficient speaker information and content information leakage in speaker embedding.To address these issues, this paper proposes MRMI-TTS, a FastSpeech2-based framework that uses speaker embedding as a conditioning variable to provide speaker information. The MRMI-TTS extracts speaker embedding and content embedding from multi-reference audios using a speaker encoder and a content encoder. To obtain sufficient speaker information, multi-reference audios are selected based on sentence similarity. The proposed model applies mutual information minimization on the two embeddings to remove entangled information within each embedding.Experiments on the public English dataset VCTK show that our method can improve synthesized speech in terms of both similarity and naturalness, even for unseen speakers. Compared to state-of-the-art reference embedding learned methods, our method achieves the best performance on the zero-shot voice cloning task. Furthermore, we demonstrate that the proposed method has a better capability of maintaining the speaker embedding in different languages. Sample outputs are available on the demo page.","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140363930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Part-of-Speech Tagging for low resource languages: Activation function for deep learning network to work with Minimal Training Data 低资源语言的语音部分标记:使用最少训练数据的深度学习网络激活功能
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-03-30 DOI: 10.1145/3655023
Diganta Baishya, Rupam Baruah
Numerous natural language processing (NLP) applications exist today, especially for the most commonly spoken languages like English, Chinese, and Spanish. Popular traditional methods like Naive Bayes classifiers, Hidden Markov models, Conditional Random field-based classifiers, and other stochastic methods have contributed to this improvement over the last three decades. Recently, deep learning has led to exciting breakthroughs in several areas of artificial intelligence, including image processing and natural language processing. It is important to label words as parts of speech to begin developing most of the NLP applications. A deep study in this area reveals that these approaches require massive training data. Therefore, these approaches have not been helpful for languages not rich in digital resources. Applying these methods with very little training data prompts the need for innovative problem-solving. This paper describes our research, which examines the strengths and weaknesses of well-known approaches, such as conditional random fields and state-of-the-art deep learning models, when applied for part-of-speech tagging using minimal training data for Assamese and English. We also examine the factors affecting them. We discuss our deep learning architecture and the proposed activation function, which shows promise with little training data. The activation function categorizes words belonging to different classes with more confidence by using the outcomes of statistical methods. With minimal training, our deep learning architecture using the proposed PSM-Taylor SoftMax improves accuracy by 4%–9%, This technique is a combination of SMTaylor SoftMax and probability distribution.
当今有许多自然语言处理(NLP)应用,尤其是英语、汉语和西班牙语等最常用的语言。在过去的三十年里,流行的传统方法,如 Naive Bayes 分类器、隐马尔可夫模型、基于条件随机场的分类器和其他随机方法,都为这一进步做出了贡献。最近,深度学习在人工智能的多个领域取得了令人振奋的突破,包括图像处理和自然语言处理。要开始开发大多数自然语言处理应用,必须将单词标记为语篇。对这一领域的深入研究表明,这些方法需要大量的训练数据。因此,这些方法对数字资源不丰富的语言没有帮助。在训练数据极少的情况下应用这些方法,促使我们需要创新性地解决问题。本文介绍了我们的研究,该研究考察了条件随机场和最先进的深度学习模型等著名方法的优缺点,这些方法在阿萨姆语和英语中使用极少的训练数据进行语音部分标记时的应用情况。我们还研究了影响这些方法的因素。我们讨论了我们的深度学习架构和所提出的激活函数,该函数在使用少量训练数据的情况下表现出了良好的前景。激活函数通过使用统计方法的结果,以更高的可信度对属于不同类别的单词进行分类。我们的深度学习架构采用了所提出的 PSM-Taylor SoftMax,在训练量极少的情况下,准确率提高了 4%-9%,这项技术是 SMTaylor SoftMax 和概率分布的结合。
{"title":"Part-of-Speech Tagging for low resource languages: Activation function for deep learning network to work with Minimal Training Data","authors":"Diganta Baishya, Rupam Baruah","doi":"10.1145/3655023","DOIUrl":"https://doi.org/10.1145/3655023","url":null,"abstract":"Numerous natural language processing (NLP) applications exist today, especially for the most commonly spoken languages like English, Chinese, and Spanish. Popular traditional methods like Naive Bayes classifiers, Hidden Markov models, Conditional Random field-based classifiers, and other stochastic methods have contributed to this improvement over the last three decades. Recently, deep learning has led to exciting breakthroughs in several areas of artificial intelligence, including image processing and natural language processing. It is important to label words as parts of speech to begin developing most of the NLP applications. A deep study in this area reveals that these approaches require massive training data. Therefore, these approaches have not been helpful for languages not rich in digital resources. Applying these methods with very little training data prompts the need for innovative problem-solving. This paper describes our research, which examines the strengths and weaknesses of well-known approaches, such as conditional random fields and state-of-the-art deep learning models, when applied for part-of-speech tagging using minimal training data for Assamese and English. We also examine the factors affecting them. We discuss our deep learning architecture and the proposed activation function, which shows promise with little training data. The activation function categorizes words belonging to different classes with more confidence by using the outcomes of statistical methods. With minimal training, our deep learning architecture using the proposed PSM-Taylor SoftMax improves accuracy by 4%–9%, This technique is a combination of SMTaylor SoftMax and probability distribution.","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140362672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Pretrained General-Purpose Vision Language Model for the Vietnamese Language 针对越南语的新型预训练通用视觉语言模型
IF 2 4区 计算机科学 Q2 Computer Science Pub Date : 2024-03-30 DOI: 10.1145/3654796
Vu Dinh Anh, Pham Quang Nhat Minh, Giang Son Tran
Lying in the cross-section of computer vision and natural language processing, vision language models are capable of processing images and text at once. These models are helpful in various tasks: text generation from image and vice versa, image-text retrieval, or visual navigation. Besides building a model trained on a dataset for a task, people also study general-purpose models to utilize many datasets for multitasks. Their two primary applications are image captioning and visual question answering. For English, large datasets and foundation models are already abundant. However, for Vietnamese, they are still limited. To expand the language range, this work proposes a pretrained general-purpose image-text model named VisualRoBERTa. A dataset of 600K images with captions (translated MS COCO 2017 from English to Vietnamese) is introduced to pretrain VisualRoBERTa. The model’s architecture is built using Convolutional Neural Network and Transformer blocks. Fine-tuning VisualRoBERTa shows promising results on the ViVQA dataset with 34.49% accuracy, 0.4173 BLEU 4, and 0.4390 RougeL (in visual question answering task), and best outcomes on the sViIC dataset with 0.6685 BLEU 4, 0.6320 RougeL (in image captioning task).
视觉语言模型是计算机视觉和自然语言处理的交叉部分,能够同时处理图像和文本。这些模型有助于完成各种任务:从图像生成文本(反之亦然)、图像-文本检索或视觉导航。除了针对某项任务在数据集上建立训练有素的模型外,人们还研究通用模型,以利用多个数据集完成多种任务。它们的两个主要应用是图像标题和视觉问题解答。在英语方面,大型数据集和基础模型已经非常丰富。然而,越南语的数据集和基础模型仍然有限。为了扩大语言范围,本研究提出了一个名为 VisualRoBERTa 的预训练通用图像-文本模型。为了对 VisualRoBERTa 进行预训练,我们引入了一个包含 60 万张图片和说明的数据集(由 MS COCO 2017 从英语翻译成越南语)。该模型的架构由卷积神经网络和变换器块构建。微调后的 VisualRoBERTa 在 ViVQA 数据集上取得了可喜的成果,准确率为 34.49%,BLEU 4 为 0.4173,RougeL 为 0.4390(视觉问题解答任务);在 sViIC 数据集上取得了最佳成果,BLEU 4 为 0.6685,RougeL 为 0.6320(图像字幕任务)。
{"title":"A Novel Pretrained General-Purpose Vision Language Model for the Vietnamese Language","authors":"Vu Dinh Anh, Pham Quang Nhat Minh, Giang Son Tran","doi":"10.1145/3654796","DOIUrl":"https://doi.org/10.1145/3654796","url":null,"abstract":"Lying in the cross-section of computer vision and natural language processing, vision language models are capable of processing images and text at once. These models are helpful in various tasks: text generation from image and vice versa, image-text retrieval, or visual navigation. Besides building a model trained on a dataset for a task, people also study general-purpose models to utilize many datasets for multitasks. Their two primary applications are image captioning and visual question answering. For English, large datasets and foundation models are already abundant. However, for Vietnamese, they are still limited. To expand the language range, this work proposes a pretrained general-purpose image-text model named VisualRoBERTa. A dataset of 600K images with captions (translated MS COCO 2017 from English to Vietnamese) is introduced to pretrain VisualRoBERTa. The model’s architecture is built using Convolutional Neural Network and Transformer blocks. Fine-tuning VisualRoBERTa shows promising results on the ViVQA dataset with 34.49% accuracy, 0.4173 BLEU 4, and 0.4390 RougeL (in visual question answering task), and best outcomes on the sViIC dataset with 0.6685 BLEU 4, 0.6320 RougeL (in image captioning task).","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140364058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ACM Transactions on Asian and Low-Resource Language Information Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1