首页 > 最新文献

Big Data Research最新文献

英文 中文
Efficient, interpretable and automated feature engineering for bank data 银行数据的高效、可解释和自动化特征工程
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-28 Epub Date: 2025-03-28 DOI: 10.1016/j.bdr.2025.100524
Atilla Karaahmetoğlu , Mehmet Yıldız , Erdem Ünal , Uğur Aydın , Murat Koraş , Barış Akgün
Banks rely on expert-generated features and simple models to have high performance and interpretability at the same time. Interpretability is needed for internal assessment and regulatory compliance for specific problems such as risk assessment and both expert generated features and simple models satisfy this need. However, feature generation by experts is a time-consuming process and susceptible to bias. In addition, features need to be generated fairly often due to the dynamic nature of bank data, and in case of significant changes or new data sources, expertise might take a while to build up. Complex models, such as deep neural networks, may be able to remedy this. However, interpretability/explainability approaches for complex models are not satisfactory from the banks' point of view. In addition, such models do not always work well with tabular data which is abundant in banking applications. This paper introduces an automated feature synthesis pipeline that creates informative and domain-interpretable features which iconsumes significantly less time than brute-force methods. We create novel feature synthesis steps, define elimination rules to rule out uninterpretable features, and combine performance-based feature selection methods to pick desirable ones to build our models. Our results on two different datasets show that the features generated with our pipeline; (1) perform on par or better than features generated by existing methods, (2) are obtained faster, and (3) are domain-interpretable.
银行依靠专家生成的特征和简单的模型来同时具有高性能和可解释性。内部评估和特定问题(如风险评估)的法规遵从性需要可解释性,专家生成的特征和简单模型都能满足这一需求。然而,由专家生成特征是一个耗时的过程,并且容易受到偏见的影响。此外,由于银行数据的动态性,需要相当频繁地生成功能,并且在发生重大更改或新数据源的情况下,可能需要一段时间才能建立专门知识。复杂的模型,如深度神经网络,可能能够弥补这一点。然而,从银行的角度来看,复杂模型的可解释性/可解释性方法并不令人满意。此外,这种模型并不总是能很好地处理银行应用中大量的表格数据。本文介绍了一种自动化的特征合成管道,它可以创建信息丰富且领域可解释的特征,比暴力方法消耗的时间要少得多。我们创建了新的特征合成步骤,定义了消除规则来排除不可解释的特征,并结合基于性能的特征选择方法来选择理想的特征来构建我们的模型。我们在两个不同的数据集上的结果表明,我们的管道生成的特征;(1)性能与现有方法生成的特征相当或更好,(2)获得速度更快,(3)可域解释。
{"title":"Efficient, interpretable and automated feature engineering for bank data","authors":"Atilla Karaahmetoğlu ,&nbsp;Mehmet Yıldız ,&nbsp;Erdem Ünal ,&nbsp;Uğur Aydın ,&nbsp;Murat Koraş ,&nbsp;Barış Akgün","doi":"10.1016/j.bdr.2025.100524","DOIUrl":"10.1016/j.bdr.2025.100524","url":null,"abstract":"<div><div>Banks rely on expert-generated features and simple models to have high performance and interpretability at the same time. Interpretability is needed for internal assessment and regulatory compliance for specific problems such as risk assessment and both expert generated features and simple models satisfy this need. However, feature generation by experts is a time-consuming process and susceptible to bias. In addition, features need to be generated fairly often due to the dynamic nature of bank data, and in case of significant changes or new data sources, expertise might take a while to build up. Complex models, such as deep neural networks, may be able to remedy this. However, interpretability/explainability approaches for complex models are not satisfactory from the banks' point of view. In addition, such models do not always work well with tabular data which is abundant in banking applications. This paper introduces an automated feature synthesis pipeline that creates informative and domain-interpretable features which iconsumes significantly less time than brute-force methods. We create novel feature synthesis steps, define elimination rules to rule out uninterpretable features, and combine performance-based feature selection methods to pick desirable ones to build our models. Our results on two different datasets show that the features generated with our pipeline; (1) perform on par or better than features generated by existing methods, (2) are obtained faster, and (3) are domain-interpretable.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"40 ","pages":"Article 100524"},"PeriodicalIF":3.5,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143790985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hourglass pattern matching for deep aware neural network text recommendation model 沙漏模式匹配的深度感知神经网络文本推荐模型
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-28 Epub Date: 2025-04-17 DOI: 10.1016/j.bdr.2025.100532
Li Gao, Hongjun Li, Qingkui Chen, Dunlu Peng
In recent years, with the rapid development of deep learning, big data mining, and natural language processing (NLP) technologies, the application of NLP in the field of recommendation systems has attracted significant attention. However, current text recommendation systems still face challenges in handling word distribution assumptions, preprocessing design, network inference models, and text perception technologies. Traditional RNN neural network layers often encounter issues such as gradient explosion or vanishing gradients, which hinder their ability to effectively handle long-term dependencies and reverse text inference among long texts. Therefore, this paper proposes a new type of depth-aware neural network recommendation model (Hourglass Deep-aware neural network Recommendation Model, HDARM), whose structure presents an hourglass shape. This model consists of three parts: The top of the hourglass uses Word Embedding for input through Fine-tune Bert to process text embeddings as word distribution assumptions, followed by utilizing bidirectional LSTM to integrate Transformer models for learning critical information. The middle of the hourglass retains key features of network outputs through CNN layers, which are combined with pooling layers to extract and enhance critical information from user text. The bottom of the hourglass avoids a decline in generalization performance through deep neural network layers. Finally, the model performs pattern matching between text vectors and word embeddings, recommending texts based on their relevance. In experiments, this model improved metrics like MSE and NDCG@10 by 8.74 % and 10.89 % respectively compared to the optimal baseline model.
近年来,随着深度学习、大数据挖掘和自然语言处理(NLP)技术的快速发展,自然语言处理在推荐系统领域的应用备受关注。然而,当前的文本推荐系统在处理词分布假设、预处理设计、网络推理模型和文本感知技术等方面仍然面临挑战。传统的RNN神经网络层经常遇到梯度爆炸或梯度消失等问题,阻碍了其有效处理长文本间的长期依赖关系和反向文本推理的能力。因此,本文提出了一种新型的深度感知神经网络推荐模型(沙漏深度感知神经网络推荐模型,HDARM),其结构呈沙漏形状。该模型由三部分组成:沙漏的顶部使用Word Embedding作为输入,通过微调Bert处理文本嵌入作为单词分布假设,然后使用双向LSTM集成Transformer模型以学习关键信息。沙漏的中间部分通过CNN层保留网络输出的关键特征,并结合池化层从用户文本中提取和增强关键信息。沙漏的底部通过深度神经网络层避免了泛化性能的下降。最后,该模型在文本向量和词嵌入之间进行模式匹配,根据它们的相关性推荐文本。在实验中,与最优基线模型相比,该模型将MSE和NDCG@10等指标分别提高了8.74%和10.89%。
{"title":"Hourglass pattern matching for deep aware neural network text recommendation model","authors":"Li Gao,&nbsp;Hongjun Li,&nbsp;Qingkui Chen,&nbsp;Dunlu Peng","doi":"10.1016/j.bdr.2025.100532","DOIUrl":"10.1016/j.bdr.2025.100532","url":null,"abstract":"<div><div>In recent years, with the rapid development of deep learning, big data mining, and natural language processing (NLP) technologies, the application of NLP in the field of recommendation systems has attracted significant attention. However, current text recommendation systems still face challenges in handling word distribution assumptions, preprocessing design, network inference models, and text perception technologies. Traditional RNN neural network layers often encounter issues such as gradient explosion or vanishing gradients, which hinder their ability to effectively handle long-term dependencies and reverse text inference among long texts. Therefore, this paper proposes a new type of depth-aware neural network recommendation model (Hourglass Deep-aware neural network Recommendation Model, HDARM), whose structure presents an hourglass shape. This model consists of three parts: The top of the hourglass uses Word Embedding for input through Fine-tune Bert to process text embeddings as word distribution assumptions, followed by utilizing bidirectional LSTM to integrate Transformer models for learning critical information. The middle of the hourglass retains key features of network outputs through CNN layers, which are combined with pooling layers to extract and enhance critical information from user text. The bottom of the hourglass avoids a decline in generalization performance through deep neural network layers. Finally, the model performs pattern matching between text vectors and word embeddings, recommending texts based on their relevance. In experiments, this model improved metrics like MSE and NDCG@10 by 8.74 % and 10.89 % respectively compared to the optimal baseline model.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"40 ","pages":"Article 100532"},"PeriodicalIF":3.5,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143923599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NoSQL data warehouse optimizing models: A comparative study of column-oriented approaches NoSQL数据仓库优化模型:面向列方法的比较研究
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-28 Epub Date: 2025-03-20 DOI: 10.1016/j.bdr.2025.100523
Mohamed Mouhiha, Abdelfettah Mabrouk
There is a great challenge when building an efficient Big Data Warehouse (DW) from the traditional data warehouse which used to handle the large datasets. Several presented solutions concentrate on the conversion of a standard DW to an columnar model, especially for direct and traditional data sources. Though there have been many successful algorithms that apply data clustering methods, these approaches also come with their fair share of limitations. This paper provides a comprehensive review of the existing methods, both tuned and out-of-the box, exposing their strengths and weaknesses. Further, a comparative study of the different options is always conducted to compare and assess them.
在传统数据仓库基础上构建高效的大数据仓库(DW)是一个巨大的挑战。提出的几个解决方案集中于将标准DW转换为柱状模型,特别是对于直接数据源和传统数据源。尽管已经有许多成功的算法应用了数据聚类方法,但这些方法也有它们的局限性。本文提供了对现有方法的全面回顾,包括已调优的和开箱即用的,揭示了它们的优点和缺点。此外,总是对不同的选择进行比较研究,以比较和评估它们。
{"title":"NoSQL data warehouse optimizing models: A comparative study of column-oriented approaches","authors":"Mohamed Mouhiha,&nbsp;Abdelfettah Mabrouk","doi":"10.1016/j.bdr.2025.100523","DOIUrl":"10.1016/j.bdr.2025.100523","url":null,"abstract":"<div><div>There is a great challenge when building an efficient Big Data Warehouse (DW) from the traditional data warehouse which used to handle the large datasets. Several presented solutions concentrate on the conversion of a standard DW to an columnar model, especially for direct and traditional data sources. Though there have been many successful algorithms that apply data clustering methods, these approaches also come with their fair share of limitations. This paper provides a comprehensive review of the existing methods, both tuned and out-of-the box, exposing their strengths and weaknesses. Further, a comparative study of the different options is always conducted to compare and assess them.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"40 ","pages":"Article 100523"},"PeriodicalIF":3.5,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143681953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging artificial intelligence for pandemic management: Case of COVID-19 in the United States 利用人工智能进行流行病管理:以美国的COVID-19为例
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-28 Epub Date: 2025-04-08 DOI: 10.1016/j.bdr.2025.100529
Ehsan Ahmadi, Reza Maihami
The COVID-19 pandemic revealed significant limitations in traditional approaches to analyzing time-series data that use one-dimensional data such as historical infection rates. Such approaches do not capture the complex, multifactor influences on disease spread. This paper addresses these challenges by proposing a comprehensive methodology that integrates multiple data sources, including community mobility, census information, Google search trends, socioeconomic variables, vaccination coverage, and political data. In addition, this paper proposes a new cross-learning (CL) methodology that allows for the training of machine learning models on multiple related time series simultaneously, enabling more accurate and robust predictions. Applying the CL approach with four machine learning algorithms, we successfully forecasted confirmed COVID-19 cases 30 days in advance with greater accuracy than the traditional ARIMAX model and the newer Transformer deep learning technique. Our findings identified daily hospital admissions as a significant predictor at the state level and vaccination status at the national level. Random Forest with CL was very effective, performing best in 44 states, while ARIMAX outperformed in seven larger states. These findings highlight the importance of advanced predictive modeling in resource optimization and response strategy development for future health emergencies.
COVID-19大流行表明,使用历史感染率等一维数据分析时间序列数据的传统方法存在重大局限性。这种方法没有捕捉到对疾病传播的复杂的多因素影响。本文通过提出一种综合的方法来解决这些挑战,该方法集成了多个数据源,包括社区流动性、人口普查信息、谷歌搜索趋势、社会经济变量、疫苗接种覆盖率和政治数据。此外,本文提出了一种新的交叉学习(CL)方法,该方法允许同时在多个相关时间序列上训练机器学习模型,从而实现更准确和稳健的预测。采用CL方法和四种机器学习算法,我们成功地提前30天预测了新冠肺炎确诊病例,其准确性高于传统的ARIMAX模型和较新的Transformer深度学习技术。我们的研究结果确定每日住院率是州一级和国家一级疫苗接种状况的重要预测因子。带有CL的随机森林非常有效,在44个州表现最好,而ARIMAX在7个较大的州表现更好。这些发现突出了先进的预测建模在未来突发卫生事件资源优化和应对策略制定中的重要性。
{"title":"Leveraging artificial intelligence for pandemic management: Case of COVID-19 in the United States","authors":"Ehsan Ahmadi,&nbsp;Reza Maihami","doi":"10.1016/j.bdr.2025.100529","DOIUrl":"10.1016/j.bdr.2025.100529","url":null,"abstract":"<div><div>The COVID-19 pandemic revealed significant limitations in traditional approaches to analyzing time-series data that use one-dimensional data such as historical infection rates. Such approaches do not capture the complex, multifactor influences on disease spread. This paper addresses these challenges by proposing a comprehensive methodology that integrates multiple data sources, including community mobility, census information, Google search trends, socioeconomic variables, vaccination coverage, and political data. In addition, this paper proposes a new cross-learning (CL) methodology that allows for the training of machine learning models on multiple related time series simultaneously, enabling more accurate and robust predictions. Applying the CL approach with four machine learning algorithms, we successfully forecasted confirmed COVID-19 cases 30 days in advance with greater accuracy than the traditional ARIMAX model and the newer Transformer deep learning technique. Our findings identified daily hospital admissions as a significant predictor at the state level and vaccination status at the national level. Random Forest with CL was very effective, performing best in 44 states, while ARIMAX outperformed in seven larger states. These findings highlight the importance of advanced predictive modeling in resource optimization and response strategy development for future health emergencies.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"40 ","pages":"Article 100529"},"PeriodicalIF":3.5,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143839334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modeling meaningful volatility events to classify monetary policy announcements 建立有意义的波动事件模型,对货币政策公告进行分类
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-28 Epub Date: 2025-02-26 DOI: 10.1016/j.bdr.2025.100517
Giampiero M. Gallo , Demetrio Lacava , Edoardo Otranto
Central Bank monetary policy interventions frequently have direct implications for financial market volatility. In this paper, we introduce an intradaily Asymmetric Multiplicative Error Model with Meaningful Volatility (MV) events (AMEM-MV), which decomposes realized variance into a base component and an MV component. A novel model-based classification of monetary announcements is developed based on their impact on the MV component of the variance. By focusing on the 30-minute window following each Federal Reserve communication, we isolate the specific impact of monetary announcements on the volatility of seven US tickers.
中央银行货币政策干预经常对金融市场波动产生直接影响。本文提出了一种包含有意义波动率(MV)事件的日内非对称乘法误差模型(AMEM-MV),该模型将实际方差分解为基分量和MV分量。基于货币公告对方差的MV分量的影响,开发了一种新的基于模型的货币公告分类。通过关注美联储每次信息发布后的30分钟窗口,我们分离出货币政策公告对7个美国股票市场波动性的具体影响。
{"title":"Modeling meaningful volatility events to classify monetary policy announcements","authors":"Giampiero M. Gallo ,&nbsp;Demetrio Lacava ,&nbsp;Edoardo Otranto","doi":"10.1016/j.bdr.2025.100517","DOIUrl":"10.1016/j.bdr.2025.100517","url":null,"abstract":"<div><div>Central Bank monetary policy interventions frequently have direct implications for financial market volatility. In this paper, we introduce an intradaily Asymmetric Multiplicative Error Model with Meaningful Volatility (MV) events (AMEM-MV), which decomposes realized variance into a base component and an MV component. A novel model-based classification of monetary announcements is developed based on their impact on the MV component of the variance. By focusing on the 30-minute window following each Federal Reserve communication, we isolate the specific impact of monetary announcements on the volatility of seven US tickers.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"40 ","pages":"Article 100517"},"PeriodicalIF":3.5,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143509938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-granularity enhanced graph convolutional network for aspect sentiment triplet extraction 面向方面情感三元组提取的多粒度增强图卷积网络
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-28 Epub Date: 2025-01-17 DOI: 10.1016/j.bdr.2025.100506
Mingwei Tang , Kun Yang , Linping Tao , Mingfeng Zhao , Wei Zhou
Aspect Sentiment Triple Extraction (ASTE) is an emerging sentiment analysis task, which describes both aspect terms and their sentiment polarity, as well as opinion terms that represent sentiment polarity. Some models have been presented to analyze sentence sentiment more accurately. Nonetheless, previous models have had problems, like inconsistent sentiment predictions for one-to-many, many-to-one, and sequence annotation. In addition, part-of-speech and contextual semantic information are ignored, resulting in the inability to identify complete multi-word aspect terms and opinion terms. To address these problems, we propose a Multi-granularity Enhanced Graph Convolutional Network (MGEGCN) to solve the problem of inaccurate multi-word term recognition. First, we propose a dual-channel enhanced graph convolutional network, which simultaneously analyzes syntactic structure and part-of-speech information and uses the combined effect of the two to enhance the deep semantic information of aspect terms and opinion terms. Second, we also design a multi-scale attention, which combines self-attention with deep separable convolution to enhance attention to aspect terms and opinion terms. In addition, a convolutional decoding strategy is used in the decoding stage to extract triples by directly detecting and classifying the relational regions in the table. In the experimental part, we conduct analysis on two public datasets (ASTE-DATA-v1 and ASTE-DATA-v2) to prove that the model improves the performance of ASTE tasks. In four subsets (14res, 14lap, 15res, and 16res), the F1 scores of the MGEGCN method are 75.65%, 61.62%, 67.62%, 74.12% and 74.69%, 62.10%, 68.18%, 74.00%, respectively.
方面情感三重提取(ASTE)是一种新兴的情感分析任务,它既描述方面术语及其情感极性,也描述代表情感极性的意见术语。为了更准确地分析句子情感,已经提出了一些模型。尽管如此,以前的模型存在一些问题,比如一对多、多对一和序列注释的情感预测不一致。此外,词性和上下文语义信息被忽略,导致无法识别完整的多词方面术语和意见术语。为了解决这些问题,我们提出了一种多粒度增强图卷积网络(MGEGCN)来解决多词术语识别不准确的问题。首先,我们提出了一种双通道增强图卷积网络,该网络同时分析句法结构和词性信息,并利用两者的联合作用增强方面词和意见词的深层语义信息。其次,我们还设计了一个多尺度注意,将自我注意与深度可分离卷积相结合,以增强对方面项和意见项的注意。此外,在解码阶段采用卷积解码策略,通过直接检测和分类表中的关系区域提取三元组。在实验部分,我们对两个公共数据集(ASTE- data -v1和ASTE- data -v2)进行了分析,证明该模型提高了ASTE任务的性能。在14res、14lap、15res和16res 4个子集中,MGEGCN方法的F1得分分别为75.65%、61.62%、67.62%、74.12%和74.69%、62.10%、68.18%、74.00%。
{"title":"Multi-granularity enhanced graph convolutional network for aspect sentiment triplet extraction","authors":"Mingwei Tang ,&nbsp;Kun Yang ,&nbsp;Linping Tao ,&nbsp;Mingfeng Zhao ,&nbsp;Wei Zhou","doi":"10.1016/j.bdr.2025.100506","DOIUrl":"10.1016/j.bdr.2025.100506","url":null,"abstract":"<div><div>Aspect Sentiment Triple Extraction (ASTE) is an emerging sentiment analysis task, which describes both aspect terms and their sentiment polarity, as well as opinion terms that represent sentiment polarity. Some models have been presented to analyze sentence sentiment more accurately. Nonetheless, previous models have had problems, like inconsistent sentiment predictions for one-to-many, many-to-one, and sequence annotation. In addition, part-of-speech and contextual semantic information are ignored, resulting in the inability to identify complete multi-word aspect terms and opinion terms. To address these problems, we propose a <em>Multi-granularity Enhanced Graph Convolutional Network</em> (MGEGCN) to solve the problem of inaccurate multi-word term recognition. First, we propose a dual-channel enhanced graph convolutional network, which simultaneously analyzes syntactic structure and part-of-speech information and uses the combined effect of the two to enhance the deep semantic information of aspect terms and opinion terms. Second, we also design a multi-scale attention, which combines self-attention with deep separable convolution to enhance attention to aspect terms and opinion terms. In addition, a convolutional decoding strategy is used in the decoding stage to extract triples by directly detecting and classifying the relational regions in the table. In the experimental part, we conduct analysis on two public datasets (ASTE-DATA-v1 and ASTE-DATA-v2) to prove that the model improves the performance of ASTE tasks. In four subsets (14res, 14lap, 15res, and 16res), the F1 scores of the MGEGCN method are 75.65%, 61.62%, 67.62%, 74.12% and 74.69%, 62.10%, 68.18%, 74.00%, respectively.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"39 ","pages":"Article 100506"},"PeriodicalIF":3.5,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143092348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Has machine paraphrasing skills approached humans? Detecting automatically and manually generated paraphrased cases 机器的释义能力已经接近人类了吗?检测自动和手动生成的释义案例
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-28 Epub Date: 2025-01-22 DOI: 10.1016/j.bdr.2025.100507
Iqra Muneer , Aysha Shehzadi , Muhammad Adnan Ashraf , Rao Muhammad Adeel Nawab
In recent years, automatic text rewriting (or paraphrasing) tools are readily and publicly available. These tools have enabled text paraphrasing as an exceptionally straightforward approach that encourages trouble-free plagiarism and text reuse. In literature, the majority of efforts have focused on detecting real cases (manual/human paraphrasing) of paraphrasing (mainly in the domain of journalism). However, the problem of paraphrase detection has not been thoroughly explored for artificial cases (machine paraphrased), mainly, due to lack of standard resources for its evaluation. To fulfill this gap, this study proposes three benchmark corpora for artificial cases of paraphrases at sentence level, and one real corpus contains examples from daily life activities. Three popular and widely used automatic text rewriting online tools have been used, i.e., paraphrasing-tools, articlerewritetool and rewritertools, to develop artificial case corpora. Further, we used two real cases corpora, including Microsoft Paraphrase Corpus (MSRP) (from the domain of journalism) and a proposed real corpus which is a combination of carefully extracted Quora question pairs and MSRP (Q-MSRP). Both real case and artificial case paraphrases were evaluated using classical machine learning, transfer learning, Large language models and a proposed model, to investigate which of the two types of paraphrasing is more difficult to detect. The results show that our proposed model outperforms all the other approaches for both artificial and real case paraphrase detection. A thorough analysis of the results suggests that, by far, manual paraphrasing is still harder to detect but certain machine paraphrased texts are equally difficult to detect. All proposed corpora are freely available to promote the research on artificial case paraphrase detection.
近年来,自动文本重写(或改写)工具很容易公开可用。这些工具使文本释义成为一种非常直接的方法,鼓励无故障的剽窃和文本重用。在文献中,大多数的努力都集中在检测释义的真实案例(手动/人工释义)(主要在新闻领域)。然而,对于人工案例(机器释义)的释义检测问题尚未深入探讨,主要原因是缺乏标准的评价资源。为了填补这一空白,本研究提出了三个句子层面的人工释义基准语料库,一个包含日常生活活动实例的真实语料库。本文利用三种流行的、广泛使用的在线自动文本改写工具,即释义工具、文章书写工具和重写工具,来开发人工案例语料库。此外,我们使用了两个真实案例语料库,包括微软释义语料库(MSRP)(来自新闻领域)和一个提议的真实语料库,该语料库是精心提取的Quora问题对和MSRP (Q-MSRP)的组合。使用经典机器学习、迁移学习、大型语言模型和一个建议模型对真实案例和人工案例释义进行评估,以研究哪一种类型的释义更难检测。结果表明,我们提出的模型在人工和真实案例释义检测方面都优于所有其他方法。对结果的全面分析表明,到目前为止,人工释义仍然难以检测,但某些机器释义的文本同样难以检测。所有建议的语料库都是免费提供的,以促进人工案例释义检测的研究。
{"title":"Has machine paraphrasing skills approached humans? Detecting automatically and manually generated paraphrased cases","authors":"Iqra Muneer ,&nbsp;Aysha Shehzadi ,&nbsp;Muhammad Adnan Ashraf ,&nbsp;Rao Muhammad Adeel Nawab","doi":"10.1016/j.bdr.2025.100507","DOIUrl":"10.1016/j.bdr.2025.100507","url":null,"abstract":"<div><div>In recent years, automatic text rewriting (or paraphrasing) tools are readily and publicly available. These tools have enabled text paraphrasing as an exceptionally straightforward approach that encourages trouble-free plagiarism and text reuse. In literature, the majority of efforts have focused on detecting real cases (manual/human paraphrasing) of paraphrasing (mainly in the domain of journalism). However, the problem of paraphrase detection has not been thoroughly explored for artificial cases (machine paraphrased), mainly, due to lack of standard resources for its evaluation. To fulfill this gap, this study proposes three benchmark corpora for artificial cases of paraphrases at sentence level, and one real corpus contains examples from daily life activities. Three popular and widely used automatic text rewriting online tools have been used, i.e., paraphrasing-tools, articlerewritetool and rewritertools, to develop artificial case corpora. Further, we used two real cases corpora, including Microsoft Paraphrase Corpus (MSRP) (from the domain of journalism) and a proposed real corpus which is a combination of carefully extracted Quora question pairs and MSRP (Q-MSRP). Both real case and artificial case paraphrases were evaluated using classical machine learning, transfer learning, Large language models and a proposed model, to investigate which of the two types of paraphrasing is more difficult to detect. The results show that our proposed model outperforms all the other approaches for both artificial and real case paraphrase detection. A thorough analysis of the results suggests that, by far, manual paraphrasing is still harder to detect but certain machine paraphrased texts are equally difficult to detect. All proposed corpora are freely available to promote the research on artificial case paraphrase detection.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"39 ","pages":"Article 100507"},"PeriodicalIF":3.5,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143092345","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel approach for job matching and skill recommendation using transformers and the O*NET database 一种利用变压器和O*NET数据库进行工作匹配和技能推荐的新方法
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-28 Epub Date: 2025-02-07 DOI: 10.1016/j.bdr.2025.100509
Rubén Alonso , Danilo Dessí , Antonello Meloni , Diego Reforgiato Recupero
Today we have tons of information posted on the web every day regarding job supply and demand which has heavily affected the job market. The online enrolling process has thus become efficient for applicants as it allows them to present their resumes using the Internet and, as such, simultaneously to numerous organizations. Online systems such as Monster.com, OfferZen, and LinkedIn contain millions of job offers and resumes of potential candidates leaving to companies with the hard task to face an enormous amount of data to manage to select the most suitable applicant. The task of assessing the resumes of candidates and providing automatic recommendations on which one suits a particular position best has, therefore, become essential to speed up the hiring process. Similarly, it is important to help applicants to quickly find a job appropriate to their skills and provide recommendations about what they need to master to become eligible for certain jobs. Our approach lies in this context and proposes a new method to identify skills from candidates' resumes and match resumes with job descriptions. We employed the O*NET database entities related to different skills and abilities required by different jobs; moreover, we leveraged deep learning technologies to compute the semantic similarity between O*NET entities and part of text extracted from candidates' resumes. The ultimate goal is to identify the most suitable job for a certain resume according to the information there contained. We have defined two scenarios: i) given a resume, identify the top O*NET occupations with the highest match with the resume, ii) given a candidate's resume and a set of job descriptions, identify which one of the input jobs is the most suitable for the candidate. The evaluation that has been carried out indicates that the proposed approach outperforms the baselines in the two scenarios. Finally, we provide a use case for candidates where it is possible to recommend courses with the goal to fill certain skills and make them qualified for a certain job.
今天,我们每天在网上发布大量关于工作供求的信息,这些信息严重影响了就业市场。因此,在线报名过程对申请人来说变得高效,因为它允许他们使用互联网提交简历,并同时向许多组织提交简历。Monster.com、OfferZen和LinkedIn等在线系统包含了数百万份潜在求职者的工作邀请和简历,这些求职者留给公司的任务艰巨,需要面对大量数据,才能选择最合适的求职者。因此,评估候选人的简历并自动推荐最适合特定职位的人,对加快招聘过程至关重要。同样,帮助求职者快速找到一份适合他们技能的工作,并提供他们需要掌握哪些技能才能胜任某些工作的建议也很重要。我们的方法就是在这种背景下,提出了一种新的方法来从候选人的简历中识别技能,并将简历与职位描述相匹配。我们采用了与不同工作所需的不同技能和能力相关的O*NET数据库实体;此外,我们利用深度学习技术来计算O*NET实体与从候选人简历中提取的部分文本之间的语义相似度。最终的目标是根据简历中包含的信息来确定最适合的工作。我们定义了两个场景:i)给出一份简历,找出与简历匹配度最高的O*NET职业;ii)给出一份求职者的简历和一组职位描述,找出输入的职位中哪一个最适合该求职者。已进行的评估表明,拟议的方法在两种情况下优于基线。最后,我们为候选人提供了一个用例,在这个用例中,可以推荐课程,以满足特定技能的要求,并使他们能够胜任特定的工作。
{"title":"A novel approach for job matching and skill recommendation using transformers and the O*NET database","authors":"Rubén Alonso ,&nbsp;Danilo Dessí ,&nbsp;Antonello Meloni ,&nbsp;Diego Reforgiato Recupero","doi":"10.1016/j.bdr.2025.100509","DOIUrl":"10.1016/j.bdr.2025.100509","url":null,"abstract":"<div><div>Today we have tons of information posted on the web every day regarding job supply and demand which has heavily affected the job market. The online enrolling process has thus become efficient for applicants as it allows them to present their resumes using the Internet and, as such, simultaneously to numerous organizations. Online systems such as Monster.com, OfferZen, and LinkedIn contain millions of job offers and resumes of potential candidates leaving to companies with the hard task to face an enormous amount of data to manage to select the most suitable applicant. The task of assessing the resumes of candidates and providing automatic recommendations on which one suits a particular position best has, therefore, become essential to speed up the hiring process. Similarly, it is important to help applicants to quickly find a job appropriate to their skills and provide recommendations about what they need to master to become eligible for certain jobs. Our approach lies in this context and proposes a new method to identify skills from candidates' resumes and match resumes with job descriptions. We employed the O*NET database entities related to different skills and abilities required by different jobs; moreover, we leveraged deep learning technologies to compute the semantic similarity between O*NET entities and part of text extracted from candidates' resumes. The ultimate goal is to identify the most suitable job for a certain resume according to the information there contained. We have defined two scenarios: i) given a resume, identify the top O*NET occupations with the highest match with the resume, ii) given a candidate's resume and a set of job descriptions, identify which one of the input jobs is the most suitable for the candidate. The evaluation that has been carried out indicates that the proposed approach outperforms the baselines in the two scenarios. Finally, we provide a use case for candidates where it is possible to recommend courses with the goal to fill certain skills and make them qualified for a certain job.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"39 ","pages":"Article 100509"},"PeriodicalIF":3.5,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143377085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Tesseract optical character recognition performance on Thai document datasets 改进泰语文档数据集上的Tesseract光学字符识别性能
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-28 Epub Date: 2025-02-08 DOI: 10.1016/j.bdr.2025.100508
Noppol Anakpluek, Watcharakorn Pasanta, Latthawan Chantharasukha, Pattanawong Chokratansombat, Pajaya Kanjanakaew, Thitirat Siriborvornratanakul
This research aims to improve the accuracy and efficiency of Optical Character Recognition (OCR) technology for the Thai language, specifically in the context of Thai government documents. OCR enables the conversion of text from images into machine-readable format, facilitating document storage and further processing. However, applying OCR to the Thai language presents unique challenges due to its complexity. This study focuses on enhancing the performance of the Tesseract OCR engine, a widely used free OCR technology, by implementing various image preprocessing techniques such as masking, adaptive thresholds, median filtering, Canny edge detection, and morphological operators. A dataset of Thai documents is utilized, and the OCR system's output is evaluated using word error rate (WER) and character error rate (CER) metrics. To improve text extraction accuracy, the research employs the original U-Net architecture [19] for image segmentation. Furthermore, the Tesseract OCR engine is finetuned, and image preprocessing is performed to optimize OCR system accuracy. The developed tools automate workflow processes, alleviate constraints on model training, and enable the effective utilization of information from official Thai documents for various purposes.
本研究旨在提高泰国语光学字符识别(OCR)技术的准确性和效率,特别是在泰国政府文件的背景下。OCR可以将图像中的文本转换为机器可读的格式,方便文档存储和进一步处理。然而,由于其复杂性,将OCR应用于泰语面临着独特的挑战。本研究的重点是通过实现各种图像预处理技术,如掩模、自适应阈值、中值滤波、Canny边缘检测和形态学算子,来增强广泛使用的免费OCR技术Tesseract OCR引擎的性能。使用泰语文档的数据集,并使用单词错误率(WER)和字符错误率(CER)指标评估OCR系统的输出。为了提高文本提取的准确性,本研究采用了原始的U-Net架构[19]进行图像分割。此外,对Tesseract OCR引擎进行了微调,并对图像进行了预处理,以优化OCR系统的精度。开发的工具使工作流程自动化,减轻了模型训练的限制,并能够有效地利用泰国官方文档中的信息用于各种目的。
{"title":"Improved Tesseract optical character recognition performance on Thai document datasets","authors":"Noppol Anakpluek,&nbsp;Watcharakorn Pasanta,&nbsp;Latthawan Chantharasukha,&nbsp;Pattanawong Chokratansombat,&nbsp;Pajaya Kanjanakaew,&nbsp;Thitirat Siriborvornratanakul","doi":"10.1016/j.bdr.2025.100508","DOIUrl":"10.1016/j.bdr.2025.100508","url":null,"abstract":"<div><div>This research aims to improve the accuracy and efficiency of Optical Character Recognition (OCR) technology for the Thai language, specifically in the context of Thai government documents. OCR enables the conversion of text from images into machine-readable format, facilitating document storage and further processing. However, applying OCR to the Thai language presents unique challenges due to its complexity. This study focuses on enhancing the performance of the Tesseract OCR engine, a widely used free OCR technology, by implementing various image preprocessing techniques such as masking, adaptive thresholds, median filtering, Canny edge detection, and morphological operators. A dataset of Thai documents is utilized, and the OCR system's output is evaluated using word error rate (WER) and character error rate (CER) metrics. To improve text extraction accuracy, the research employs the original U-Net architecture [<span><span>19</span></span>] for image segmentation. Furthermore, the Tesseract OCR engine is finetuned, and image preprocessing is performed to optimize OCR system accuracy. The developed tools automate workflow processes, alleviate constraints on model training, and enable the effective utilization of information from official Thai documents for various purposes.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"39 ","pages":"Article 100508"},"PeriodicalIF":3.5,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143396246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Positional-attention based bidirectional deep stacked AutoEncoder for aspect based sentimental analysis 基于位置注意力的双向深度堆叠自编码器,用于面向情感分析
IF 3.5 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-28 Epub Date: 2024-12-16 DOI: 10.1016/j.bdr.2024.100505
S. Anjali Devi , M. Sitha Ram , Pulugu Dileep , Sasibhushana Rao Pappu , T. Subha Mastan Rao , Mula Malyadri
With the rapid growth of Internet technology and social networks, the generation of text-based information on the web is increased. To ease the Natural Language Processing (NLP) tasks, analyzing the sentiments behind the provided input text is highly important. To effectively analyze the polarities of sentiments (positive, negative and neutral), categorizing the aspects in the text is an essential task. Several existing studies have attempted to accurately classify aspects based on sentiments in text inputs. However, the existing methods attained limited performance because of reduced aspect coverage, inefficiency in handling ambiguous language, inappropriate feature extraction, lack of contextual understanding and overfitting issues. Thus, the proposed study intends to develop an effective word embedding scheme with a novel hybrid deep learning technique for performing aspect-based sentimental analysis in a social media text. Initially, the collected raw input text data are pre-processed to reduce the undesirable data by initiating tokenization, stemming, lemmatization, duplicate removal, stop words removal, empty sets removal and empty rows removal. The required information from the pre-processed text is extracted using three varied word-level embedding methods: Scored-Lexicon based Word2Vec, Glove modelling and Extended Bidirectional Encoder Representation from Transformers (E-BERT). After extracting sufficient features, the aspects are analyzed, and the exact sentimental polarities are classified through a novel Positional-Attention-based Bidirectional Deep Stacked AutoEncoder (PA_BiDSAE) model. In this proposed classification, the BiLSTM network is hybridized with a deep stacked autoencoder (DSAE) model to categorize sentiment. The experimental analysis is done by using Python software, and the proposed model is simulated with three publicly available datasets: SemEval Challenge 2014 (Restaurant), SemEval Challenge 2014 (Laptop) and SemEval Challenge 2015 (Restaurant). The performance analysis proves that the proposed hybrid deep learning model obtains improved classification performance in accuracy, precision, recall, specificity, F1 score and kappa measure.
随着互联网技术和社交网络的快速发展,网络上基于文本的信息的产生越来越多。为了简化自然语言处理(NLP)任务,分析提供的输入文本背后的情感是非常重要的。为了有效地分析情感的极性(积极、消极和中性),对文本中的极性进行分类是一项必不可少的工作。现有的一些研究试图根据文本输入中的情感对方面进行准确分类。然而,现有的方法由于方面覆盖率低、处理歧义语言效率低、特征提取不当、缺乏上下文理解和过拟合等问题而性能有限。因此,本研究旨在开发一种有效的词嵌入方案,采用一种新的混合深度学习技术,在社交媒体文本中进行基于方面的情感分析。首先,对收集到的原始输入文本数据进行预处理,通过启动标记化、词干提取、词序化、重复删除、停止词删除、空集删除和空行删除来减少不需要的数据。从预处理文本中提取所需信息使用三种不同的词级嵌入方法:基于评分词典的Word2Vec,手套建模和来自变形金刚的扩展双向编码器表示(E-BERT)。在提取足够的特征后,对这些方面进行分析,并通过一种新的基于位置-注意力的双向深度堆叠自动编码器(PA_BiDSAE)模型分类出准确的情感极性。在该分类中,BiLSTM网络与深度堆叠自编码器(DSAE)模型相结合,对情感进行分类。实验分析使用Python软件完成,并使用三个公开可用的数据集模拟所提出的模型:SemEval Challenge 2014 (Restaurant), SemEval Challenge 2014 (Laptop)和SemEval Challenge 2015 (Restaurant)。性能分析表明,所提出的混合深度学习模型在准确率、精密度、召回率、特异性、F1评分和kappa测度等方面都取得了较好的分类性能。
{"title":"Positional-attention based bidirectional deep stacked AutoEncoder for aspect based sentimental analysis","authors":"S. Anjali Devi ,&nbsp;M. Sitha Ram ,&nbsp;Pulugu Dileep ,&nbsp;Sasibhushana Rao Pappu ,&nbsp;T. Subha Mastan Rao ,&nbsp;Mula Malyadri","doi":"10.1016/j.bdr.2024.100505","DOIUrl":"10.1016/j.bdr.2024.100505","url":null,"abstract":"<div><div>With the rapid growth of Internet technology and social networks, the generation of text-based information on the web is increased. To ease the Natural Language Processing (NLP) tasks, analyzing the sentiments behind the provided input text is highly important. To effectively analyze the polarities of sentiments (positive, negative and neutral), categorizing the aspects in the text is an essential task. Several existing studies have attempted to accurately classify aspects based on sentiments in text inputs. However, the existing methods attained limited performance because of reduced aspect coverage, inefficiency in handling ambiguous language, inappropriate feature extraction, lack of contextual understanding and overfitting issues. Thus, the proposed study intends to develop an effective word embedding scheme with a novel hybrid deep learning technique for performing aspect-based sentimental analysis in a social media text. Initially, the collected raw input text data are pre-processed to reduce the undesirable data by initiating tokenization, stemming, lemmatization, duplicate removal, stop words removal, empty sets removal and empty rows removal. The required information from the pre-processed text is extracted using three varied word-level embedding methods: Scored-Lexicon based Word2Vec, Glove modelling and Extended Bidirectional Encoder Representation from Transformers (E-BERT). After extracting sufficient features, the aspects are analyzed, and the exact sentimental polarities are classified through a novel Positional-Attention-based Bidirectional Deep Stacked AutoEncoder (PA_BiDSAE) model. In this proposed classification, the BiLSTM network is hybridized with a deep stacked autoencoder (DSAE) model to categorize sentiment. The experimental analysis is done by using Python software, and the proposed model is simulated with three publicly available datasets: SemEval Challenge 2014 (Restaurant), SemEval Challenge 2014 (Laptop) and SemEval Challenge 2015 (Restaurant). The performance analysis proves that the proposed hybrid deep learning model obtains improved classification performance in accuracy, precision, recall, specificity, F1 score and kappa measure.</div></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"39 ","pages":"Article 100505"},"PeriodicalIF":3.5,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143092347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Big Data Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1