首页 > 最新文献

Data & Knowledge Engineering最新文献

英文 中文
Hermes, a low-latency transactional storage for binary data streams from remote devices Hermes,一种用于远程设备二进制数据流的低延迟事务存储设备
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-16 DOI: 10.1016/j.datak.2024.102315
Gabriele Scaffidi Militone, Daniele Apiletti, Giovanni Malnati

In many contexts where data is streamed on a large scale, such as video surveillance systems, there is a dual requirement: secure data storage and continuous access to audio and video content by third parties, such as human operators or specific business logic, even while the media files are still being collected. However, using transactions to ensure data persistence often limits system throughput and latency. This paper presents a solution that enables both high ingestion rates with transactional data persistence and near real-time, low-latency access to the stream during collection. This immediate access enables the prompt application of specialized data engineering algorithms during data acquisition. The proposed solution is particularly suitable for binary data sources such as audio and video recordings in surveillance systems, and it can be extended to various big data scenarios via well-defined general interfaces. The scalability of the approach is based on the microservice architecture. Preliminary results obtained with Apache Kafka and MongoDB replica sets show that the proposed solution provides up to 3 times higher throughput and 2.2 times lower latency compared to standard multi-document transactions.

在视频监控系统等大规模流式传输数据的许多情况下,存在着双重需求:安全的数据存储和第三方(如人工操作员或特定业务逻辑)对音频和视频内容的持续访问,即使媒体文件仍在收集过程中。然而,使用事务来确保数据持久性往往会限制系统的吞吐量和延迟。本文提出了一种解决方案,既能通过事务数据持久性实现高摄取率,又能在采集过程中对数据流进行近乎实时的低延迟访问。这种即时访问可在数据采集期间迅速应用专门的数据工程算法。所提出的解决方案特别适用于二进制数据源,如监控系统中的音频和视频记录,并可通过定义明确的通用接口扩展到各种大数据场景。该方法的可扩展性基于微服务架构。使用 Apache Kafka 和 MongoDB 复制集获得的初步结果表明,与标准多文档事务相比,拟议解决方案的吞吐量提高了 3 倍,延迟降低了 2.2 倍。
{"title":"Hermes, a low-latency transactional storage for binary data streams from remote devices","authors":"Gabriele Scaffidi Militone,&nbsp;Daniele Apiletti,&nbsp;Giovanni Malnati","doi":"10.1016/j.datak.2024.102315","DOIUrl":"10.1016/j.datak.2024.102315","url":null,"abstract":"<div><p>In many contexts where data is streamed on a large scale, such as video surveillance systems, there is a dual requirement: secure data storage and continuous access to audio and video content by third parties, such as human operators or specific business logic, even while the media files are still being collected. However, using transactions to ensure data persistence often limits system throughput and latency. This paper presents a solution that enables both high ingestion rates with transactional data persistence and near real-time, low-latency access to the stream during collection. This immediate access enables the prompt application of specialized data engineering algorithms during data acquisition. The proposed solution is particularly suitable for binary data sources such as audio and video recordings in surveillance systems, and it can be extended to various big data scenarios via well-defined general interfaces. The scalability of the approach is based on the microservice architecture. Preliminary results obtained with Apache Kafka and MongoDB replica sets show that the proposed solution provides up to 3 times higher throughput and 2.2 times lower latency compared to standard multi-document transactions.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"153 ","pages":"Article 102315"},"PeriodicalIF":2.5,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141042236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analyzing fuzzy semantics of reviews for multi-criteria recommendations 为多标准推荐分析评论的模糊语义
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-16 DOI: 10.1016/j.datak.2024.102314
Navreen Kaur Boparai , Himanshu Aggarwal , Rinkle Rani

Hotel reviews play a vital role in tourism recommender system. They should be analyzed effectively to enhance the accuracy of recommendations which can be generated either from crisp ratings on a fixed scale or real sentiments of reviews. But crisp ratings cannot represent the actual feelings of reviewers. Existing tourism recommender systems mostly recommend hotels on the basis of vague and sparse ratings resulting in inaccurate recommendations or preferences for online users. This paper presents a semantic approach to analyze the online reviews being crawled from tripadvisor.in. It discovers the underlying fuzzy semantics of reviews with respect to the multiple criteria of hotels rather than using the crisp ratings. The crawled reviews are preprocessed via data cleaning such as stopword and punctuation removal, tokenization, lemmatization, pos tagging to understand the semantics efficiently. Nouns representing frequent features of hotels are extracted from pre-processed reviews which are further used to identify opinion phrases. Fuzzy weights are derived from normalized frequency of frequent nouns and combined with sentiment score of all the synonyms of adjectives in the identified opinion phrases. This results in fuzzy semantics which form an ideal representation of reviews for a multi-criteria tourism recommender system. The proposed work is implemented in python by crawling the recent reviews of Jaipur hotels from TripAdvisor and analyzing their semantics. The resultant fuzzy semantics form a manually tagged dataset of reviews tagged with sentiments of identified aspects, respectively. Experimental results show improved sentiment score while considering all the synonyms of adjectives. The results are further used to fine-tune BERT models to form encodings for a query-based recommender system. The proposed approach can help tourism and hospitality service providers to take advantage of such sentiment analysis to examine the negative comments or unpleasant experiences of tourists and making appropriate improvements. Moreover, it will help online users to get better recommendations while planning their trips.

酒店评论在旅游推荐系统中起着至关重要的作用。应有效地分析这些评论,以提高推荐的准确性,而推荐的准确性可以通过固定比例的清晰评分或评论的真实情感来生成。但清晰的评分并不能代表评论者的真实感受。现有的旅游推荐系统大多根据模糊和稀疏的评分来推荐酒店,结果导致对在线用户的推荐或偏好不准确。本文提出了一种语义方法来分析从 tripadvisor.in 抓取的在线评论。它能根据酒店的多种标准发现评论的基本模糊语义,而不是使用清晰的评分。抓取到的评论会经过数据清理预处理,如删除停顿词和标点符号、标记化、词法化、pos 标记等,以便有效地理解语义。从预处理后的评论中提取代表酒店常见特征的名词,并进一步用于识别意见短语。根据频繁出现的名词的归一化频率得出模糊权重,并结合已识别意见短语中所有形容词同义词的情感得分。这就形成了模糊语义,为多标准旅游推荐系统提供了理想的评论表示。通过从 TripAdvisor 抓取斋浦尔酒店最近的评论并分析其语义,用 python 实现了提议的工作。由此产生的模糊语义形成了一个人工标记的评论数据集,分别标记了已识别方面的情感。实验结果表明,在考虑所有形容词同义词的情况下,情感评分有所提高。实验结果进一步用于微调 BERT 模型,为基于查询的推荐系统形成编码。所提出的方法可以帮助旅游和酒店服务提供商利用这种情感分析来检查游客的负面评论或不愉快经历,并做出适当的改进。此外,它还能帮助在线用户在规划旅行时获得更好的推荐。
{"title":"Analyzing fuzzy semantics of reviews for multi-criteria recommendations","authors":"Navreen Kaur Boparai ,&nbsp;Himanshu Aggarwal ,&nbsp;Rinkle Rani","doi":"10.1016/j.datak.2024.102314","DOIUrl":"10.1016/j.datak.2024.102314","url":null,"abstract":"<div><p>Hotel reviews play a vital role in tourism recommender system. They should be analyzed effectively to enhance the accuracy of recommendations which can be generated either from crisp ratings on a fixed scale or real sentiments of reviews. But crisp ratings cannot represent the actual feelings of reviewers. Existing tourism recommender systems mostly recommend hotels on the basis of vague and sparse ratings resulting in inaccurate recommendations or preferences for online users. This paper presents a semantic approach to analyze the online reviews being crawled from tripadvisor.in. It discovers the underlying fuzzy semantics of reviews with respect to the multiple criteria of hotels rather than using the crisp ratings. The crawled reviews are preprocessed via data cleaning such as stopword and punctuation removal, tokenization, lemmatization, pos tagging to understand the semantics efficiently. Nouns representing frequent features of hotels are extracted from pre-processed reviews which are further used to identify opinion phrases. Fuzzy weights are derived from normalized frequency of frequent nouns and combined with sentiment score of all the synonyms of adjectives in the identified opinion phrases. This results in fuzzy semantics which form an ideal representation of reviews for a multi-criteria tourism recommender system. The proposed work is implemented in python by crawling the recent reviews of Jaipur hotels from TripAdvisor and analyzing their semantics. The resultant fuzzy semantics form a manually tagged dataset of reviews tagged with sentiments of identified aspects, respectively. Experimental results show improved sentiment score while considering all the synonyms of adjectives. The results are further used to fine-tune BERT models to form encodings for a query-based recommender system. The proposed approach can help tourism and hospitality service providers to take advantage of such sentiment analysis to examine the negative comments or unpleasant experiences of tourists and making appropriate improvements. Moreover, it will help online users to get better recommendations while planning their trips.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"152 ","pages":"Article 102314"},"PeriodicalIF":2.5,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141034319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Business intelligence and cognitive loads: Proposition of a dashboard adoption model 商业智能与认知负荷:仪表盘应用模型的提出
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-11 DOI: 10.1016/j.datak.2024.102310
Corentin Burnay, Mathieu Lega, Sarah Bouraga

Decision makers in organizations strive to improve the quality of their decisions. One way to improve that process is to objectify the decisions with facts. Data-driven Decision Support Systems (data-driven DSS), and more specifically business intelligence (BI) intend to achieve this. Organizations invest massively in the development of BI data-driven DSS and expect them to be adopted and to effectively support decision makers. This raises many technical and methodological challenges, especially regarding the design of BI dashboards, which can be seen as the visible tip of the BI data-driven DSS iceberg and which play a major role in the adoption of the entire system. In this paper, the dashboard content is investigated as one possible root cause for BI data-driven DSS dashboard adoption or rejection through an early empirical research. More precisely, this work is composed of three parts. In the first part, the concept of cognitive loads is studied in the context of BI dashboards and the informational, the representational and the non-informational loads are introduced. In the second part, the effects of these loads on the adoption of BI dashboards are then studied through an experiment with 167 respondents and a Structural Equation Modeling (SEM) analysis. The result is a Dashboard Adoption Model, enriching the seminal Technology Acceptance Model with new content-oriented variables to support the design of more supportive BI data-driven DSS dashboards. Finally, in the third part, a set of indicators is proposed to help dashboards designers in the monitoring of the loads of their dashboards practically.

组织中的决策者都在努力提高决策质量。改进这一过程的方法之一就是用事实将决策客观化。数据驱动的决策支持系统(DSS),更具体地说就是商业智能(BI),就是为了实现这一目标。各组织在开发商业智能数据驱动型决策支持系统方面投入了大量资金,并期望这些系统能够被采用并为决策者提供有效支持。这就提出了许多技术和方法上的挑战,尤其是在商业智能仪表盘的设计方面,它可以被视为商业智能数据驱动型数据支持系统的冰山一角,在整个系统的采用方面发挥着重要作用。本文通过早期实证研究,将仪表盘内容作为 BI 数据驱动的 DSS 仪表盘采用或拒绝的可能根源之一进行调查。更确切地说,这项工作由三部分组成。在第一部分中,研究了 BI 面板背景下的认知负荷概念,并介绍了信息负荷、表征负荷和非信息负荷。在第二部分中,通过对 167 名受访者进行实验和结构方程建模(SEM)分析,研究了这些负载对采用商业智能仪表盘的影响。研究结果是仪表盘采用模型,该模型用新的内容导向变量丰富了开创性的技术接受模型,以支持设计更具支持性的商业智能数据驱动的 DSS 仪表盘。最后,第三部分提出了一套指标,以帮助仪表盘设计者切实监测仪表盘的负载情况。
{"title":"Business intelligence and cognitive loads: Proposition of a dashboard adoption model","authors":"Corentin Burnay,&nbsp;Mathieu Lega,&nbsp;Sarah Bouraga","doi":"10.1016/j.datak.2024.102310","DOIUrl":"https://doi.org/10.1016/j.datak.2024.102310","url":null,"abstract":"<div><p>Decision makers in organizations strive to improve the quality of their decisions. One way to improve that process is to objectify the decisions with facts. Data-driven Decision Support Systems (data-driven DSS), and more specifically business intelligence (BI) intend to achieve this. Organizations invest massively in the development of BI data-driven DSS and expect them to be adopted and to effectively support decision makers. This raises many technical and methodological challenges, especially regarding the design of BI dashboards, which can be seen as the visible tip of the BI data-driven DSS iceberg and which play a major role in the adoption of the entire system. In this paper, the dashboard content is investigated as one possible root cause for BI data-driven DSS dashboard adoption or rejection through an early empirical research. More precisely, this work is composed of three parts. In the first part, the concept of cognitive loads is studied in the context of BI dashboards and the informational, the representational and the non-informational loads are introduced. In the second part, the effects of these loads on the adoption of BI dashboards are then studied through an experiment with 167 respondents and a Structural Equation Modeling (SEM) analysis. The result is a Dashboard Adoption Model, enriching the seminal Technology Acceptance Model with new content-oriented variables to support the design of more supportive BI data-driven DSS dashboards. Finally, in the third part, a set of indicators is proposed to help dashboards designers in the monitoring of the loads of their dashboards practically.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"152 ","pages":"Article 102310"},"PeriodicalIF":2.5,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140951807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning for predicting off-block delays: A case study at Paris — Charles de Gaulle International Airport 预测非阻塞延误的机器学习:巴黎戴高乐国际机场案例研究
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-08 DOI: 10.1016/j.datak.2024.102303
Thibault Falque , Bertrand Mazure , Karim Tabia

Punctuality is a sensitive issue in large airports and hubs for passenger experience and for controlling operational costs. This paper presents a real and challenging problem of predicting and explaining flight off-block delays. We study the case of the international airport Paris Charles de Gaulle (Paris-CDG) starting from the specificities of this problem at Paris-CDG until the proposal of modelings then solutions and the analysis of the results on real data covering an entire year of activity. The proof of concept provided in this paper allows us to believe that the proposed approach could help improve the management of delays and reduce the impact of the resulting consequences.

准点率是大型机场和枢纽的一个敏感问题,关系到乘客体验和运营成本控制。本文提出了一个具有挑战性的实际问题,即如何预测和解释航班延误。我们以巴黎戴高乐国际机场(Paris Charles de Gaulle,简称 "Paris-CDG")为例进行研究,从巴黎戴高乐机场这一问题的特殊性入手,到提出模型和解决方案,再到对全年活动的真实数据进行结果分析。本文所提供的概念证明让我们相信,所提出的方法有助于改善延误管理并减少由此造成的影响。
{"title":"Machine learning for predicting off-block delays: A case study at Paris — Charles de Gaulle International Airport","authors":"Thibault Falque ,&nbsp;Bertrand Mazure ,&nbsp;Karim Tabia","doi":"10.1016/j.datak.2024.102303","DOIUrl":"10.1016/j.datak.2024.102303","url":null,"abstract":"<div><p>Punctuality is a sensitive issue in large airports and hubs for passenger experience and for controlling operational costs. This paper presents a real and challenging problem of predicting and explaining flight off-block delays. We study the case of the international airport Paris Charles de Gaulle (Paris-CDG) starting from the specificities of this problem at Paris-CDG until the proposal of modelings then solutions and the analysis of the results on real data covering an entire year of activity. The proof of concept provided in this paper allows us to believe that the proposed approach could help improve the management of delays and reduce the impact of the resulting consequences.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"152 ","pages":"Article 102303"},"PeriodicalIF":2.5,"publicationDate":"2024-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0169023X24000272/pdfft?md5=ff8c7468240914b3ce61469a0954468c&pid=1-s2.0-S0169023X24000272-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141043841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
To prompt or not to prompt: Navigating the use of Large Language Models for integrating and modeling heterogeneous data 提示还是不提示?使用大型语言模型对异构数据进行整合和建模的导航
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-06 DOI: 10.1016/j.datak.2024.102313
Adel Remadi , Karim El Hage , Yasmina Hobeika , Francesca Bugiotti

Manually integrating data of diverse formats and languages is vital to many artificial intelligence applications. However, the task itself remains challenging and time-consuming. This paper highlights the potential of Large Language Models (LLMs) to streamline data extraction and resolution processes. Our approach aims to address the ongoing challenge of integrating heterogeneous data sources, encouraging advancements in the field of data engineering. Applied on the specific use case of learning disorders in higher education, our research demonstrates LLMs’ capability to effectively extract data from unstructured sources. It is then further highlighted that LLMs can enhance data integration by providing the ability to resolve entities originating from multiple data sources. Crucially, the paper underscores the necessity of preliminary data modeling decisions to ensure the success of such technological applications. By merging human expertise with LLM-driven automation, this study advocates for the further exploration of semi-autonomous data engineering pipelines.

手动整合不同格式和语言的数据对许多人工智能应用来说都至关重要。然而,这项任务本身仍然具有挑战性且耗时。本文强调了大型语言模型(LLM)在简化数据提取和解析过程方面的潜力。我们的方法旨在解决整合异构数据源的持续挑战,推动数据工程领域的进步。我们的研究以高等教育中的学习障碍为特定用例,展示了 LLMs 从非结构化数据源中有效提取数据的能力。论文还进一步强调,LLM 可以提供解析来自多个数据源的实体的能力,从而加强数据整合。最重要的是,本文强调了初步数据建模决策的必要性,以确保此类技术应用的成功。通过将人类专业知识与 LLM 驱动的自动化相结合,本研究主张进一步探索半自主数据工程管道。
{"title":"To prompt or not to prompt: Navigating the use of Large Language Models for integrating and modeling heterogeneous data","authors":"Adel Remadi ,&nbsp;Karim El Hage ,&nbsp;Yasmina Hobeika ,&nbsp;Francesca Bugiotti","doi":"10.1016/j.datak.2024.102313","DOIUrl":"https://doi.org/10.1016/j.datak.2024.102313","url":null,"abstract":"<div><p>Manually integrating data of diverse formats and languages is vital to many artificial intelligence applications. However, the task itself remains challenging and time-consuming. This paper highlights the potential of Large Language Models (LLMs) to streamline data extraction and resolution processes. Our approach aims to address the ongoing challenge of integrating heterogeneous data sources, encouraging advancements in the field of data engineering. Applied on the specific use case of learning disorders in higher education, our research demonstrates LLMs’ capability to effectively extract data from unstructured sources. It is then further highlighted that LLMs can enhance data integration by providing the ability to resolve entities originating from multiple data sources. Crucially, the paper underscores the necessity of preliminary data modeling decisions to ensure the success of such technological applications. By merging human expertise with LLM-driven automation, this study advocates for the further exploration of semi-autonomous data engineering pipelines.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"152 ","pages":"Article 102313"},"PeriodicalIF":2.5,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0169023X24000375/pdfft?md5=11ee9c76542d55fac49075892a9a8c7d&pid=1-s2.0-S0169023X24000375-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140918204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using Berlin SPARQL benchmark to evaluate virtual SPARQL endpoints over relational databases 使用柏林 SPARQL 基准评估关系数据库上的虚拟 SPARQL 端点
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-03 DOI: 10.1016/j.datak.2024.102309
Milos Chaloupka, Martin Necasky

The RDF is a popular and well-documented format for publishing structured data on the web. It enables data to be consumed without the knowledge of how the data is internally stored. There are already several native RDF storage solutions that provide a SPARQL endpoint. However, native RDF stores are not widely adopted. It is still more common to store data in a relational database. One of the useful features of native RDF storage solutions is providing a SPARQL endpoint, a web service to query RDF data with SPARQL. To provide this feature also on top of prevalent relational databases, solutions for virtual SPARQL endpoints on top of a relational database have appeared. To benchmark these solutions, a state-of-the-art tool, the Berlin SPARQL Benchmark (BSBM), is used. However, BSBM was designed primarily to benchmark native RDF stores. It can also be used to benchmark solutions for virtual SPARQL endpoints. However, since BSBM was not designed for virtual SPARQL endpoints, each implementation uses that tool differently for evaluation. As a result, the evaluation is not consistent and therefore hardly comparable. In this paper, we demonstrate how this well-defined benchmarking tool for SPARQL endpoints can be used to evaluate virtual endpoints over relational databases, perform the evaluation on the available implementations, and provide instructions on how to repeat the same evaluation in the future.

RDF 是一种在网络上发布结构化数据的流行且文档齐全的格式。它能让数据在不知道数据内部存储方式的情况下被消费。目前已经有几种本地 RDF 存储解决方案提供 SPARQL 端点。然而,本地 RDF 存储并没有被广泛采用。将数据存储在关系数据库中仍然更为常见。原生 RDF 存储解决方案的一个有用功能是提供 SPARQL 端点,即使用 SPARQL 查询 RDF 数据的网络服务。为了在流行的关系数据库之上也能提供这一功能,出现了在关系数据库之上提供虚拟 SPARQL 端点的解决方案。为了对这些解决方案进行基准测试,我们使用了最先进的工具--柏林 SPARQL 基准(BSBM)。不过,BSBM 主要是为本地 RDF 存储基准而设计的。它也可用于对虚拟 SPARQL 端点的解决方案进行基准测试。但是,由于 BSBM 不是为虚拟 SPARQL 端点设计的,因此每个实施方案使用该工具进行评估的方式都不同。因此,评估结果并不一致,很难进行比较。在本文中,我们将演示如何使用这一定义明确的 SPARQL 端点基准测试工具来评估关系数据库上的虚拟端点,对可用的实现进行评估,并就将来如何重复相同的评估提供指导。
{"title":"Using Berlin SPARQL benchmark to evaluate virtual SPARQL endpoints over relational databases","authors":"Milos Chaloupka,&nbsp;Martin Necasky","doi":"10.1016/j.datak.2024.102309","DOIUrl":"https://doi.org/10.1016/j.datak.2024.102309","url":null,"abstract":"<div><p>The RDF is a popular and well-documented format for publishing structured data on the web. It enables data to be consumed without the knowledge of how the data is internally stored. There are already several native RDF storage solutions that provide a SPARQL endpoint. However, native RDF stores are not widely adopted. It is still more common to store data in a relational database. One of the useful features of native RDF storage solutions is providing a SPARQL endpoint, a web service to query RDF data with SPARQL. To provide this feature also on top of prevalent relational databases, solutions for virtual SPARQL endpoints on top of a relational database have appeared. To benchmark these solutions, a state-of-the-art tool, the Berlin SPARQL Benchmark (BSBM), is used. However, BSBM was designed primarily to benchmark native RDF stores. It can also be used to benchmark solutions for virtual SPARQL endpoints. However, since BSBM was not designed for virtual SPARQL endpoints, each implementation uses that tool differently for evaluation. As a result, the evaluation is not consistent and therefore hardly comparable. In this paper, we demonstrate how this well-defined benchmarking tool for SPARQL endpoints can be used to evaluate virtual endpoints over relational databases, perform the evaluation on the available implementations, and provide instructions on how to repeat the same evaluation in the future.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"152 ","pages":"Article 102309"},"PeriodicalIF":2.5,"publicationDate":"2024-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140905621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective text classification using BERT, MTM LSTM, and DT 使用 BERT、MTM LSTM 和 DT 进行有效的文本分类
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-01 DOI: 10.1016/j.datak.2024.102306
Saman Jamshidi , Mahin Mohammadi , Saeed Bagheri , Hamid Esmaeili Najafabadi , Alireza Rezvanian , Mehdi Gheisari , Mustafa Ghaderzadeh , Amir Shahab Shahabi , Zongda Wu

Text classification plays a critical role in managing large volumes of electronically produced texts. As the number of such texts increases, manual analysis becomes impractical, necessitating an intelligent approach for processing information. Deep learning models have witnessed widespread application in text classification, including the use of recurrent neural networks like Many to One Long Short-Term Memory (MTO LSTM). Nonetheless, this model is limited by its reliance on only the last token for text labelling. To overcome this limitation, this study introduces a novel hybrid model that combines Bidirectional Encoder Representations from Transformers (BERT), Many To Many Long Short-Term Memory (MTM LSTM), and Decision Templates (DT) for text classification. In this new model, the text is first embedded using the BERT model and then trained using MTM LSTM to approximate the target at each token. Finally, the approximations are fused using DT. The proposed model is evaluated using the well-known IMDB movie review dataset for binary classification and Drug Review Dataset for multiclass classification. The results demonstrate superior performance in terms of accuracy, recall, precision, and F1 score compared to previous models. The hybrid model presented in this study holds significant potential for a wide range of text classification tasks and stands as a valuable contribution to the field.

文本分类在管理大量电子文本方面发挥着至关重要的作用。随着此类文本数量的增加,人工分析变得不切实际,因此需要一种智能方法来处理信息。深度学习模型在文本分类中得到了广泛应用,包括使用多对一长短时记忆(MTO LSTM)等递归神经网络。然而,这种模型的局限性在于仅依赖最后一个标记进行文本标注。为了克服这一局限,本研究引入了一种新型混合模型,该模型结合了变压器双向编码器表征(BERT)、多对多长短期记忆(MTM LSTM)和决策模板(DT),用于文本分类。在这个新模型中,首先使用 BERT 模型嵌入文本,然后使用 MTM LSTM 进行训练,以近似每个标记的目标。最后,使用 DT 对近似值进行融合。我们使用著名的 IMDB 电影评论数据集进行了二分类评估,并使用药物评论数据集进行了多分类评估。结果表明,与之前的模型相比,该模型在准确率、召回率、精确度和 F1 分数等方面都表现出色。本研究提出的混合模型在广泛的文本分类任务中具有巨大的潜力,是对该领域的宝贵贡献。
{"title":"Effective text classification using BERT, MTM LSTM, and DT","authors":"Saman Jamshidi ,&nbsp;Mahin Mohammadi ,&nbsp;Saeed Bagheri ,&nbsp;Hamid Esmaeili Najafabadi ,&nbsp;Alireza Rezvanian ,&nbsp;Mehdi Gheisari ,&nbsp;Mustafa Ghaderzadeh ,&nbsp;Amir Shahab Shahabi ,&nbsp;Zongda Wu","doi":"10.1016/j.datak.2024.102306","DOIUrl":"https://doi.org/10.1016/j.datak.2024.102306","url":null,"abstract":"<div><p>Text classification plays a critical role in managing large volumes of electronically produced texts. As the number of such texts increases, manual analysis becomes impractical, necessitating an intelligent approach for processing information. Deep learning models have witnessed widespread application in text classification, including the use of recurrent neural networks like Many to One Long Short-Term Memory (MTO LSTM). Nonetheless, this model is limited by its reliance on only the last token for text labelling. To overcome this limitation, this study introduces a novel hybrid model that combines Bidirectional Encoder Representations from Transformers (BERT), Many To Many Long Short-Term Memory (MTM LSTM), and Decision Templates (DT) for text classification. In this new model, the text is first embedded using the BERT model and then trained using MTM LSTM to approximate the target at each token. Finally, the approximations are fused using DT. The proposed model is evaluated using the well-known IMDB movie review dataset for binary classification and Drug Review Dataset for multiclass classification. The results demonstrate superior performance in terms of accuracy, recall, precision, and F1 score compared to previous models. The hybrid model presented in this study holds significant potential for a wide range of text classification tasks and stands as a valuable contribution to the field.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"151 ","pages":"Article 102306"},"PeriodicalIF":2.5,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140825257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating Transformers and Linguistic Features integration for Author Profiling tasks in Spanish 评估转换器和语言特征整合在西班牙语作者分析任务中的应用
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-01 DOI: 10.1016/j.datak.2024.102307
José Antonio García-Díaz , Ghassan Beydoun , Rafel Valencia-García

Author profiling consists of extracting their demographic and psychographic information by examining their writings. This information can then be used to improve the reader experience and to detect bots or propagators of hoaxes and/or hate speech. Therefore, author profiling can be applied to build more robust and efficient Knowledge-Based Systems for tasks such as content moderation, user profiling, and information retrieval. Author profiling is typically performed automatically as a document classification task. Recently, language models based on transformers have also proven to be quite effective in this task. However, the size and heterogeneity of novel language models, makes it necessary to evaluate them in context. The contributions we make in this paper are four-fold: First, we evaluate which language models are best suited to perform author profiling in Spanish. These experiments include basic, distilled, and multilingual models. Second, we evaluate how feature integration can improve performance for this task. We evaluate two distinct strategies: knowledge integration and ensemble learning. Third, we evaluate the ability of linguistic features to improve the interpretability of the results. Fourth, we evaluate the performance of each language model in terms of memory, training, and inference times. Our results indicate that the use of lightweight models can indeed achieve similar performance to heavy models and that multilingual models are actually less effective than models trained with one language. Finally, we confirm that the best models and strategies for integrating features ultimately depend on the context of the task.

作者分析包括通过研究作者的著作来提取其人口统计学和心理学信息。这些信息可用于改善读者体验,检测机器人或恶作剧和/或仇恨言论的传播者。因此,作者特征描述可用于为内容管理、用户特征描述和信息检索等任务构建更强大、更高效的知识型系统。作者特征描述通常作为文档分类任务自动执行。最近,基于转换器的语言模型也被证明在这项任务中相当有效。然而,由于新型语言模型的规模和异质性,有必要在上下文中对其进行评估。我们在本文中做出了四方面的贡献:首先,我们评估了哪些语言模型最适合在西班牙语中执行作者剖析。这些实验包括基本模型、提炼模型和多语言模型。其次,我们评估了特征整合如何提高这项任务的性能。我们评估了两种不同的策略:知识整合和集合学习。第三,我们评估了语言特征提高结果可解释性的能力。第四,我们评估了每个语言模型在内存、训练和推理时间方面的性能。我们的结果表明,使用轻量级模型确实可以达到与重型模型相似的性能,而多语言模型的效果实际上不如用一种语言训练的模型。最后,我们证实,整合特征的最佳模型和策略最终取决于任务的背景。
{"title":"Evaluating Transformers and Linguistic Features integration for Author Profiling tasks in Spanish","authors":"José Antonio García-Díaz ,&nbsp;Ghassan Beydoun ,&nbsp;Rafel Valencia-García","doi":"10.1016/j.datak.2024.102307","DOIUrl":"https://doi.org/10.1016/j.datak.2024.102307","url":null,"abstract":"<div><p>Author profiling consists of extracting their demographic and psychographic information by examining their writings. This information can then be used to improve the reader experience and to detect bots or propagators of hoaxes and/or hate speech. Therefore, author profiling can be applied to build more robust and efficient Knowledge-Based Systems for tasks such as content moderation, user profiling, and information retrieval. Author profiling is typically performed automatically as a document classification task. Recently, language models based on transformers have also proven to be quite effective in this task. However, the size and heterogeneity of novel language models, makes it necessary to evaluate them in context. The contributions we make in this paper are four-fold: First, we evaluate which language models are best suited to perform author profiling in Spanish. These experiments include basic, distilled, and multilingual models. Second, we evaluate how feature integration can improve performance for this task. We evaluate two distinct strategies: knowledge integration and ensemble learning. Third, we evaluate the ability of linguistic features to improve the interpretability of the results. Fourth, we evaluate the performance of each language model in terms of memory, training, and inference times. Our results indicate that the use of lightweight models can indeed achieve similar performance to heavy models and that multilingual models are actually less effective than models trained with one language. Finally, we confirm that the best models and strategies for integrating features ultimately depend on the context of the task.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"151 ","pages":"Article 102307"},"PeriodicalIF":2.5,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0169023X24000314/pdfft?md5=42a482dbed2e2a640c46e89a6f3a69c8&pid=1-s2.0-S0169023X24000314-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140825258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Managerial risk data analytics applications using grey influence analysis (GINA) 利用灰色影响分析(GINA)进行管理风险数据分析应用
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-01 DOI: 10.1016/j.datak.2024.102312
R. Rajesh

We observe and analyze the causal relations among risk factors in a system, considering the manufacturing supply chains. Seven major categories of risks were identified and scrutinized and the detailed analysis of causal relations using the grey influence analysis (GINA) methodology is outlined. With expert response based survey, we conduct an initial analysis of the risks using risk matrix analysis (RMA) and the risks under high priority are identified. Later, the GINA is implemented to understand the causal relations among various categories of risks, which is particularly useful in group decision-making environments. The results from RMA concludes that the capacity risks (CR) and delays (DL) are in the category of very high priority risks. GINA results also ratify the conclusions from RMA and observes that managers need to control and manage capacity risks (CR) and delays (DL) with high priorities. Additionally from the results of GINA, the causal factors disruptions (DS) and forecast risks (FR) appear to be primary importance and if unattended can lead to the initiation of several other risks in supply chains. Managers are recommended to identify disruptions at an early stage in supply chains and reduce the forecast errors to avoid bullwhips in supply chains.

考虑到制造业供应链,我们观察并分析了系统中风险因素之间的因果关系。我们确定并审查了七大类风险,并使用灰色影响分析(GINA)方法对因果关系进行了详细分析。通过基于专家回复的调查,我们使用风险矩阵分析法(RMA)对风险进行了初步分析,并确定了高度优先的风险。随后,我们采用灰色分析法来了解各类风险之间的因果关系,这在群体决策环境中尤为有用。RMA 分析结果表明,能力风险(CR)和延误风险(DL)属于高度优先风险。GINA 的结果也验证了 RMA 的结论,并指出管理人员需要以高度优先的方式控制和管理产能风险(CR)和延误风险(DL)。此外,从 GINA 的结果来看,因果因素中断(DS)和预测风险(FR)似乎是最重要的,如果不加注意,可能会导致供应链中其他一些风险的发生。建议管理者及早识别供应链中的中断,减少预测误差,避免供应链中的牛鞭现象。
{"title":"Managerial risk data analytics applications using grey influence analysis (GINA)","authors":"R. Rajesh","doi":"10.1016/j.datak.2024.102312","DOIUrl":"https://doi.org/10.1016/j.datak.2024.102312","url":null,"abstract":"<div><p>We observe and analyze the causal relations among risk factors in a system, considering the manufacturing supply chains. Seven major categories of risks were identified and scrutinized and the detailed analysis of causal relations using the grey influence analysis (GINA) methodology is outlined. With expert response based survey, we conduct an initial analysis of the risks using risk matrix analysis (RMA) and the risks under high priority are identified. Later, the GINA is implemented to understand the causal relations among various categories of risks, which is particularly useful in group decision-making environments. The results from RMA concludes that the <em>capacity risks (CR)</em> and <em>delays (DL)</em> are in the category of very high priority risks. GINA results also ratify the conclusions from RMA and observes that managers need to control and manage <em>capacity risks (CR)</em> and <em>delays (DL)</em> with high priorities. Additionally from the results of GINA, the causal factors <em>disruptions (DS)</em> and <em>forecast risks (FR)</em> appear to be primary importance and if unattended can lead to the initiation of several other risks in supply chains. Managers are recommended to identify disruptions at an early stage in supply chains and reduce the forecast errors to avoid bullwhips in supply chains.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"151 ","pages":"Article 102312"},"PeriodicalIF":2.5,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140879377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A graph based named entity disambiguation using clique partitioning and semantic relatedness 利用小块分割和语义相关性进行基于图的命名实体消歧
IF 2.5 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-30 DOI: 10.1016/j.datak.2024.102308
Ramla Belalta , Mouhoub Belazzoug , Farid Meziane

Disambiguating name mentions in texts is a crucial task in Natural Language Processing, especially in entity linking. The credibility and efficiency of such systems depend largely on this task. For a given name entity mention in a text, there are many potential candidate entities that may refer to it in the knowledge base. Therefore, it is very difficult to assign the correct candidate from the whole set of candidate entities of this mention. To solve this problem, collective entity disambiguation is a prominent approach. In this paper, we present a novel algorithm called CPSR for collective entity disambiguation, which is based on a graph approach and semantic relatedness. A clique partitioning algorithm is used to find the best clique that contains a set of candidate entities. These candidate entities provide the answers to the corresponding mentions in the disambiguation process. To evaluate our algorithm, we carried out a series of experiments on seven well-known datasets, namely, AIDA/CoNLL2003-TestB, IITB, MSNBC, AQUAINT, ACE2004, Cweb, and Wiki. The Kensho Derived Wikimedia Dataset (KDWD) is used as the knowledge base for our system. From the experimental results, our CPSR algorithm outperforms both the baselines and other well-known state-of-the-art approaches.

对文本中提到的名称进行消歧是自然语言处理中的一项重要任务,尤其是在实体链接中。此类系统的可信度和效率在很大程度上取决于这项任务。对于文本中提到的某个名称实体,知识库中可能有许多潜在的候选实体。因此,要从这一提及的全部候选实体中指定正确的候选实体是非常困难的。为了解决这个问题,集体实体消歧是一种突出的方法。本文提出了一种名为 CPSR 的新型集体实体消歧算法,该算法基于图方法和语义相关性。该算法基于图方法和语义相关性,使用簇划分算法来找到包含一组候选实体的最佳簇。这些候选实体在消歧过程中为相应的提及提供答案。为了评估我们的算法,我们在七个著名的数据集上进行了一系列实验,即 AIDA/CoNLL2003-TestB、IITB、MSNBC、AQUAINT、ACE2004、Cweb 和 Wiki。Kensho Derived Wikimedia Dataset (KDWD) 被用作我们系统的知识库。从实验结果来看,我们的 CPSR 算法优于基线和其他著名的先进方法。
{"title":"A graph based named entity disambiguation using clique partitioning and semantic relatedness","authors":"Ramla Belalta ,&nbsp;Mouhoub Belazzoug ,&nbsp;Farid Meziane","doi":"10.1016/j.datak.2024.102308","DOIUrl":"https://doi.org/10.1016/j.datak.2024.102308","url":null,"abstract":"<div><p>Disambiguating name mentions in texts is a crucial task in Natural Language Processing, especially in entity linking. The credibility and efficiency of such systems depend largely on this task. For a given name entity mention in a text, there are many potential candidate entities that may refer to it in the knowledge base. Therefore, it is very difficult to assign the correct candidate from the whole set of candidate entities of this mention. To solve this problem, collective entity disambiguation is a prominent approach. In this paper, we present a novel algorithm called CPSR for collective entity disambiguation, which is based on a graph approach and semantic relatedness. A clique partitioning algorithm is used to find the best clique that contains a set of candidate entities. These candidate entities provide the answers to the corresponding mentions in the disambiguation process. To evaluate our algorithm, we carried out a series of experiments on seven well-known datasets, namely, AIDA/CoNLL2003-TestB, IITB, MSNBC, AQUAINT, ACE2004, Cweb, and Wiki. The Kensho Derived Wikimedia Dataset (KDWD) is used as the knowledge base for our system. From the experimental results, our CPSR algorithm outperforms both the baselines and other well-known state-of-the-art approaches.</p></div>","PeriodicalId":55184,"journal":{"name":"Data & Knowledge Engineering","volume":"152 ","pages":"Article 102308"},"PeriodicalIF":2.5,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140901817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Data & Knowledge Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1