首页 > 最新文献

Journal of Big Data最新文献

英文 中文
Exploring AI-driven approaches for unstructured document analysis and future horizons 探索人工智能驱动的非结构化文件分析方法及未来展望
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-07-05 DOI: 10.1186/s40537-024-00948-z
Supriya V. Mahadevkar, Shruti Patil, Ketan Kotecha, Lim Way Soong, Tanupriya Choudhury

In the current industrial landscape, a significant number of sectors are grappling with the challenges posed by unstructured data, which incurs financial losses amounting to millions annually. If harnessed effectively, this data has the potential to substantially boost operational efficiency. Traditional methods for extracting information have their limitations; however, solutions powered by artificial intelligence (AI) could provide a more fitting alternative. There is an evident gap in scholarly research concerning a comprehensive evaluation of AI-driven techniques for the extraction of information from unstructured content. This systematic literature review aims to identify, assess, and deliberate on prospective research directions within the field of unstructured document information extraction. It has been observed that prevailing extraction methods primarily depend on static patterns or rules, often proving inadequate when faced with complex document structures typically encountered in real-world scenarios, such as medical records. Datasets currently available to the public suffer from low quality and are tailored for specific tasks only. This underscores an urgent need for developing new datasets that accurately reflect complex issues encountered in practical settings. The review reveals that AI-based techniques show promise in autonomously extracting information from diverse unstructured documents, encompassing both printed and handwritten text. Challenges arise, however, when dealing with varied document layouts. Proposing a framework through hybrid AI-based approaches, this review envisions processing a high-quality dataset for automatic information extraction from unstructured documents. Additionally, it emphasizes the importance of collaborative efforts between organizations and researchers to address the diverse challenges associated with unstructured data analysis.

在当前的工业领域,许多部门都在努力应对非结构化数据带来的挑战,这些数据每年造成的经济损失高达数百万美元。如果能有效利用这些数据,就有可能大幅提高运营效率。传统的信息提取方法有其局限性;然而,人工智能(AI)驱动的解决方案可以提供更合适的替代方案。在全面评估从非结构化内容中提取信息的人工智能驱动技术方面,学术研究明显存在空白。本系统性文献综述旨在确定、评估和审议非结构化文档信息提取领域的前瞻性研究方向。据观察,目前流行的信息提取方法主要依赖于静态模式或规则,当面对现实世界中常见的复杂文档结构(如医疗记录)时,这些方法往往显得力不从心。目前向公众提供的数据集质量不高,而且只针对特定任务。这突出表明,迫切需要开发新的数据集,以准确反映实际环境中遇到的复杂问题。综述显示,基于人工智能的技术在从各种非结构化文档(包括印刷文本和手写文本)中自主提取信息方面大有可为。然而,在处理不同的文档布局时会遇到挑战。本综述通过基于人工智能的混合方法提出了一个框架,设想处理一个高质量的数据集,以便从非结构化文档中自动提取信息。此外,它还强调了组织和研究人员合作应对与非结构化数据分析相关的各种挑战的重要性。
{"title":"Exploring AI-driven approaches for unstructured document analysis and future horizons","authors":"Supriya V. Mahadevkar, Shruti Patil, Ketan Kotecha, Lim Way Soong, Tanupriya Choudhury","doi":"10.1186/s40537-024-00948-z","DOIUrl":"https://doi.org/10.1186/s40537-024-00948-z","url":null,"abstract":"<p>In the current industrial landscape, a significant number of sectors are grappling with the challenges posed by unstructured data, which incurs financial losses amounting to millions annually. If harnessed effectively, this data has the potential to substantially boost operational efficiency. Traditional methods for extracting information have their limitations; however, solutions powered by artificial intelligence (AI) could provide a more fitting alternative. There is an evident gap in scholarly research concerning a comprehensive evaluation of AI-driven techniques for the extraction of information from unstructured content. This systematic literature review aims to identify, assess, and deliberate on prospective research directions within the field of unstructured document information extraction. It has been observed that prevailing extraction methods primarily depend on static patterns or rules, often proving inadequate when faced with complex document structures typically encountered in real-world scenarios, such as medical records. Datasets currently available to the public suffer from low quality and are tailored for specific tasks only. This underscores an urgent need for developing new datasets that accurately reflect complex issues encountered in practical settings. The review reveals that AI-based techniques show promise in autonomously extracting information from diverse unstructured documents, encompassing both printed and handwritten text. Challenges arise, however, when dealing with varied document layouts. Proposing a framework through hybrid AI-based approaches, this review envisions processing a high-quality dataset for automatic information extraction from unstructured documents. Additionally, it emphasizes the importance of collaborative efforts between organizations and researchers to address the diverse challenges associated with unstructured data analysis.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"31 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141576733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
New custom rating for improving recommendation system performance 用于提高推荐系统性能的新自定义评级
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-07-02 DOI: 10.1186/s40537-024-00952-3
Tora Fahrudin, Dedy Rahman Wijaya

Recommendation system is currently attracting the interest of many explorers. Various new businesses have surfaced with the rise of online marketing (E-Commerce) in response to Covid-19 pandemic. This phenomenon allows recommendation items through a system called Collaborative Filtering (CF), aiming to improve shopping experience of users. Typically, the effectiveness of CF relies on the precise identification of similar profile users by similarity algorithms. Traditional similarity measures are based on the user-item rating matrix. Approximately, four custom ratings (CR) were used along with a new rating formula, termed New Custom Rating (NCR), derived from the popularity of users and items in addition to the original rating. Specifically, NCR optimized recommendation system performance by using the popularity of users and items to determine new ratings value, rather than solely relying on the original rating. Additionally, the formulas improved the representativeness of the new rating values and the accuracy of similarity algorithm calculations. Consequently, the increased accuracy of recommendation system was achieved. The implementation of NCR across four CR algorithms and recommendation system using five public datasets was examined. Consequently, the experimental results showed that NCR significantly increased recommendation system accuracy, as evidenced by reductions in RMSE, MSE, and MAE as well as increasing FCP and Hit Rate. Moreover, by combining the popularity of users and items into rating calculations, NCR improved the accuracy of various recommendation system algorithms reducing RMSE, MSE, and MAE up to 62.10%, 53.62%, 65.97%, respectively, while also increasing FCP and Hit Rate up to 11.89% and 31.42%, respectively.

推荐系统目前正吸引着众多探索者的兴趣。随着网络营销(电子商务)的兴起,各种新业务也随之浮出水面,以应对 Covid-19 的流行。这种现象允许通过一种称为协同过滤(CF)的系统来推荐商品,目的是改善用户的购物体验。通常情况下,CF 的有效性依赖于通过相似性算法精确识别相似资料用户。传统的相似性测量方法基于用户-物品评级矩阵。大约使用了四个自定义评级(CR)和一个新的评级公式,称为新自定义评级(NCR),新自定义评级是在原始评级的基础上,根据用户和物品的受欢迎程度得出的。具体来说,NCR 通过使用用户和项目的受欢迎程度来确定新的评级值,而不是仅仅依赖于原始评级,从而优化了推荐系统的性能。此外,这些公式还提高了新评分值的代表性和相似性算法计算的准确性。因此,推荐系统的准确性得到了提高。我们使用五个公共数据集对 NCR 在四种 CR 算法和推荐系统中的实施情况进行了检验。实验结果表明,NCR 显著提高了推荐系统的准确性,这体现在 RMSE、MSE 和 MAE 的降低以及 FCP 和命中率的提高上。此外,通过将用户和项目的受欢迎程度纳入评级计算,NCR 提高了各种推荐系统算法的准确性,RMSE、MSE 和 MAE 分别降低了 62.10%、53.62% 和 65.97%,FCP 和命中率也分别提高了 11.89% 和 31.42%。
{"title":"New custom rating for improving recommendation system performance","authors":"Tora Fahrudin, Dedy Rahman Wijaya","doi":"10.1186/s40537-024-00952-3","DOIUrl":"https://doi.org/10.1186/s40537-024-00952-3","url":null,"abstract":"<p>Recommendation system is currently attracting the interest of many explorers. Various new businesses have surfaced with the rise of online marketing (E-Commerce) in response to Covid-19 pandemic. This phenomenon allows recommendation items through a system called Collaborative Filtering (CF), aiming to improve shopping experience of users. Typically, the effectiveness of CF relies on the precise identification of similar profile users by similarity algorithms. Traditional similarity measures are based on the user-item rating matrix. Approximately, four custom ratings (CR) were used along with a new rating formula, termed New Custom Rating (NCR), derived from the popularity of users and items in addition to the original rating. Specifically, NCR optimized recommendation system performance by using the popularity of users and items to determine new ratings value, rather than solely relying on the original rating. Additionally, the formulas improved the representativeness of the new rating values and the accuracy of similarity algorithm calculations. Consequently, the increased accuracy of recommendation system was achieved. The implementation of NCR across four CR algorithms and recommendation system using five public datasets was examined. Consequently, the experimental results showed that NCR significantly increased recommendation system accuracy, as evidenced by reductions in RMSE, MSE, and MAE as well as increasing FCP and Hit Rate. Moreover, by combining the popularity of users and items into rating calculations, NCR improved the accuracy of various recommendation system algorithms reducing RMSE, MSE, and MAE up to 62.10%, 53.62%, 65.97%, respectively, while also increasing FCP and Hit Rate up to 11.89% and 31.42%, respectively.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"23 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141520245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimization-based convolutional neural model for the classification of white blood cells 基于优化的卷积神经模型用于白细胞分类
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-06-26 DOI: 10.1186/s40537-024-00949-y
Tulasi Gayatri Devi, Nagamma Patil

White blood cells (WBCs) are one of the most significant parts of the human immune system, and they play a crucial role in diagnosing the characteristics of pathologists and blood-related diseases. The characteristics of WBCs are well-defined based on the morphological behavior of their nuclei, and the number and types of WBCs can often determine the presence of diseases or illnesses. Generally, there are different types of WBCs, and the accurate classification of WBCs helps in proper diagnosis and treatment. Although various classification models were developed in the past, they face issues like less classification accuracy, high error rate, and large execution. Hence, a novel classification strategy named the African Buffalo-based Convolutional Neural Model (ABCNM) is proposed to classify the types of WBCs accurately. The proposed strategy commences with collecting WBC sample databases, which are preprocessed and trained into the system for classification. The preprocessing phase removes the noises and training flaws, which helps improve the dataset's quality and consistency. Further, feature extraction is performed to segment the WBCs, and African Buffalo fitness is updated in the classification layer for the correct classification of WBCs. The proposed framework is modeled in Python, and the experimental analysis depicts that it achieved 99.12% accuracy, 98.16% precision, 99% sensitivity, 99.04% specificity, and 99.02% f-measure. Furthermore, a comparative assessment with the existing techniques validated that the proposed strategy obtained better performances than the conventional models.

白细胞(WBC)是人体免疫系统中最重要的组成部分之一,在诊断病理特征和血液相关疾病方面起着至关重要的作用。白细胞的特征是根据其细胞核的形态行为明确界定的,白细胞的数量和类型往往可以判断疾病或病症的存在。一般来说,白细胞有不同的类型,对白细胞进行准确分类有助于正确诊断和治疗。虽然过去开发了各种分类模型,但它们都面临着分类准确率低、错误率高和执行量大等问题。因此,我们提出了一种名为 "基于非洲水牛的卷积神经模型"(ABCNM)的新型分类策略,用于对白细胞类型进行准确分类。所提出的策略首先要收集白细胞样本数据库,然后对其进行预处理,并训练系统进行分类。预处理阶段可去除噪音和训练缺陷,有助于提高数据集的质量和一致性。此外,还将进行特征提取以分割白细胞,并在分类层中更新非洲水牛的适配性,以便对白细胞进行正确分类。所提出的框架以 Python 为模型,实验分析表明其准确率达到 99.12%,精确率达到 98.16%,灵敏度达到 99%,特异性达到 99.04%,f-measure 达到 99.02%。此外,通过与现有技术的比较评估,验证了所提出的策略比传统模型获得了更好的性能。
{"title":"Optimization-based convolutional neural model for the classification of white blood cells","authors":"Tulasi Gayatri Devi, Nagamma Patil","doi":"10.1186/s40537-024-00949-y","DOIUrl":"https://doi.org/10.1186/s40537-024-00949-y","url":null,"abstract":"<p>White blood cells (WBCs) are one of the most significant parts of the human immune system, and they play a crucial role in diagnosing the characteristics of pathologists and blood-related diseases. The characteristics of WBCs are well-defined based on the morphological behavior of their nuclei, and the number and types of WBCs can often determine the presence of diseases or illnesses. Generally, there are different types of WBCs, and the accurate classification of WBCs helps in proper diagnosis and treatment. Although various classification models were developed in the past, they face issues like less classification accuracy, high error rate, and large execution. Hence, a novel classification strategy named the African Buffalo-based Convolutional Neural Model (ABCNM) is proposed to classify the types of WBCs accurately. The proposed strategy commences with collecting WBC sample databases, which are preprocessed and trained into the system for classification. The preprocessing phase removes the noises and training flaws, which helps improve the dataset's quality and consistency. Further, feature extraction is performed to segment the WBCs, and African Buffalo fitness is updated in the classification layer for the correct classification of WBCs. The proposed framework is modeled in Python, and the experimental analysis depicts that it achieved 99.12% accuracy, 98.16% precision, 99% sensitivity, 99.04% specificity, and 99.02% f-measure. Furthermore, a comparative assessment with the existing techniques validated that the proposed strategy obtained better performances than the conventional models.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"38 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141503493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature reduction for hepatocellular carcinoma prediction using machine learning algorithms 利用机器学习算法减少预测肝细胞癌的特征
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-06-18 DOI: 10.1186/s40537-024-00944-3
Ghada Mostafa, Hamdi Mahmoud, Tarek Abd El-Hafeez, Mohamed E. ElAraby

Hepatocellular carcinoma (HCC) is a highly prevalent form of liver cancer that necessitates accurate prediction models for early diagnosis and effective treatment. Machine learning algorithms have demonstrated promising results in various medical domains, including cancer prediction. In this study, we propose a comprehensive approach for HCC prediction by comparing the performance of different machine learning algorithms before and after applying feature reduction methods. We employ popular feature reduction techniques, such as weighting features, hidden features correlation, feature selection, and optimized selection, to extract a reduced feature subset that captures the most relevant information related to HCC. Subsequently, we apply multiple algorithms, including Naive Bayes, support vector machines (SVM), Neural Networks, Decision Tree, and K nearest neighbors (KNN), to both the original high-dimensional dataset and the reduced feature set. By comparing the predictive accuracy, precision, F Score, recall, and execution time of each algorithm, we assess the effectiveness of feature reduction in enhancing the performance of HCC prediction models. Our experimental results, obtained using a comprehensive dataset comprising clinical features of HCC patients, demonstrate that feature reduction significantly improves the performance of all examined algorithms. Notably, the reduced feature set consistently outperforms the original high-dimensional dataset in terms of prediction accuracy and execution time. After applying feature reduction techniques, the employed algorithms, namely decision trees, Naive Bayes, KNN, neural networks, and SVM achieved accuracies of 96%, 97.33%, 94.67%, 96%, and 96.00%, respectively.

肝细胞癌(HCC)是一种高发的肝癌,需要准确的预测模型来进行早期诊断和有效治疗。机器学习算法在包括癌症预测在内的各种医疗领域都取得了可喜的成果。在本研究中,我们通过比较不同机器学习算法在应用特征缩减方法前后的性能,提出了一种用于 HCC 预测的综合方法。我们采用了流行的特征缩减技术,如加权特征、隐藏特征相关性、特征选择和优化选择,以提取能捕捉与 HCC 最相关信息的缩减特征子集。随后,我们对原始高维数据集和缩减后的特征集应用了多种算法,包括奈维贝叶、支持向量机(SVM)、神经网络、决策树和 K 近邻(KNN)。通过比较每种算法的预测准确率、精确度、F Score、召回率和执行时间,我们评估了特征缩减在提高 HCC 预测模型性能方面的有效性。我们使用包含 HCC 患者临床特征的综合数据集获得的实验结果表明,特征缩减显著提高了所有研究算法的性能。值得注意的是,缩减后的特征集在预测准确性和执行时间方面始终优于原始高维数据集。在应用特征缩减技术后,所采用的算法,即决策树、Naive Bayes、KNN、神经网络和 SVM 的准确率分别达到了 96%、97.33%、94.67%、96% 和 96.00%。
{"title":"Feature reduction for hepatocellular carcinoma prediction using machine learning algorithms","authors":"Ghada Mostafa, Hamdi Mahmoud, Tarek Abd El-Hafeez, Mohamed E. ElAraby","doi":"10.1186/s40537-024-00944-3","DOIUrl":"https://doi.org/10.1186/s40537-024-00944-3","url":null,"abstract":"<p>Hepatocellular carcinoma (HCC) is a highly prevalent form of liver cancer that necessitates accurate prediction models for early diagnosis and effective treatment. Machine learning algorithms have demonstrated promising results in various medical domains, including cancer prediction. In this study, we propose a comprehensive approach for HCC prediction by comparing the performance of different machine learning algorithms before and after applying feature reduction methods. We employ popular feature reduction techniques, such as weighting features, hidden features correlation, feature selection, and optimized selection, to extract a reduced feature subset that captures the most relevant information related to HCC. Subsequently, we apply multiple algorithms, including Naive Bayes, support vector machines (SVM), Neural Networks, Decision Tree, and K nearest neighbors (KNN), to both the original high-dimensional dataset and the reduced feature set. By comparing the predictive accuracy, precision, F Score, recall, and execution time of each algorithm, we assess the effectiveness of feature reduction in enhancing the performance of HCC prediction models. Our experimental results, obtained using a comprehensive dataset comprising clinical features of HCC patients, demonstrate that feature reduction significantly improves the performance of all examined algorithms. Notably, the reduced feature set consistently outperforms the original high-dimensional dataset in terms of prediction accuracy and execution time. After applying feature reduction techniques, the employed algorithms, namely decision trees, Naive Bayes, KNN, neural networks, and SVM achieved accuracies of 96%, 97.33%, 94.67%, 96%, and 96.00%, respectively.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"22 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141503495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advanced RIME architecture for global optimization and feature selection 用于全局优化和特征选择的先进 RIME 架构
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-06-18 DOI: 10.1186/s40537-024-00931-8
Ruba Abu Khurma, Malik Braik, Abdullah Alzaqebah, Krishna Gopal Dhal, Robertas Damaševičius, Bilal Abu-Salih

The article introduces an innovative approach to global optimization and feature selection (FS) using the RIME algorithm, inspired by RIME-ice formation. The RIME algorithm employs a soft-RIME search strategy and a hard-RIME puncture mechanism, along with an improved positive greedy selection mechanism, to resist getting trapped in local optima and enhance its overall search capabilities. The article also introduces Binary modified RIME (mRIME), a binary adaptation of the RIME algorithm to address the unique challenges posed by FS problems, which typically involve binary search spaces. Four different types of transfer functions (TFs) were selected for FS issues, and their efficacy was investigated for global optimization using CEC2011 and CEC2017 and FS tasks related to disease diagnosis. The results of the proposed mRIME were tested on ten reliable optimization algorithms. The advanced RIME architecture demonstrated superior performance in global optimization and FS tasks, providing an effective solution to complex optimization problems in various domains.

文章介绍了一种利用 RIME 算法进行全局优化和特征选择(FS)的创新方法,其灵感来自 RIME 冰的形成。RIME 算法采用了软 RIME 搜索策略和硬 RIME 穿刺机制,以及改进的正贪婪选择机制,以防止陷入局部最优并增强其整体搜索能力。文章还介绍了二进制修正 RIME(mRIME),这是对 RIME 算法的二进制调整,以应对 FS 问题带来的独特挑战,这些问题通常涉及二进制搜索空间。针对 FS 问题选择了四种不同类型的传递函数 (TF),并利用 CEC2011 和 CEC2017 以及与疾病诊断相关的 FS 任务研究了它们在全局优化方面的功效。在十种可靠的优化算法上测试了所提出的 mRIME 的结果。先进的 RIME 架构在全局优化和 FS 任务中表现出卓越的性能,为不同领域的复杂优化问题提供了有效的解决方案。
{"title":"Advanced RIME architecture for global optimization and feature selection","authors":"Ruba Abu Khurma, Malik Braik, Abdullah Alzaqebah, Krishna Gopal Dhal, Robertas Damaševičius, Bilal Abu-Salih","doi":"10.1186/s40537-024-00931-8","DOIUrl":"https://doi.org/10.1186/s40537-024-00931-8","url":null,"abstract":"<p>The article introduces an innovative approach to global optimization and feature selection (FS) using the RIME algorithm, inspired by RIME-ice formation. The RIME algorithm employs a soft-RIME search strategy and a hard-RIME puncture mechanism, along with an improved positive greedy selection mechanism, to resist getting trapped in local optima and enhance its overall search capabilities. The article also introduces Binary modified RIME (mRIME), a binary adaptation of the RIME algorithm to address the unique challenges posed by FS problems, which typically involve binary search spaces. Four different types of transfer functions (TFs) were selected for FS issues, and their efficacy was investigated for global optimization using CEC2011 and CEC2017 and FS tasks related to disease diagnosis. The results of the proposed mRIME were tested on ten reliable optimization algorithms. The advanced RIME architecture demonstrated superior performance in global optimization and FS tasks, providing an effective solution to complex optimization problems in various domains.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"22 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141503494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PoLYTC: a novel BERT-based classifier to detect political leaning of YouTube videos based on their titles PoLYTC:基于 BERT 的新型分类器,根据标题检测 YouTube 视频的政治倾向性
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-06-05 DOI: 10.1186/s40537-024-00946-1
Nouar AlDahoul, Talal Rahwan, Yasir Zaki

Over two-thirds of the U.S. population uses YouTube, and a quarter of U.S. adults regularly receive their news from it. Despite the massive political content available on the platform, to date, no classifier has been proposed to classify the political leaning of YouTube videos. The only exception is a classifier that requires extensive information about each video (rather than just the title) and classifies the videos into just three classes (rather than the widely-used categorization into six classes). To fill this gap, “PoLYTC” (Political Leaning YouTube Classifier) is proposed to classify YouTube videos based on their titles into six political classes. PoLYTC utilizes a large language model, namely BERT, and is fine-tuned on a public dataset of 11.5 million YouTube videos. Experiments reveal that the proposed solution achieves high accuracy (75%) and high F1-score (77%), thereby outperforming the state of the art. To further validate the solution’s classification performance, several videos were collected from numerous prominent news agencies’ YouTube channels, such as Fox News and The New York Times, which have widely known political leanings. These videos were classified based on their titles, and the results have shown that, in the vast majority of cases, the predicted political leaning matches that of the news agency. PoLYTC can help YouTube users make informed decisions about which videos to watch and can help researchers analyze the political content on YouTube.

超过三分之二的美国人使用 YouTube,四分之一的美国成年人经常从 YouTube 上获取新闻。尽管该平台上有大量的政治内容,但迄今为止,还没有人提出过分类器来对 YouTube 视频的政治倾向进行分类。唯一的例外是,分类器需要每个视频的大量信息(而不仅仅是标题),并将视频分为三类(而不是广泛使用的六类)。为了填补这一空白,我们提出了 "PoLYTC"(政治倾向 YouTube 分类器),根据标题将 YouTube 视频分为六个政治类别。PoLYTC 采用了一个大型语言模型,即 BERT,并在一个包含 1150 万个 YouTube 视频的公共数据集上进行了微调。实验表明,所提出的解决方案实现了较高的准确率(75%)和较高的 F1 分数(77%),从而超越了现有技术水平。为了进一步验证该解决方案的分类性能,我们从福克斯新闻和《纽约时报》等众多知名新闻机构的 YouTube 频道中收集了一些视频,这些视频具有广为人知的政治倾向。根据标题对这些视频进行了分类,结果表明,在绝大多数情况下,预测的政治倾向与新闻机构的政治倾向相吻合。PoLYTC可以帮助YouTube用户在观看视频时做出明智的决定,也可以帮助研究人员分析YouTube上的政治内容。
{"title":"PoLYTC: a novel BERT-based classifier to detect political leaning of YouTube videos based on their titles","authors":"Nouar AlDahoul, Talal Rahwan, Yasir Zaki","doi":"10.1186/s40537-024-00946-1","DOIUrl":"https://doi.org/10.1186/s40537-024-00946-1","url":null,"abstract":"<p>Over two-thirds of the U.S. population uses YouTube, and a quarter of U.S. adults regularly receive their news from it. Despite the massive political content available on the platform, to date, no classifier has been proposed to classify the political leaning of YouTube videos. The only exception is a classifier that requires extensive information about each video (rather than just the title) and classifies the videos into just three classes (rather than the widely-used categorization into six classes). To fill this gap, “PoLYTC” (Political Leaning YouTube Classifier) is proposed to classify YouTube videos based on their titles into six political classes. PoLYTC utilizes a large language model, namely BERT, and is fine-tuned on a public dataset of 11.5 million YouTube videos. Experiments reveal that the proposed solution achieves high accuracy (75%) and high F1-score (77%), thereby outperforming the state of the art. To further validate the solution’s classification performance, several videos were collected from numerous prominent news agencies’ YouTube channels, such as Fox News and The New York Times, which have widely known political leanings. These videos were classified based on their titles, and the results have shown that, in the vast majority of cases, the predicted political leaning matches that of the news agency. PoLYTC can help YouTube users make informed decisions about which videos to watch and can help researchers analyze the political content on YouTube.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"74 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141520246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GB-AFS: graph-based automatic feature selection for multi-class classification via Mean Simplified Silhouette GB-AFS:基于图的自动特征选择,通过平均简化剪影实现多类分类
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-05-31 DOI: 10.1186/s40537-024-00934-5
David Levin, Gonen Singer

This paper introduces a novel graph-based filter method for automatic feature selection (abbreviated as GB-AFS) for multi-class classification tasks. The method determines the minimum combination of features required to sustain prediction performance while maintaining complementary discriminating abilities between different classes. It does not require any user-defined parameters such as the number of features to select. The minimum number of features is selected using our newly developed Mean Simplified Silhouette (abbreviated as MSS) index, designed to evaluate the clustering results for the feature selection task. To illustrate the effectiveness and generality of the method, we applied the GB-AFS method using various combinations of statistical measures and dimensionality reduction techniques. The experimental results demonstrate the superior performance of the proposed GB-AFS over other filter-based techniques and automatic feature selection approaches, and demonstrate that the GB-AFS method is independent of the statistical measure or the dimensionality reduction technique chosen by the user. Moreover, the proposed method maintained the accuracy achieved when utilizing all features while using only 7–(30%) of the original features. This resulted in an average time saving ranging from (15%) for the smallest dataset to (70%) for the largest. Our code is available at https://github.com/davidlevinwork/gbfs/.

本文介绍了一种新颖的基于图的自动特征选择滤波方法(简称 GB-AFS),用于多类分类任务。该方法可确定维持预测性能所需的最小特征组合,同时保持不同类别之间的互补分辨能力。它不需要任何用户定义的参数,如要选择的特征数量。最小特征数是通过我们新开发的平均简化轮廓(缩写为 MSS)指数来选择的,该指数旨在评估特征选择任务的聚类结果。为了说明该方法的有效性和通用性,我们使用各种统计量和降维技术组合来应用 GB-AFS 方法。实验结果表明,与其他基于滤波器的技术和自动特征选择方法相比,所提出的 GB-AFS 方法性能优越,并证明了 GB-AFS 方法与用户选择的统计量或降维技术无关。此外,所提出的方法保持了利用所有特征时所达到的准确度,同时只使用了 7(30%)个原始特征。这使得最小数据集的平均时间节省了15%,最大数据集的平均时间节省了70%。我们的代码见 https://github.com/davidlevinwork/gbfs/。
{"title":"GB-AFS: graph-based automatic feature selection for multi-class classification via Mean Simplified Silhouette","authors":"David Levin, Gonen Singer","doi":"10.1186/s40537-024-00934-5","DOIUrl":"https://doi.org/10.1186/s40537-024-00934-5","url":null,"abstract":"<p>This paper introduces a novel graph-based filter method for automatic feature selection (abbreviated as GB-AFS) for multi-class classification tasks. The method determines the minimum combination of features required to sustain prediction performance while maintaining complementary discriminating abilities between different classes. It does not require any user-defined parameters such as the number of features to select. The minimum number of features is selected using our newly developed Mean Simplified Silhouette (abbreviated as MSS) index, designed to evaluate the clustering results for the feature selection task. To illustrate the effectiveness and generality of the method, we applied the GB-AFS method using various combinations of statistical measures and dimensionality reduction techniques. The experimental results demonstrate the superior performance of the proposed GB-AFS over other filter-based techniques and automatic feature selection approaches, and demonstrate that the GB-AFS method is independent of the statistical measure or the dimensionality reduction technique chosen by the user. Moreover, the proposed method maintained the accuracy achieved when utilizing all features while using only 7–<span>(30%)</span> of the original features. This resulted in an average time saving ranging from <span>(15%)</span> for the smallest dataset to <span>(70%)</span> for the largest. Our code is available at https://github.com/davidlevinwork/gbfs/.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"117 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141192655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integration of feature enhancement technique in Google inception network for breast cancer detection and classification 在谷歌感知网络中整合特征增强技术,用于乳腺癌检测和分类
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-05-28 DOI: 10.1186/s40537-024-00936-3
Wasyihun Sema Admass, Yirga Yayeh Munaye, Ayodeji Olalekan Salau

Breast cancer is a major public health concern, and early detection and classification are essential for improving patient outcomes. However, breast tumors can be difficult to distinguish from benign tumors, leading to high false positive rates in screening. The reason is that both benign and malignant tumors have no consistent shape, are found at the same position, have variable sizes, and have high correlations. The ambiguity of the correlation challenges the computer-aided system, and the inconsistency of morphology challenges an expert in identifying and classifying what is positive and what is negative. Due to this, most of the time, breast cancer screen is prone to false positive rates. This research paper presents the introduction of a feature enhancement method into the Google inception network for breast cancer detection and classification. The proposed model preserves both local and global information, which is important for addressing the variability of breast tumor morphology and their complex correlations. A locally preserving projection transformation function is introduced to retain local information that might be lost in the intermediate output of the inception model. Additionally, transfer learning is used to improve the performance of the proposed model on limited datasets. The proposed model is evaluated on a dataset of ultrasound images and achieves an accuracy of 99.81%, recall of 96.48%, and sensitivity of 93.0%. These results demonstrate the effectiveness of the proposed method for breast cancer detection and classification.

乳腺癌是一个重大的公共卫生问题,早期检测和分类对改善患者预后至关重要。然而,乳腺肿瘤很难与良性肿瘤区分开来,导致筛查的假阳性率很高。究其原因,良性肿瘤和恶性肿瘤形状不一致、位置相同、大小不一、相关性高。相关性的模糊性给计算机辅助系统带来了挑战,而形态的不一致性又给专家鉴别和分类哪些是阳性哪些是阴性带来了挑战。因此,在大多数情况下,乳腺癌筛查容易出现假阳性率。本研究论文将一种特征增强方法引入谷歌萌芽网络,用于乳腺癌的检测和分类。所提出的模型同时保留了局部和全局信息,这对于解决乳腺肿瘤形态的多变性及其复杂的相关性非常重要。该模型引入了局部保存投影变换函数,以保留初始模型中间输出中可能丢失的局部信息。此外,还使用迁移学习来提高所提模型在有限数据集上的性能。该模型在超声波图像数据集上进行了评估,准确率达到 99.81%,召回率达到 96.48%,灵敏度达到 93.0%。这些结果证明了所提方法在乳腺癌检测和分类方面的有效性。
{"title":"Integration of feature enhancement technique in Google inception network for breast cancer detection and classification","authors":"Wasyihun Sema Admass, Yirga Yayeh Munaye, Ayodeji Olalekan Salau","doi":"10.1186/s40537-024-00936-3","DOIUrl":"https://doi.org/10.1186/s40537-024-00936-3","url":null,"abstract":"<p>Breast cancer is a major public health concern, and early detection and classification are essential for improving patient outcomes. However, breast tumors can be difficult to distinguish from benign tumors, leading to high false positive rates in screening. The reason is that both benign and malignant tumors have no consistent shape, are found at the same position, have variable sizes, and have high correlations. The ambiguity of the correlation challenges the computer-aided system, and the inconsistency of morphology challenges an expert in identifying and classifying what is positive and what is negative. Due to this, most of the time, breast cancer screen is prone to false positive rates. This research paper presents the introduction of a feature enhancement method into the Google inception network for breast cancer detection and classification. The proposed model preserves both local and global information, which is important for addressing the variability of breast tumor morphology and their complex correlations. A locally preserving projection transformation function is introduced to retain local information that might be lost in the intermediate output of the inception model. Additionally, transfer learning is used to improve the performance of the proposed model on limited datasets. The proposed model is evaluated on a dataset of ultrasound images and achieves an accuracy of 99.81%, recall of 96.48%, and sensitivity of 93.0%. These results demonstrate the effectiveness of the proposed method for breast cancer detection and classification.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"29 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141170319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficiently approaching vertical federated learning by combining data reduction and conditional computation techniques 结合数据缩减和条件计算技术,高效实现垂直联合学习
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-05-28 DOI: 10.1186/s40537-024-00933-6
Francesco Folino, Gianluigi Folino, Francesco Sergio Pisani, Luigi Pontieri, Pietro Sabatino

In this paper, a framework based on a sparse Mixture of Experts (MoE) architecture is proposed for the federated learning and application of a distributed classification model in domains (like cybersecurity and healthcare) where different parties of the federation store different subsets of features for a number of data instances. The framework is designed to limit the risk of information leakage and computation/communication costs in both model training (through data sampling) and application (leveraging the conditional-computation abilities of sparse MoEs). Experiments on real data have shown the proposed approach to ensure a better balance between efficiency and model accuracy, compared to other VFL-based solutions. Notably, in a real-life cybersecurity case study focused on malware classification (the KronoDroid dataset), the proposed method surpasses competitors even though it utilizes only 50% and 75% of the training set, which is fully utilized by the other approaches in the competition. This method achieves reductions in the rate of false positives by 16.9% and 18.2%, respectively, and also delivers satisfactory results on the other evaluation metrics. These results showcase our framework’s potential to significantly enhance cybersecurity threat detection and prevention in a collaborative yet secure manner.

本文提出了一个基于稀疏专家混合物(MoE)架构的框架,用于分布式分类模型在不同领域(如网络安全和医疗保健)的联合学习和应用,在这些领域中,联合体的不同成员为大量数据实例存储了不同的特征子集。该框架旨在限制模型训练(通过数据采样)和应用(利用稀疏 MoE 的条件计算能力)中的信息泄露风险和计算/通信成本。对真实数据的实验表明,与其他基于 VFL 的解决方案相比,所提出的方法能确保在效率和模型准确性之间取得更好的平衡。值得注意的是,在以恶意软件分类为重点的真实网络安全案例研究(KronoDroid 数据集)中,所提出的方法超越了竞争对手,尽管它只使用了训练集的 50%和 75%,而其他方法在竞争中已经充分利用了训练集。该方法的误报率分别降低了 16.9% 和 18.2%,在其他评估指标上也取得了令人满意的结果。这些结果表明,我们的框架具有以协作而安全的方式显著提高网络安全威胁检测和预防能力的潜力。
{"title":"Efficiently approaching vertical federated learning by combining data reduction and conditional computation techniques","authors":"Francesco Folino, Gianluigi Folino, Francesco Sergio Pisani, Luigi Pontieri, Pietro Sabatino","doi":"10.1186/s40537-024-00933-6","DOIUrl":"https://doi.org/10.1186/s40537-024-00933-6","url":null,"abstract":"<p>In this paper, a framework based on a sparse Mixture of Experts (MoE) architecture is proposed for the federated learning and application of a distributed classification model in domains (like cybersecurity and healthcare) where different parties of the federation store different subsets of features for a number of data instances. The framework is designed to limit the risk of information leakage and computation/communication costs in both model training (through data sampling) and application (leveraging the conditional-computation abilities of sparse MoEs). Experiments on real data have shown the proposed approach to ensure a better balance between efficiency and model accuracy, compared to other VFL-based solutions. Notably, in a real-life cybersecurity case study focused on malware classification (the KronoDroid dataset), the proposed method surpasses competitors even though it utilizes only 50% and 75% of the training set, which is fully utilized by the other approaches in the competition. This method achieves reductions in the rate of false positives by 16.9% and 18.2%, respectively, and also delivers satisfactory results on the other evaluation metrics. These results showcase our framework’s potential to significantly enhance cybersecurity threat detection and prevention in a collaborative yet secure manner.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"23 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141170198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
15 years of Big Data: a systematic literature review 大数据 15 年:系统文献综述
IF 8.1 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2024-05-14 DOI: 10.1186/s40537-024-00914-9
Davide Tosi, Redon Kokaj, Marco Roccetti

Big Data is still gaining attention as a fundamental building block of the Artificial Intelligence and Machine Learning world. Therefore, a lot of effort has been pushed into Big Data research in the last 15 years. The objective of this Systematic Literature Review is to summarize the current state of the art of the previous 15 years of research about Big Data by providing answers to a set of research questions related to the main application domains for Big Data analytics; the significant challenges and limitations researchers have encountered in Big Data analysis, and emerging research trends and future directions in Big Data. The review follows a predefined procedure that automatically searches five well-known digital libraries. After applying the selection criteria to the results, 189 primary studies were identified as relevant, of which 32 were Systematic Literature Reviews. Required information was extracted from the 32 studies and summarized. Our Systematic Literature Review sketched the picture of 15 years of research in Big Data, identifying application domains, challenges, and future directions in this research field. We believe that a substantial amount of work remains to be done to align and seamlessly integrate Big Data into data-driven advanced software solutions of the future.

大数据作为人工智能和机器学习领域的基本构件,仍然受到越来越多的关注。因此,在过去的 15 年中,人们在大数据研究方面投入了大量精力。本系统性文献综述的目的是总结过去 15 年有关大数据的研究现状,回答一系列研究问题,这些问题涉及大数据分析的主要应用领域、研究人员在大数据分析中遇到的重大挑战和限制,以及大数据的新兴研究趋势和未来方向。综述按照预定程序自动搜索了五个著名的数字图书馆。在对结果应用选择标准后,确定了 189 项相关的主要研究,其中 32 项为系统文献综述。我们从这 32 项研究中提取了所需的信息并进行了总结。我们的系统文献综述勾勒出大数据 15 年来的研究图景,确定了这一研究领域的应用领域、挑战和未来方向。我们认为,要将大数据与数据驱动的未来先进软件解决方案进行协调和无缝集成,仍有大量工作要做。
{"title":"15 years of Big Data: a systematic literature review","authors":"Davide Tosi, Redon Kokaj, Marco Roccetti","doi":"10.1186/s40537-024-00914-9","DOIUrl":"https://doi.org/10.1186/s40537-024-00914-9","url":null,"abstract":"<p>Big Data is still gaining attention as a fundamental building block of the Artificial Intelligence and Machine Learning world. Therefore, a lot of effort has been pushed into Big Data research in the last 15 years. The objective of this Systematic Literature Review is to summarize the current state of the art of the previous 15 years of research about Big Data by providing answers to a set of research questions related to the main application domains for Big Data analytics; the significant challenges and limitations researchers have encountered in Big Data analysis, and emerging research trends and future directions in Big Data. The review follows a predefined procedure that automatically searches five well-known digital libraries. After applying the selection criteria to the results, 189 primary studies were identified as relevant, of which 32 were Systematic Literature Reviews. Required information was extracted from the 32 studies and summarized. Our Systematic Literature Review sketched the picture of 15 years of research in Big Data, identifying application domains, challenges, and future directions in this research field. We believe that a substantial amount of work remains to be done to align and seamlessly integrate Big Data into data-driven advanced software solutions of the future.</p>","PeriodicalId":15158,"journal":{"name":"Journal of Big Data","volume":"100 1","pages":""},"PeriodicalIF":8.1,"publicationDate":"2024-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140931958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Big Data
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1