首页 > 最新文献

Big Data and Cognitive Computing最新文献

英文 中文
Improving Clothing Product Quality and Reducing Waste Based on Consumer Review Using RoBERTa and BERTopic Language Model 基于RoBERTa和BERTopic语言模型的消费者评论提高服装产品质量和减少浪费
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-25 DOI: 10.3390/bdcc7040168
Andry Alamsyah, Nadhif Ditertian Girawan
The disposability of clothing has emerged as a critical concern, precipitating waste accumulation due to product quality degradation. Such consequences exert significant pressure on resources and challenge sustainability efforts. In response, this research focuses on empowering clothing companies to elevate product excellence by harnessing consumer feedback. Beyond insights, this research extends to sustainability by providing suggestions on refining product quality by improving material handling, gradually mitigating waste production, and cultivating longevity, therefore decreasing discarded clothes. Managing a vast influx of diverse reviews necessitates sophisticated natural language processing (NLP) techniques. Our study introduces a Robustly optimized BERT Pretraining Approach (RoBERTa) model calibrated for multilabel classification and BERTopic for topic modeling. The model adeptly distills vital themes from consumer reviews, exhibiting astounding accuracy in projecting concerns across various dimensions of clothing quality. NLP’s potential lies in endowing companies with insights into consumer review, augmented by the BERTopic to facilitate immersive exploration of harvested review topics. This research presents a thorough case for integrating machine learning to foster sustainability and waste reduction. The contribution of this research is notable for its integration of RoBERTa and BERTopic in multilabel classification tasks and topic modeling in the fashion industry. The results indicate that the RoBERTa model exhibits remarkable performance, as demonstrated by its macro-averaged F1 score of 0.87 and micro-averaged F1 score of 0.87. Likewise, BERTopic achieves a coherence score of 0.67, meaning the model can form an insightful topic.
服装的可丢弃性已成为一个关键问题,由于产品质量下降而导致废物堆积。这种后果对资源造成巨大压力,并对可持续性努力提出挑战。作为回应,本研究的重点是授权服装公司通过利用消费者反馈来提升产品的卓越性。除了见解之外,本研究还延伸到可持续性,通过改善物料处理,逐步减少废物产生,培养寿命,从而减少丢弃的衣服,从而提高产品质量。管理大量涌入的不同评论需要复杂的自然语言处理(NLP)技术。我们的研究引入了一个鲁棒优化的BERT预训练方法(RoBERTa)模型,该模型用于多标签分类,BERTopic用于主题建模。该模型熟练地从消费者评论中提炼出重要的主题,在服装质量的各个方面表现出惊人的准确性。NLP的潜力在于赋予公司洞察消费者评论的能力,并通过BERTopic进行增强,以促进对收获的评论主题的沉浸式探索。这项研究为整合机器学习以促进可持续性和减少浪费提供了一个彻底的案例。本研究的贡献在于将RoBERTa和BERTopic集成到时尚行业的多标签分类任务和主题建模中。结果表明,RoBERTa模型具有显著的性能,其宏观平均F1得分为0.87,微观平均F1得分为0.87。同样,BERTopic的相干性得分为0.67,这意味着该模型可以形成一个有洞察力的主题。
{"title":"Improving Clothing Product Quality and Reducing Waste Based on Consumer Review Using RoBERTa and BERTopic Language Model","authors":"Andry Alamsyah, Nadhif Ditertian Girawan","doi":"10.3390/bdcc7040168","DOIUrl":"https://doi.org/10.3390/bdcc7040168","url":null,"abstract":"The disposability of clothing has emerged as a critical concern, precipitating waste accumulation due to product quality degradation. Such consequences exert significant pressure on resources and challenge sustainability efforts. In response, this research focuses on empowering clothing companies to elevate product excellence by harnessing consumer feedback. Beyond insights, this research extends to sustainability by providing suggestions on refining product quality by improving material handling, gradually mitigating waste production, and cultivating longevity, therefore decreasing discarded clothes. Managing a vast influx of diverse reviews necessitates sophisticated natural language processing (NLP) techniques. Our study introduces a Robustly optimized BERT Pretraining Approach (RoBERTa) model calibrated for multilabel classification and BERTopic for topic modeling. The model adeptly distills vital themes from consumer reviews, exhibiting astounding accuracy in projecting concerns across various dimensions of clothing quality. NLP’s potential lies in endowing companies with insights into consumer review, augmented by the BERTopic to facilitate immersive exploration of harvested review topics. This research presents a thorough case for integrating machine learning to foster sustainability and waste reduction. The contribution of this research is notable for its integration of RoBERTa and BERTopic in multilabel classification tasks and topic modeling in the fashion industry. The results indicate that the RoBERTa model exhibits remarkable performance, as demonstrated by its macro-averaged F1 score of 0.87 and micro-averaged F1 score of 0.87. Likewise, BERTopic achieves a coherence score of 0.67, meaning the model can form an insightful topic.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"16 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135113017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identifying Probable Dementia in Undiagnosed Black and White Americans Using Machine Learning in Veterans Health Administration Electronic Health Records 使用退伍军人健康管理局电子健康记录中的机器学习识别未确诊的黑人和白人美国人可能的痴呆症
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-19 DOI: 10.3390/bdcc7040167
Yijun Shao, Kaitlin Todd, Andrew Shutes-David, Steven P. Millard, Karl Brown, Amy Thomas, Kathryn Chen, Katherine Wilson, Qing T. Zeng, Debby W. Tsuang
The application of natural language processing and machine learning (ML) in electronic health records (EHRs) may help reduce dementia underdiagnosis, but models that are not designed to reflect minority populations may instead perpetuate underdiagnosis. To improve the identification of undiagnosed dementia, particularly in Black Americans (BAs), we developed support vector machine (SVM) ML models to assign dementia risk scores based on features identified in unstructured EHR data (via latent Dirichlet allocation and stable topic extraction in n = 1 M notes) and structured EHR data. We hypothesized that separate models would show differentiation between racial groups, so the models were fit separately for BAs (n = 5 K with dementia ICD codes, n = 5 K without) and White Americans (WAs; n = 5 K with codes, n = 5 K without). To validate our method, scores were generated for separate samples of BAs (n = 10 K) and WAs (n = 10 K) without dementia codes, and the EHRs of 1.2 K of these patients were reviewed by dementia experts. All subjects were age 65+ and drawn from the VA, which meant that the samples were disproportionately male. A strong positive relationship was observed between SVM-generated risk scores and undiagnosed dementia. BAs were more likely than WAs to have undiagnosed dementia per chart review, both overall (15.3% vs. 9.5%) and among Veterans with >90th percentile cutoff scores (25.6% vs. 15.3%). With chart reviews as the reference standard and varied cutoff scores, the BA model performed slightly better than the WA model (AUC = 0.86 with negative predictive value [NPV] = 0.98, positive predictive value [PPV] = 0.26, sensitivity = 0.61, specificity = 0.92 and accuracy = 0.91 at >90th percentile cutoff vs. AUC = 0.77 with NPV = 0.98, PPV = 0.15, sensitivity = 0.43, specificity = 0.91 and accuracy = 0.89 at >90th). Our findings suggest that race-specific ML models can help identify BAs who may have undiagnosed dementia. Future studies should examine model generalizability in settings with more females and test whether incorporating these models into clinical settings increases the referral of undiagnosed BAs to specialists.
在电子健康记录(EHRs)中应用自然语言处理和机器学习(ML)可能有助于减少痴呆症的诊断不足,但不是为反映少数群体而设计的模型可能会导致诊断不足。为了提高对未确诊痴呆的识别,特别是在美国黑人(BAs)中,我们开发了支持向量机(SVM) ML模型,根据非结构化EHR数据(通过n = 1 M笔记的潜在狄利克雷分配和稳定主题提取)和结构化EHR数据中识别的特征分配痴呆风险评分。我们假设单独的模型会显示种族群体之间的差异,因此模型分别适合BAs (n = 5k, n = 5k,没有痴呆症ICD代码)和White Americans (WAs;n = 5k带代码,n = 5k不带代码)。为了验证我们的方法,我们对没有痴呆代码的BAs (n = 10 K)和WAs (n = 10 K)的单独样本进行了评分,并由痴呆专家对这些患者的1.2 K的电子病历进行了审查。所有研究对象的年龄都在65岁以上,都来自退伍军人管理局,这意味着样本中男性的比例不成比例。支持向量机生成的风险评分与未诊断的痴呆之间存在强烈的正相关。在每个图表回顾中,ba比WAs更有可能患有未确诊的痴呆症,无论是总体(15.3%对9.5%)还是在第90百分位分值的退伍军人中(25.6%对15.3%)。以图表回顾作为参考标准和不同的截止点评分,BA模型的表现略好于WA模型(在第90百分位截止点,AUC = 0.86,阴性预测值[NPV] = 0.98,阳性预测值[PPV] = 0.26,敏感性= 0.61,特异性= 0.92,准确性= 0.91;在第90百分位截止点,AUC = 0.77, NPV = 0.98, PPV = 0.15,敏感性= 0.43,特异性= 0.91,准确性= 0.89)。我们的研究结果表明,种族特异性ML模型可以帮助识别可能患有未确诊痴呆的BAs。未来的研究应该在更多女性的环境中检验模型的普遍性,并测试将这些模型纳入临床环境是否会增加未确诊的BAs转诊给专家。
{"title":"Identifying Probable Dementia in Undiagnosed Black and White Americans Using Machine Learning in Veterans Health Administration Electronic Health Records","authors":"Yijun Shao, Kaitlin Todd, Andrew Shutes-David, Steven P. Millard, Karl Brown, Amy Thomas, Kathryn Chen, Katherine Wilson, Qing T. Zeng, Debby W. Tsuang","doi":"10.3390/bdcc7040167","DOIUrl":"https://doi.org/10.3390/bdcc7040167","url":null,"abstract":"The application of natural language processing and machine learning (ML) in electronic health records (EHRs) may help reduce dementia underdiagnosis, but models that are not designed to reflect minority populations may instead perpetuate underdiagnosis. To improve the identification of undiagnosed dementia, particularly in Black Americans (BAs), we developed support vector machine (SVM) ML models to assign dementia risk scores based on features identified in unstructured EHR data (via latent Dirichlet allocation and stable topic extraction in n = 1 M notes) and structured EHR data. We hypothesized that separate models would show differentiation between racial groups, so the models were fit separately for BAs (n = 5 K with dementia ICD codes, n = 5 K without) and White Americans (WAs; n = 5 K with codes, n = 5 K without). To validate our method, scores were generated for separate samples of BAs (n = 10 K) and WAs (n = 10 K) without dementia codes, and the EHRs of 1.2 K of these patients were reviewed by dementia experts. All subjects were age 65+ and drawn from the VA, which meant that the samples were disproportionately male. A strong positive relationship was observed between SVM-generated risk scores and undiagnosed dementia. BAs were more likely than WAs to have undiagnosed dementia per chart review, both overall (15.3% vs. 9.5%) and among Veterans with >90th percentile cutoff scores (25.6% vs. 15.3%). With chart reviews as the reference standard and varied cutoff scores, the BA model performed slightly better than the WA model (AUC = 0.86 with negative predictive value [NPV] = 0.98, positive predictive value [PPV] = 0.26, sensitivity = 0.61, specificity = 0.92 and accuracy = 0.91 at >90th percentile cutoff vs. AUC = 0.77 with NPV = 0.98, PPV = 0.15, sensitivity = 0.43, specificity = 0.91 and accuracy = 0.89 at >90th). Our findings suggest that race-specific ML models can help identify BAs who may have undiagnosed dementia. Future studies should examine model generalizability in settings with more females and test whether incorporating these models into clinical settings increases the referral of undiagnosed BAs to specialists.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135779428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HAMCap: A Weak-Supervised Hybrid Attention-Based Capsule Neural Network for Fine-Grained Climate Change Debate Analysis HAMCap:用于细粒度气候变化辩论分析的弱监督混合基于注意力的胶囊神经网络
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-17 DOI: 10.3390/bdcc7040166
Kun Xiang, Akihiro Fujii
Climate change (CC) has become a central global topic within the multiple branches of social disciplines. Natural Language Processing (NLP) plays a superior role since it has achieved marvelous accomplishments in various application scenarios. However, CC debates are ambiguous and complicated to interpret even for humans, especially when it comes to the aspect-oriented fine-grained level. Furthermore, the lack of large-scale effective labeled datasets is always a plight encountered in NLP. In this work, we propose a novel weak-supervised Hybrid Attention Masking Capsule Neural Network (HAMCap) for fine-grained CC debate analysis. Specifically, we use vectors with allocated different weights instead of scalars, and a hybrid attention mechanism is designed in order to better capture and represent information. By randomly masking with a Partial Context Mask (PCM) mechanism, we can better construct the internal relationship between the aspects and entities and easily obtain a large-scale generated dataset. Considering the uniqueness of linguistics, we propose a Reinforcement Learning-based Generator-Selector mechanism to automatically update and select data that are beneficial to model training. Empirical results indicate that our proposed ensemble model outperforms baselines on downstream tasks with a maximum of 50.08% on accuracy and 49.48% on F1 scores. Finally, we draw interpretable conclusions about the climate change debate, which is a widespread global concern.
气候变化(CC)已成为社会学科多个分支中的一个核心全球话题。自然语言处理(Natural Language Processing, NLP)在各种应用场景中都取得了令人瞩目的成就,在其中发挥着重要的作用。然而,CC争论是模棱两可的,即使对于人类来说也很难解释,尤其是在面向方面的细粒度级别。此外,缺乏大规模有效的标记数据集一直是自然语言处理中遇到的困境。在这项工作中,我们提出了一种新的弱监督混合注意掩蔽胶囊神经网络(HAMCap),用于细粒度CC辩论分析。具体来说,我们使用分配不同权重的向量代替标量,并设计了一种混合注意机制,以便更好地捕获和表示信息。利用部分上下文掩码(Partial Context Mask, PCM)机制进行随机掩码,可以更好地构建方面与实体之间的内部关系,方便地获得大规模生成的数据集。考虑到语言学的独特性,我们提出了一种基于强化学习的生成器-选择器机制来自动更新和选择有利于模型训练的数据。实证结果表明,我们提出的集成模型在下游任务上的准确率最高为50.08%,F1分数最高为49.48%。最后,我们对气候变化辩论得出了可解释的结论,这是一个广泛的全球关注。
{"title":"HAMCap: A Weak-Supervised Hybrid Attention-Based Capsule Neural Network for Fine-Grained Climate Change Debate Analysis","authors":"Kun Xiang, Akihiro Fujii","doi":"10.3390/bdcc7040166","DOIUrl":"https://doi.org/10.3390/bdcc7040166","url":null,"abstract":"Climate change (CC) has become a central global topic within the multiple branches of social disciplines. Natural Language Processing (NLP) plays a superior role since it has achieved marvelous accomplishments in various application scenarios. However, CC debates are ambiguous and complicated to interpret even for humans, especially when it comes to the aspect-oriented fine-grained level. Furthermore, the lack of large-scale effective labeled datasets is always a plight encountered in NLP. In this work, we propose a novel weak-supervised Hybrid Attention Masking Capsule Neural Network (HAMCap) for fine-grained CC debate analysis. Specifically, we use vectors with allocated different weights instead of scalars, and a hybrid attention mechanism is designed in order to better capture and represent information. By randomly masking with a Partial Context Mask (PCM) mechanism, we can better construct the internal relationship between the aspects and entities and easily obtain a large-scale generated dataset. Considering the uniqueness of linguistics, we propose a Reinforcement Learning-based Generator-Selector mechanism to automatically update and select data that are beneficial to model training. Empirical results indicate that our proposed ensemble model outperforms baselines on downstream tasks with a maximum of 50.08% on accuracy and 49.48% on F1 scores. Finally, we draw interpretable conclusions about the climate change debate, which is a widespread global concern.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135993146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ZeroTrustBlock: Enhancing Security, Privacy, and Interoperability of Sensitive Data through ZeroTrust Permissioned Blockchain ZeroTrustBlock:通过ZeroTrust许可区块链增强敏感数据的安全性、隐私性和互操作性
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-17 DOI: 10.3390/bdcc7040165
Pratik Thantharate, Anurag Thantharate
With the digitization of healthcare, an immense amount of sensitive medical data are generated and shared between various healthcare stakeholders—however, traditional health data management mechanisms present interoperability, security, and privacy challenges. The centralized nature of current health information systems leads to single points of failure, making the data vulnerable to cyberattacks. Patients also have little control over their medical records, raising privacy concerns. Blockchain technology presents a promising solution to these challenges through its decentralized, transparent, and immutable properties. This research proposes ZeroTrustBlock, a comprehensive blockchain framework for secure and private health information exchange. The decentralized ledger enhances integrity, while permissioned access and smart contracts enable patient-centric control over medical data sharing. A hybrid on-chain and off-chain storage model balances transparency with confidentiality. Integration gateways bridge ZeroTrustBlock protocols with existing systems like EHRs. Implemented on Hyperledger Fabric, ZeroTrustBlock demonstrates substantial security improvements over mainstream databases via cryptographic mechanisms, formal privacy-preserving protocols, and access policies enacting patient consent. Results validate the architecture’s effectiveness in achieving 14,200 TPS average throughput, 480 ms average latency for 100,000 concurrent transactions, and linear scalability up to 20 nodes. However, enhancements around performance, advanced cryptography, and real-world pilots are future work. Overall, ZeroTrustBlock provides a robust application of blockchain capabilities to transform security, privacy, interoperability, and patient agency in health data management.
随着医疗保健的数字化,大量敏感医疗数据被生成,并在各种医疗保健利益相关者之间共享——然而,传统的健康数据管理机制存在互操作性、安全性和隐私方面的挑战。当前卫生信息系统的集中性质导致单点故障,使数据容易受到网络攻击。患者对自己的医疗记录也几乎没有控制权,这引发了对隐私的担忧。区块链技术通过其分散、透明和不可变的特性,为这些挑战提供了一个有希望的解决方案。本研究提出了ZeroTrustBlock,这是一个用于安全和私人健康信息交换的综合区块链框架。去中心化的分类账增强了完整性,而允许访问和智能合约可以实现以患者为中心的医疗数据共享控制。链上和链下混合存储模型平衡了透明性和保密性。集成网关将ZeroTrustBlock协议与现有系统(如ehr)连接起来。ZeroTrustBlock在Hyperledger Fabric上实现,通过加密机制、正式的隐私保护协议和制定患者同意的访问策略,在主流数据库的基础上展示了实质性的安全性改进。结果验证了该架构在实现14,200 TPS平均吞吐量、100,000个并发事务的480 ms平均延迟以及最多20个节点的线性可伸缩性方面的有效性。然而,围绕性能、高级加密和现实世界试点的增强是未来的工作。总体而言,ZeroTrustBlock提供了区块链功能的强大应用,以改变健康数据管理中的安全性、隐私性、互操作性和患者代理。
{"title":"ZeroTrustBlock: Enhancing Security, Privacy, and Interoperability of Sensitive Data through ZeroTrust Permissioned Blockchain","authors":"Pratik Thantharate, Anurag Thantharate","doi":"10.3390/bdcc7040165","DOIUrl":"https://doi.org/10.3390/bdcc7040165","url":null,"abstract":"With the digitization of healthcare, an immense amount of sensitive medical data are generated and shared between various healthcare stakeholders—however, traditional health data management mechanisms present interoperability, security, and privacy challenges. The centralized nature of current health information systems leads to single points of failure, making the data vulnerable to cyberattacks. Patients also have little control over their medical records, raising privacy concerns. Blockchain technology presents a promising solution to these challenges through its decentralized, transparent, and immutable properties. This research proposes ZeroTrustBlock, a comprehensive blockchain framework for secure and private health information exchange. The decentralized ledger enhances integrity, while permissioned access and smart contracts enable patient-centric control over medical data sharing. A hybrid on-chain and off-chain storage model balances transparency with confidentiality. Integration gateways bridge ZeroTrustBlock protocols with existing systems like EHRs. Implemented on Hyperledger Fabric, ZeroTrustBlock demonstrates substantial security improvements over mainstream databases via cryptographic mechanisms, formal privacy-preserving protocols, and access policies enacting patient consent. Results validate the architecture’s effectiveness in achieving 14,200 TPS average throughput, 480 ms average latency for 100,000 concurrent transactions, and linear scalability up to 20 nodes. However, enhancements around performance, advanced cryptography, and real-world pilots are future work. Overall, ZeroTrustBlock provides a robust application of blockchain capabilities to transform security, privacy, interoperability, and patient agency in health data management.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136033313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Cognitive Assessment Based on Electroencephalography Analysis in Virtual and Augmented Reality Environments, Using Head Mounted Displays: A Systematic Review 基于脑电图分析的认知评估在虚拟和增强现实环境中,使用头戴式显示器:系统综述
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-13 DOI: 10.3390/bdcc7040163
Foteini Gramouseni, Katerina D. Tzimourta, Pantelis Angelidis, Nikolaos Giannakeas, Markos G. Tsipouras
The objective of this systematic review centers on cognitive assessment based on electroencephalography (EEG) analysis in Virtual Reality (VR), Augmented Reality (AR) and Mixed Reality (MR) environments, projected on Head Mounted Displays (HMD), in healthy individuals. A range of electronic databases were searched (Scopus, ScienceDirect, IEEE Explore and PubMed), using PRISMA research method and 82 experimental studies were included in the final report. Specific aspects of cognitive function were evaluated, including cognitive load, immersion, spatial awareness, interaction with the digital environment and attention. These were analyzed based on various aspects of the analysis, including the number of participants, stimuli, frequency bands range, data preprocessing and data analysis. Based on the analysis conducted, significant findings have emerged both in terms of the experimental structure related to cognitive neuroscience and the key parameters considered in the research. Also, numerous significant avenues and domains requiring more extensive exploration have been identified within neuroscience and cognition research in digital environments. These encompass factors such as the experimental setup, including issues like narrow participant populations and the feasibility of using EEG equipment with a limited number of sensors to overcome the challenges posed by the time-consuming placement of a multi-electrode EEG cap. There is a clear need for more in-depth exploration in signal analysis, especially concerning the α, β, and γ sub-bands and their role in providing more precise insights for evaluating cognitive states. Finally, further research into augmented and mixed reality environments will enable the extraction of more accurate conclusions regarding their utility in cognitive neuroscience.
本系统综述的目的是在虚拟现实(VR)、增强现实(AR)和混合现实(MR)环境中基于脑电图(EEG)分析的认知评估,投影在头戴式显示器(HMD)上,健康个体。检索电子数据库(Scopus、ScienceDirect、IEEE Explore和PubMed),采用PRISMA研究方法,最终报告纳入82项实验研究。评估了认知功能的具体方面,包括认知负荷、沉浸感、空间意识、与数字环境的互动和注意力。这些分析是基于分析的各个方面,包括参与者的数量,刺激,频带范围,数据预处理和数据分析。根据所进行的分析,在与认知神经科学相关的实验结构和研究中考虑的关键参数方面都出现了重要的发现。此外,在数字环境中的神经科学和认知研究中,已经确定了许多需要更广泛探索的重要途径和领域。这些因素包括实验设置,包括参与者群体狭窄以及使用具有有限数量传感器的脑电图设备的可行性等问题,以克服耗时放置多电极脑电图帽所带来的挑战。显然需要在信号分析方面进行更深入的探索,特别是关于α, β和γ子带及其在提供更精确的认知状态评估中的作用。最后,对增强现实和混合现实环境的进一步研究将有助于提取有关其在认知神经科学中的效用的更准确的结论。
{"title":"Cognitive Assessment Based on Electroencephalography Analysis in Virtual and Augmented Reality Environments, Using Head Mounted Displays: A Systematic Review","authors":"Foteini Gramouseni, Katerina D. Tzimourta, Pantelis Angelidis, Nikolaos Giannakeas, Markos G. Tsipouras","doi":"10.3390/bdcc7040163","DOIUrl":"https://doi.org/10.3390/bdcc7040163","url":null,"abstract":"The objective of this systematic review centers on cognitive assessment based on electroencephalography (EEG) analysis in Virtual Reality (VR), Augmented Reality (AR) and Mixed Reality (MR) environments, projected on Head Mounted Displays (HMD), in healthy individuals. A range of electronic databases were searched (Scopus, ScienceDirect, IEEE Explore and PubMed), using PRISMA research method and 82 experimental studies were included in the final report. Specific aspects of cognitive function were evaluated, including cognitive load, immersion, spatial awareness, interaction with the digital environment and attention. These were analyzed based on various aspects of the analysis, including the number of participants, stimuli, frequency bands range, data preprocessing and data analysis. Based on the analysis conducted, significant findings have emerged both in terms of the experimental structure related to cognitive neuroscience and the key parameters considered in the research. Also, numerous significant avenues and domains requiring more extensive exploration have been identified within neuroscience and cognition research in digital environments. These encompass factors such as the experimental setup, including issues like narrow participant populations and the feasibility of using EEG equipment with a limited number of sensors to overcome the challenges posed by the time-consuming placement of a multi-electrode EEG cap. There is a clear need for more in-depth exploration in signal analysis, especially concerning the α, β, and γ sub-bands and their role in providing more precise insights for evaluating cognitive states. Finally, further research into augmented and mixed reality environments will enable the extraction of more accurate conclusions regarding their utility in cognitive neuroscience.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135855609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MM-EMOR: Multi-Modal Emotion Recognition of Social Media Using Concatenated Deep Learning Networks MM-EMOR:使用连接深度学习网络的社交媒体多模态情感识别
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-13 DOI: 10.3390/bdcc7040164
Omar Adel, Karma M.Fathalla, Ahmed Abo ElFarag
Emotion recognition is crucial in artificial intelligence, particularly in the domain of human–computer interaction. The ability to accurately discern and interpret emotions plays a critical role in helping machines to effectively decipher users’ underlying intentions, allowing for a more streamlined interaction process that invariably translates into an elevated user experience. The recent increase in social media usage, as well as the availability of an immense amount of unstructured data, has resulted in a significant demand for the deployment of automated emotion recognition systems. Artificial intelligence (AI) techniques have emerged as a powerful solution to this pressing concern in this context. In particular, the incorporation of multimodal AI-driven approaches for emotion recognition has proven beneficial in capturing the intricate interplay of diverse human expression cues that manifest across multiple modalities. The current study aims to develop an effective multimodal emotion recognition system known as MM-EMOR in order to improve the efficacy of emotion recognition efforts focused on audio and text modalities. The use of Mel spectrogram features, Chromagram features, and the Mobilenet Convolutional Neural Network (CNN) for processing audio data are central to the operation of this system, while an attention-based Roberta model caters to the text modality. The methodology of this study is based on an exhaustive evaluation of this approach across three different datasets. Notably, the empirical findings show that MM-EMOR outperforms competing models across the same datasets. This performance boost is noticeable, with accuracy gains of an impressive 7% on one dataset and a substantial 8% on another. Most significantly, the observed increase in accuracy for the final dataset was an astounding 18%.
情感识别在人工智能领域,尤其是人机交互领域具有重要意义。准确识别和解释情绪的能力在帮助机器有效地破译用户潜在意图方面起着至关重要的作用,从而允许更精简的交互过程,从而始终转化为更高的用户体验。最近社交媒体使用量的增加,以及大量非结构化数据的可用性,导致了对自动情绪识别系统部署的巨大需求。在这种背景下,人工智能(AI)技术已经成为解决这一紧迫问题的有力解决方案。特别是,将多模态人工智能驱动的方法用于情感识别,已被证明有助于捕捉跨越多种模态的各种人类表达线索的复杂相互作用。本研究旨在开发一种有效的多模态情绪识别系统MM-EMOR,以提高专注于音频和文本模态的情绪识别的效率。使用Mel谱图特征、Chromagram特征和Mobilenet卷积神经网络(CNN)来处理音频数据是该系统操作的核心,而基于注意力的Roberta模型则迎合了文本模式。本研究的方法是基于对三种不同数据集的方法进行详尽的评估。值得注意的是,实证结果表明,在相同的数据集上,MM-EMOR优于竞争模型。这种性能提升是明显的,在一个数据集上获得了令人印象深刻的7%的准确性,在另一个数据集上获得了可观的8%。最重要的是,观察到最终数据集的准确性提高了惊人的18%。
{"title":"MM-EMOR: Multi-Modal Emotion Recognition of Social Media Using Concatenated Deep Learning Networks","authors":"Omar Adel, Karma M.Fathalla, Ahmed Abo ElFarag","doi":"10.3390/bdcc7040164","DOIUrl":"https://doi.org/10.3390/bdcc7040164","url":null,"abstract":"Emotion recognition is crucial in artificial intelligence, particularly in the domain of human–computer interaction. The ability to accurately discern and interpret emotions plays a critical role in helping machines to effectively decipher users’ underlying intentions, allowing for a more streamlined interaction process that invariably translates into an elevated user experience. The recent increase in social media usage, as well as the availability of an immense amount of unstructured data, has resulted in a significant demand for the deployment of automated emotion recognition systems. Artificial intelligence (AI) techniques have emerged as a powerful solution to this pressing concern in this context. In particular, the incorporation of multimodal AI-driven approaches for emotion recognition has proven beneficial in capturing the intricate interplay of diverse human expression cues that manifest across multiple modalities. The current study aims to develop an effective multimodal emotion recognition system known as MM-EMOR in order to improve the efficacy of emotion recognition efforts focused on audio and text modalities. The use of Mel spectrogram features, Chromagram features, and the Mobilenet Convolutional Neural Network (CNN) for processing audio data are central to the operation of this system, while an attention-based Roberta model caters to the text modality. The methodology of this study is based on an exhaustive evaluation of this approach across three different datasets. Notably, the empirical findings show that MM-EMOR outperforms competing models across the same datasets. This performance boost is noticeable, with accuracy gains of an impressive 7% on one dataset and a substantial 8% on another. Most significantly, the observed increase in accuracy for the final dataset was an astounding 18%.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135856260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contemporary Art Authentication with Large-Scale Classification 大规模分类的当代艺术鉴定
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-09 DOI: 10.3390/bdcc7040162
Todd Dobbs, Abdullah-Al-Raihan Nayeem, Isaac Cho, Zbigniew Ras
Art authentication is the process of identifying the artist who created a piece of artwork and is manifested through events of provenance, such as art gallery exhibitions and financial transactions. Art authentication has visual influence via the uniqueness of the artist’s style in contrast to the style of another artist. The significance of this contrast is proportional to the number of artists involved and the degree of uniqueness of an artist’s collection. This visual uniqueness of style can be captured in a mathematical model produced by a machine learning (ML) algorithm on painting images. Art authentication is not always possible as provenance can be obscured or lost through anonymity, forgery, gifting, or theft of artwork. This paper presents an image-only art authentication attribute marker of contemporary art paintings for a very large number of artists. The experiments in this paper demonstrate that it is possible to use ML-generated models to authenticate contemporary art from 2368 to 100 artists with an accuracy of 48.97% to 91.23%, respectively. This is the largest effort for image-only art authentication to date, with respect to the number of artists involved and the accuracy of authentication.
艺术鉴定是鉴定创作一件艺术品的艺术家的过程,通过美术馆展览和金融交易等事件的来源来体现。艺术鉴定通过艺术家风格的独特性与其他艺术家风格的对比而产生视觉影响。这种对比的重要性与参与的艺术家数量和艺术家收藏的独特性程度成正比。这种风格的视觉独特性可以通过机器学习(ML)算法对绘画图像产生的数学模型来捕捉。艺术认证并不总是可能的,因为出处可能会因匿名、伪造、赠送或盗窃而模糊或丢失。本文为非常多的艺术家提供了一种当代艺术绘画的纯图像艺术认证属性标记。本文的实验表明,使用ml生成的模型对2368 ~ 100位艺术家的当代艺术作品进行认证是可能的,准确率分别为48.97% ~ 91.23%。就涉及的艺术家数量和认证的准确性而言,这是迄今为止在纯图像艺术认证方面所做的最大努力。
{"title":"Contemporary Art Authentication with Large-Scale Classification","authors":"Todd Dobbs, Abdullah-Al-Raihan Nayeem, Isaac Cho, Zbigniew Ras","doi":"10.3390/bdcc7040162","DOIUrl":"https://doi.org/10.3390/bdcc7040162","url":null,"abstract":"Art authentication is the process of identifying the artist who created a piece of artwork and is manifested through events of provenance, such as art gallery exhibitions and financial transactions. Art authentication has visual influence via the uniqueness of the artist’s style in contrast to the style of another artist. The significance of this contrast is proportional to the number of artists involved and the degree of uniqueness of an artist’s collection. This visual uniqueness of style can be captured in a mathematical model produced by a machine learning (ML) algorithm on painting images. Art authentication is not always possible as provenance can be obscured or lost through anonymity, forgery, gifting, or theft of artwork. This paper presents an image-only art authentication attribute marker of contemporary art paintings for a very large number of artists. The experiments in this paper demonstrate that it is possible to use ML-generated models to authenticate contemporary art from 2368 to 100 artists with an accuracy of 48.97% to 91.23%, respectively. This is the largest effort for image-only art authentication to date, with respect to the number of artists involved and the accuracy of authentication.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135095107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Empirical Study on Core Data Asset Identification in Data Governance 数据治理中核心数据资产识别的实证研究
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-10-07 DOI: 10.3390/bdcc7040161
Yunpeng Chen, Ying Zhao, Wenxuan Xie, Yanbo Zhai, Xin Zhao, Jiang Zhang, Jiang Long, Fangfang Zhou
Data governance aims to optimize the value derived from data assets and effectively mitigate data-related risks. The rapid growth of data assets increases the risk of data breaches. One key solution to reduce this risk is to classify data assets according to their business value and criticality to the enterprises, allocating limited resources to protect core data assets. The existing methods rely on the experience of professionals and cannot identify core data assets across business scenarios. This work conducts an empirical study to address this issue. First, we utilized data lineage graphs with expert-labeled core data assets to investigate the experience of data users on core data asset identification from a scenario perspective. Then, we explored the structural features of core data assets on data lineage graphs from an abstraction perspective. Finally, one expert seminar was conducted to derive a set of universal indicators to identify core data assets by synthesizing the results from the two perspectives. User and field studies were conducted to demonstrate the effectiveness of the indicators.
数据治理旨在优化数据资产的价值,并有效降低数据相关风险。数据资产的快速增长增加了数据泄露的风险。降低这种风险的一个关键解决方案是根据数据资产的业务价值和对企业的重要性对其进行分类,分配有限的资源来保护核心数据资产。现有的方法依赖于专业人员的经验,不能跨业务场景识别核心数据资产。本文对这一问题进行了实证研究。首先,我们利用带有专家标记的核心数据资产的数据谱系图,从场景的角度调查数据用户在核心数据资产识别方面的体验。然后,我们从抽象的角度探讨了数据沿袭图上核心数据资产的结构特征。最后,举办了一次专家研讨会,通过综合两个方面的结果,得出一套通用指标,以确定核心数据资产。进行了用户和实地研究,以证明这些指标的有效性。
{"title":"An Empirical Study on Core Data Asset Identification in Data Governance","authors":"Yunpeng Chen, Ying Zhao, Wenxuan Xie, Yanbo Zhai, Xin Zhao, Jiang Zhang, Jiang Long, Fangfang Zhou","doi":"10.3390/bdcc7040161","DOIUrl":"https://doi.org/10.3390/bdcc7040161","url":null,"abstract":"Data governance aims to optimize the value derived from data assets and effectively mitigate data-related risks. The rapid growth of data assets increases the risk of data breaches. One key solution to reduce this risk is to classify data assets according to their business value and criticality to the enterprises, allocating limited resources to protect core data assets. The existing methods rely on the experience of professionals and cannot identify core data assets across business scenarios. This work conducts an empirical study to address this issue. First, we utilized data lineage graphs with expert-labeled core data assets to investigate the experience of data users on core data asset identification from a scenario perspective. Then, we explored the structural features of core data assets on data lineage graphs from an abstraction perspective. Finally, one expert seminar was conducted to derive a set of universal indicators to identify core data assets by synthesizing the results from the two perspectives. User and field studies were conducted to demonstrate the effectiveness of the indicators.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135251931","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Defining Semantically Close Words of Kazakh Language with Distributed System Apache Spark 用分布式系统Apache Spark定义哈萨克语语义封闭词
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-09-27 DOI: 10.3390/bdcc7040160
Dauren Ayazbayev, Andrey Bogdanchikov, Kamila Orynbekova, Iraklis Varlamis
This work focuses on determining semantically close words and using semantic similarity in general in order to improve performance in information retrieval tasks. The semantic similarity of words is an important task with many applications from information retrieval to spell checking or even document clustering and classification. Although, in languages with rich linguistic resources, the methods and tools for this task are well established, some languages do not have such tools. The first step in our experiment is to represent the words in a collection in a vector form and then define the semantic similarity of the terms using a vector similarity method. In order to tame the complexity of the task, which relies on the number of word (and, consequently, of the vector) pairs that have to be combined in order to define the semantically closest word pairs, A distributed method that runs on Apache Spark is designed to reduce the calculation time by running comparison tasks in parallel. Three alternative implementations are proposed and tested using a list of target words and seeking the most semantically similar words from a lexicon for each one of them. In a second step, we employ pre-trained multilingual sentence transformers to capture the content semantics at a sentence level and a vector-based semantic index to accelerate the searches. The code is written in MapReduce, and the experiments and results show that the proposed methods can provide an interesting solution for finding similar words or texts in the Kazakh language.
本研究的重点是确定语义相近的词,并利用语义相似度来提高信息检索任务的性能。从信息检索到拼写检查,甚至是文档聚类和分类,单词的语义相似度是一项重要的任务。虽然在语言资源丰富的语言中,完成这项任务的方法和工具已经很好地建立起来,但有些语言没有这样的工具。我们实验的第一步是以向量形式表示集合中的单词,然后使用向量相似度方法定义术语的语义相似度。为了驯服任务的复杂性,它依赖于为了定义语义上最接近的词对而必须组合的词对的数量(以及向量的数量),在Apache Spark上运行的分布式方法被设计为通过并行运行比较任务来减少计算时间。提出并测试了三种备选实现,使用目标单词列表,并从词典中为每个单词寻找语义上最相似的单词。在第二步中,我们使用预先训练的多语言句子转换器在句子级别捕获内容语义,并使用基于向量的语义索引来加速搜索。代码是用MapReduce编写的,实验和结果表明,所提出的方法可以为寻找哈萨克语中相似的单词或文本提供一个有趣的解决方案。
{"title":"Defining Semantically Close Words of Kazakh Language with Distributed System Apache Spark","authors":"Dauren Ayazbayev, Andrey Bogdanchikov, Kamila Orynbekova, Iraklis Varlamis","doi":"10.3390/bdcc7040160","DOIUrl":"https://doi.org/10.3390/bdcc7040160","url":null,"abstract":"This work focuses on determining semantically close words and using semantic similarity in general in order to improve performance in information retrieval tasks. The semantic similarity of words is an important task with many applications from information retrieval to spell checking or even document clustering and classification. Although, in languages with rich linguistic resources, the methods and tools for this task are well established, some languages do not have such tools. The first step in our experiment is to represent the words in a collection in a vector form and then define the semantic similarity of the terms using a vector similarity method. In order to tame the complexity of the task, which relies on the number of word (and, consequently, of the vector) pairs that have to be combined in order to define the semantically closest word pairs, A distributed method that runs on Apache Spark is designed to reduce the calculation time by running comparison tasks in parallel. Three alternative implementations are proposed and tested using a list of target words and seeking the most semantically similar words from a lexicon for each one of them. In a second step, we employ pre-trained multilingual sentence transformers to capture the content semantics at a sentence level and a vector-based semantic index to accelerate the searches. The code is written in MapReduce, and the experiments and results show that the proposed methods can provide an interesting solution for finding similar words or texts in the Kazakh language.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135580073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Pruning Method Based on Feature Map Similarity Score 一种基于特征映射相似度评分的剪枝方法
Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-09-26 DOI: 10.3390/bdcc7040159
Jihua Cui, Zhenbang Wang, Ziheng Yang, Xin Guan
As the number of layers of deep learning models increases, the number of parameters and computation increases, making it difficult to deploy on edge devices. Pruning has the potential to significantly reduce the number of parameters and computations in a deep learning model. Existing pruning methods frequently require a specific distribution of network parameters to achieve good results when measuring filter importance. As a result, a feature map similarity score-based pruning method is proposed. We calculate the similarity score of each feature map to measure the importance of the filter and guide filter pruning using the similarity between the filter output feature maps to measure the redundancy of the corresponding filter. Pruning experiments on ResNet-56 and ResNet-110 networks on Cifar-10 datasets can compress the model by more than 70% while maintaining a higher compression ratio and accuracy than traditional methods.
随着深度学习模型层数的增加,参数和计算量的增加,使其难以部署在边缘设备上。修剪有可能显著减少深度学习模型中的参数数量和计算量。现有的剪枝方法往往需要特定的网络参数分布,才能在测量滤波器重要性时获得良好的结果。为此,提出了一种基于特征映射相似度分数的剪枝方法。我们计算每个特征映射的相似度分数来衡量滤波器的重要性,并使用滤波器输出特征映射之间的相似度来衡量相应滤波器的冗余度来指导滤波器修剪。在Cifar-10数据集上进行ResNet-56和ResNet-110网络的剪枝实验,可以将模型压缩70%以上,同时保持比传统方法更高的压缩比和精度。
{"title":"A Pruning Method Based on Feature Map Similarity Score","authors":"Jihua Cui, Zhenbang Wang, Ziheng Yang, Xin Guan","doi":"10.3390/bdcc7040159","DOIUrl":"https://doi.org/10.3390/bdcc7040159","url":null,"abstract":"As the number of layers of deep learning models increases, the number of parameters and computation increases, making it difficult to deploy on edge devices. Pruning has the potential to significantly reduce the number of parameters and computations in a deep learning model. Existing pruning methods frequently require a specific distribution of network parameters to achieve good results when measuring filter importance. As a result, a feature map similarity score-based pruning method is proposed. We calculate the similarity score of each feature map to measure the importance of the filter and guide filter pruning using the similarity between the filter output feature maps to measure the redundancy of the corresponding filter. Pruning experiments on ResNet-56 and ResNet-110 networks on Cifar-10 datasets can compress the model by more than 70% while maintaining a higher compression ratio and accuracy than traditional methods.","PeriodicalId":36397,"journal":{"name":"Big Data and Cognitive Computing","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134886460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Big Data and Cognitive Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1