首页 > 最新文献

Frontiers in Computer Science最新文献

英文 中文
Exploring Deep Transfer Learning Techniques for Alzheimer's Dementia Detection. 探索用于阿尔茨海默氏症痴呆症检测的深度迁移学习技术。
IF 2.6 Q3 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2021-05-01 Epub Date: 2021-05-12 DOI: 10.3389/fcomp.2021.624683
Youxiang Zhu, Xiaohui Liang, John A Batsis, Robert M Roth

Examination of speech datasets for detecting dementia, collected via various speech tasks, has revealed links between speech and cognitive abilities. However, the speech dataset available for this research is extremely limited because the collection process of speech and baseline data from patients with dementia in clinical settings is expensive. In this paper, we study the spontaneous speech dataset from a recent ADReSS challenge, a Cookie Theft Picture (CTP) dataset with balanced groups of participants in age, gender, and cognitive status. We explore state-of-the-art deep transfer learning techniques from image, audio, speech, and language domains. We envision that one advantage of transfer learning is to eliminate the design of handcrafted features based on the tasks and datasets. Transfer learning further mitigates the limited dementia-relevant speech data problem by inheriting knowledge from similar but much larger datasets. Specifically, we built a variety of transfer learning models using commonly employed MobileNet (image), YAMNet (audio), Mockingjay (speech), and BERT (text) models. Results indicated that the transfer learning models of text data showed significantly better performance than those of audio data. Performance gains of the text models may be due to the high similarity between the pre-training text dataset and the CTP text dataset. Our multi-modal transfer learning introduced a slight improvement in accuracy, demonstrating that audio and text data provide limited complementary information. Multi-task transfer learning resulted in limited improvements in classification and a negative impact in regression. By analyzing the meaning behind the AD/non-AD labels and Mini-Mental State Examination (MMSE) scores, we observed that the inconsistency between labels and scores could limit the performance of the multi-task learning, especially when the outputs of the single-task models are highly consistent with the corresponding labels/scores. In sum, we conducted a large comparative analysis of varying transfer learning models focusing less on model customization but more on pre-trained models and pre-training datasets. We revealed insightful relations among models, data types, and data labels in this research area.

通过各种语音任务收集到的用于检测痴呆症的语音数据集显示,语音与认知能力之间存在联系。然而,由于在临床环境中收集痴呆症患者的语音和基线数据的过程非常昂贵,因此可用于这项研究的语音数据集非常有限。在本文中,我们研究了最近一次 ADReSS 挑战赛中的自发语音数据集,即 Cookie Theft Picture(CTP)数据集,该数据集的参与者在年龄、性别和认知状态上都是均衡的。我们探索了图像、音频、语音和语言领域最先进的深度迁移学习技术。我们认为,迁移学习的一个优势是消除了基于任务和数据集的手工特征设计。迁移学习通过继承类似但规模更大的数据集的知识,进一步缓解了痴呆症相关语音数据有限的问题。具体来说,我们使用常用的 MobileNet(图像)、YAMNet(音频)、Mockingjay(语音)和 BERT(文本)模型建立了各种迁移学习模型。结果表明,文本数据迁移学习模型的性能明显优于音频数据迁移学习模型。文本模型的性能提升可能是由于预训练文本数据集与 CTP 文本数据集之间的高度相似性。我们的多模态迁移学习略微提高了准确率,这表明音频和文本数据提供的互补信息有限。多任务迁移学习在分类方面的改进有限,而在回归方面则产生了负面影响。通过分析注意力缺失/非注意力缺失(AD/non-AD)标签和迷你精神状态检查(MMSE)分数背后的含义,我们发现标签和分数之间的不一致性可能会限制多任务学习的性能,尤其是当单任务模型的输出与相应的标签/分数高度一致时。总之,我们对不同的迁移学习模型进行了大量比较分析,重点不是模型定制,而是预训练模型和预训练数据集。我们揭示了这一研究领域中模型、数据类型和数据标签之间的深刻关系。
{"title":"Exploring Deep Transfer Learning Techniques for Alzheimer's Dementia Detection.","authors":"Youxiang Zhu, Xiaohui Liang, John A Batsis, Robert M Roth","doi":"10.3389/fcomp.2021.624683","DOIUrl":"10.3389/fcomp.2021.624683","url":null,"abstract":"<p><p>Examination of speech datasets for detecting dementia, collected via various speech tasks, has revealed links between speech and cognitive abilities. However, the speech dataset available for this research is extremely limited because the collection process of speech and baseline data from patients with dementia in clinical settings is expensive. In this paper, we study the spontaneous speech dataset from a recent ADReSS challenge, a Cookie Theft Picture (CTP) dataset with balanced groups of participants in age, gender, and cognitive status. We explore state-of-the-art deep transfer learning techniques from image, audio, speech, and language domains. We envision that one advantage of transfer learning is to eliminate the design of handcrafted features based on the tasks and datasets. Transfer learning further mitigates the limited dementia-relevant speech data problem by inheriting knowledge from similar but much larger datasets. Specifically, we built a variety of transfer learning models using commonly employed MobileNet (image), YAMNet (audio), Mockingjay (speech), and BERT (text) models. Results indicated that the transfer learning models of text data showed significantly better performance than those of audio data. Performance gains of the text models may be due to the high similarity between the pre-training text dataset and the CTP text dataset. Our multi-modal transfer learning introduced a slight improvement in accuracy, demonstrating that audio and text data provide limited complementary information. Multi-task transfer learning resulted in limited improvements in classification and a negative impact in regression. By analyzing the meaning behind the AD/non-AD labels and Mini-Mental State Examination (MMSE) scores, we observed that the inconsistency between labels and scores could limit the performance of the multi-task learning, especially when the outputs of the single-task models are highly consistent with the corresponding labels/scores. In sum, we conducted a large comparative analysis of varying transfer learning models focusing less on model customization but more on pre-trained models and pre-training datasets. We revealed insightful relations among models, data types, and data labels in this research area.</p>","PeriodicalId":52823,"journal":{"name":"Frontiers in Computer Science","volume":"3 ","pages":""},"PeriodicalIF":2.6,"publicationDate":"2021-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8153512/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39027802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integration of the ImageJ Ecosystem in the KNIME Analytics Platform. ImageJ生态系统在KNIME分析平台的集成。
IF 2.6 Q3 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-03-01 Epub Date: 2020-03-17 DOI: 10.3389/fcomp.2020.00008
Christian Dietz, Curtis T Rueden, Stefan Helfrich, Ellen T A Dobson, Martin Horn, Jan Eglinger, Edward L Evans, Dalton T McLean, Tatiana Novitskaya, William A Ricke, Nathan M Sherer, Andries Zijlstra, Michael R Berthold, Kevin W Eliceiri

Open-source software tools are often used for analysis of scientific image data due to their flexibility and transparency in dealing with rapidly evolving imaging technologies. The complex nature of image analysis problems frequently requires many tools to be used in conjunction, including image processing and analysis, data processing, machine learning and deep learning, statistical analysis of the results, visualization, correlation to heterogeneous but related data, and more. However, the development, and therefore application, of these computational tools is impeded by a lack of integration across platforms. Integration of tools goes beyond convenience, as it is impractical for one tool to anticipate and accommodate the current and future needs of every user. This problem is emphasized in the field of bioimage analysis, where various rapidly emerging methods are quickly being adopted by researchers. ImageJ is a popular open-source image analysis platform, with contributions from a global community resulting in hundreds of specialized routines for a wide array of scientific tasks. ImageJ's strength lies in its accessibility and extensibility, allowing researchers to easily improve the software to solve their image analysis tasks. However, ImageJ is not designed for development of complex end-to-end image analysis workflows. Scientists are often forced to create highly specialized and hard-to-reproduce scripts to orchestrate individual software fragments and cover the entire life-cycle of an analysis of an image dataset. KNIME Analytics Platform, a user-friendly data integration, analysis, and exploration workflow system, was designed to handle huge amounts of heterogeneous data in a platform-agnostic, computing environment and has been successful in meeting complex end-to-end demands in several communities, such as cheminformatics and mass spectrometry. Similar needs within the bioimage analysis community led to the creation of the KNIME Image Processing extension which integrates ImageJ into KNIME Analytics Platform, enabling researchers to develop reproducible and scalable workflows, integrating a diverse range of analysis tools. Here we present how users and developers alike can leverage the ImageJ ecosystem via the KNIME Image Processing extension to provide robust and extensible image analysis within KNIME workflows. We illustrate the benefits of this integration with examples, as well as representative scientific use cases.

开源软件工具经常用于分析科学图像数据,因为它们在处理快速发展的成像技术方面具有灵活性和透明度。图像分析问题的复杂性经常需要许多工具一起使用,包括图像处理和分析、数据处理、机器学习和深度学习、结果的统计分析、可视化、异构但相关数据的相关性等等。然而,由于缺乏跨平台的集成,这些计算工具的开发和应用受到了阻碍。工具的集成超越了便利性,因为一个工具预测和适应每个用户当前和未来的需求是不切实际的。这个问题在生物图像分析领域得到了强调,研究人员正在迅速采用各种快速出现的方法。ImageJ是一个流行的开源图像分析平台,来自全球社区的贡献产生了数百个专门的例程,用于广泛的科学任务。ImageJ的优势在于它的可访问性和可扩展性,使研究人员可以轻松地改进软件来解决他们的图像分析任务。然而,ImageJ不是为开发复杂的端到端图像分析工作流而设计的。科学家经常被迫创建高度专业化且难以复制的脚本来编排单个软件片段,并覆盖图像数据集分析的整个生命周期。KNIME分析平台是一个用户友好的数据集成、分析和勘探工作流程系统,旨在处理与平台无关的计算环境中的大量异构数据,并已成功满足化学信息学和质谱等多个领域的复杂端到端需求。生物图像分析社区的类似需求导致了KNIME图像处理扩展的创建,该扩展将ImageJ集成到KNIME分析平台中,使研究人员能够开发可重复和可扩展的工作流程,集成了各种分析工具。在这里,我们介绍了用户和开发人员如何通过KNIME图像处理扩展来利用ImageJ生态系统,在KNIME工作流程中提供健壮和可扩展的图像分析。我们通过示例以及具有代表性的科学用例来说明这种集成的好处。
{"title":"Integration of the ImageJ Ecosystem in the KNIME Analytics Platform.","authors":"Christian Dietz,&nbsp;Curtis T Rueden,&nbsp;Stefan Helfrich,&nbsp;Ellen T A Dobson,&nbsp;Martin Horn,&nbsp;Jan Eglinger,&nbsp;Edward L Evans,&nbsp;Dalton T McLean,&nbsp;Tatiana Novitskaya,&nbsp;William A Ricke,&nbsp;Nathan M Sherer,&nbsp;Andries Zijlstra,&nbsp;Michael R Berthold,&nbsp;Kevin W Eliceiri","doi":"10.3389/fcomp.2020.00008","DOIUrl":"https://doi.org/10.3389/fcomp.2020.00008","url":null,"abstract":"<p><p>Open-source software tools are often used for analysis of scientific image data due to their flexibility and transparency in dealing with rapidly evolving imaging technologies. The complex nature of image analysis problems frequently requires many tools to be used in conjunction, including image processing and analysis, data processing, machine learning and deep learning, statistical analysis of the results, visualization, correlation to heterogeneous but related data, and more. However, the development, and therefore application, of these computational tools is impeded by a lack of integration across platforms. Integration of tools goes beyond convenience, as it is impractical for one tool to anticipate and accommodate the current and future needs of every user. This problem is emphasized in the field of bioimage analysis, where various rapidly emerging methods are quickly being adopted by researchers. ImageJ is a popular open-source image analysis platform, with contributions from a global community resulting in hundreds of specialized routines for a wide array of scientific tasks. ImageJ's strength lies in its accessibility and extensibility, allowing researchers to easily improve the software to solve their image analysis tasks. However, ImageJ is not designed for development of complex end-to-end image analysis workflows. Scientists are often forced to create highly specialized and hard-to-reproduce scripts to orchestrate individual software fragments and cover the entire life-cycle of an analysis of an image dataset. KNIME Analytics Platform, a user-friendly data integration, analysis, and exploration workflow system, was designed to handle huge amounts of heterogeneous data in a platform-agnostic, computing environment and has been successful in meeting complex end-to-end demands in several communities, such as cheminformatics and mass spectrometry. Similar needs within the bioimage analysis community led to the creation of the KNIME Image Processing extension which integrates ImageJ into KNIME Analytics Platform, enabling researchers to develop reproducible and scalable workflows, integrating a diverse range of analysis tools. Here we present how users and developers alike can leverage the ImageJ ecosystem via the KNIME Image Processing extension to provide robust and extensible image analysis within KNIME workflows. We illustrate the benefits of this integration with examples, as well as representative scientific use cases.</p>","PeriodicalId":52823,"journal":{"name":"Frontiers in Computer Science","volume":"2 ","pages":""},"PeriodicalIF":2.6,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.3389/fcomp.2020.00008","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38359736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
D-PAttNet: Dynamic Patch-Attentive Deep Network for Action Unit Detection. D-PAttNet:用于动作单元检测的动态补丁-注意力深度网络
IF 2.6 Q3 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2019-11-01 Epub Date: 2019-11-29 DOI: 10.3389/fcomp.2019.00011
Itir Onal Ertugrul, Le Yang, László A Jeni, Jeffrey F Cohn

Facial action units (AUs) relate to specific local facial regions. Recent efforts in automated AU detection have focused on learning the facial patch representations to detect specific AUs. These efforts have encountered three hurdles. First, they implicitly assume that facial patches are robust to head rotation; yet non-frontal rotation is common. Second, mappings between AUs and patches are defined a priori, which ignores co-occurrences among AUs. And third, the dynamics of AUs are either ignored or modeled sequentially rather than simultaneously as in human perception. Inspired by recent advances in human perception, we propose a dynamic patch-attentive deep network, called D-PAttNet, for AU detection that (i) controls for 3D head and face rotation, (ii) learns mappings of patches to AUs, and (iii) models spatiotemporal dynamics. D-PAttNet approach significantly improves upon existing state of the art.

面部动作单元(AU)与特定的局部面部区域有关。最近在自动 AU 检测方面所做的努力主要集中在学习面部斑块表征以检测特定的 AU。这些努力遇到了三个障碍。首先,它们隐含地假定面部补丁对头部旋转具有鲁棒性;然而非正面旋转是很常见的。其次,AU 和斑块之间的映射是先验定义的,忽略了 AU 之间的共现。第三,AUs 的动态要么被忽略,要么被顺序建模,而不是像人类感知那样同时建模。受人类感知领域最新进展的启发,我们提出了一种动态斑块注意力深度网络(称为 D-PAttNet),用于 AU 检测,该网络(i)控制三维头部和面部旋转,(ii)学习斑块到 AU 的映射,(iii)建立时空动态模型。D-PAttNet 方法大大改进了现有的技术水平。
{"title":"D-PAttNet: Dynamic Patch-Attentive Deep Network for Action Unit Detection.","authors":"Itir Onal Ertugrul, Le Yang, László A Jeni, Jeffrey F Cohn","doi":"10.3389/fcomp.2019.00011","DOIUrl":"10.3389/fcomp.2019.00011","url":null,"abstract":"<p><p>Facial action units (AUs) relate to specific local facial regions. Recent efforts in automated AU detection have focused on learning the facial patch representations to detect specific AUs. These efforts have encountered three hurdles. First, they implicitly assume that facial patches are robust to head rotation; yet non-frontal rotation is common. Second, mappings between AUs and patches are defined a priori, which ignores co-occurrences among AUs. And third, the dynamics of AUs are either ignored or modeled sequentially rather than simultaneously as in human perception. Inspired by recent advances in human perception, we propose a dynamic patch-attentive deep network, called D-PAttNet, for AU detection that (i) controls for 3D head and face rotation, (ii) learns mappings of patches to AUs, and (iii) models spatiotemporal dynamics. D-PAttNet approach significantly improves upon existing state of the art.</p>","PeriodicalId":52823,"journal":{"name":"Frontiers in Computer Science","volume":"1 ","pages":""},"PeriodicalIF":2.6,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6953909/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"37536194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Frontiers in Computer Science
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1