Pub Date : 2025-05-09DOI: 10.1016/j.patter.2025.101264
Xiaoyi Chen, Haixu Tang
In a recent issue of Cell Reports Physical Science, Zhao et al. introduced ChemDFM, a foundational large language model designed specifically for chemistry. The model bridges the gap between general-purpose language models and specialized chemical knowledge, including the integration of multimodal capabilities for spectroscopic data interpretation, improved numerical reasoning, and connectivity with chemical tools and databases to enhance practical research applications. This approach demonstrates how domain adaptation can transform AI tools into collaborative research partners for scientific discovery.
{"title":"Designing a large language model for chemists.","authors":"Xiaoyi Chen, Haixu Tang","doi":"10.1016/j.patter.2025.101264","DOIUrl":"10.1016/j.patter.2025.101264","url":null,"abstract":"<p><p>In a recent issue of <i>Cell Reports Physical Science</i>, Zhao et al. introduced ChemDFM, a foundational large language model designed specifically for chemistry. The model bridges the gap between general-purpose language models and specialized chemical knowledge, including the integration of multimodal capabilities for spectroscopic data interpretation, improved numerical reasoning, and connectivity with chemical tools and databases to enhance practical research applications. This approach demonstrates how domain adaptation can transform AI tools into collaborative research partners for scientific discovery.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 5","pages":"101264"},"PeriodicalIF":6.7,"publicationDate":"2025-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12142642/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144250064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-08eCollection Date: 2025-06-13DOI: 10.1016/j.patter.2025.101261
Brandon Theodorou, Cao Xiao, Lucas Glass, Jimeng Sun
We introduce MediSim, a multi-modal generative model for simulating and augmenting electronic health records across multiple modalities, including structured codes, clinical notes, and medical imaging. MediSim employs a multi-granular, autoregressive architecture to simulate missing modalities and visits and iterative, reinforcement learning-based training to improve simulation in low-data settings. Additionally, it utilizes encoder-decoder model pairs to handle complex modalities like notes and images. Experiments on outpatient claims and inpatient ICU datasets have demonstrated MediSim's superiority over baselines in predicting missing codes, creating enriched data, and improving downstream predictive modeling. Specifically, MediSim improved over 74% on missing code prediction, enabled up to 65% better downstream predictive performance compared to original deficient records missing either some visits or entire data modalities, and successfully produced realistic note and X-ray samples for use in downstream tasks. MediSim's ability to generate comprehensive, high-dimensional EHR data has the potential to significantly improve AI applications throughout healthcare.
{"title":"MediSim: Multi-granular simulation for enriching longitudinal, multi-modal electronic health records.","authors":"Brandon Theodorou, Cao Xiao, Lucas Glass, Jimeng Sun","doi":"10.1016/j.patter.2025.101261","DOIUrl":"10.1016/j.patter.2025.101261","url":null,"abstract":"<p><p>We introduce MediSim, a multi-modal generative model for simulating and augmenting electronic health records across multiple modalities, including structured codes, clinical notes, and medical imaging. MediSim employs a multi-granular, autoregressive architecture to simulate missing modalities and visits and iterative, reinforcement learning-based training to improve simulation in low-data settings. Additionally, it utilizes encoder-decoder model pairs to handle complex modalities like notes and images. Experiments on outpatient claims and inpatient ICU datasets have demonstrated MediSim's superiority over baselines in predicting missing codes, creating enriched data, and improving downstream predictive modeling. Specifically, MediSim improved over 74% on missing code prediction, enabled up to 65% better downstream predictive performance compared to original deficient records missing either some visits or entire data modalities, and successfully produced realistic note and X-ray samples for use in downstream tasks. MediSim's ability to generate comprehensive, high-dimensional EHR data has the potential to significantly improve AI applications throughout healthcare.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 6","pages":"101261"},"PeriodicalIF":6.7,"publicationDate":"2025-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12191717/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144508674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-08eCollection Date: 2025-06-13DOI: 10.1016/j.patter.2025.101260
Banghao Chen, Zhaofeng Zhang, Nicolas Langrené, Shengxin Zhu
This review explores the role of prompt engineering in unleashing the capabilities of large language models (LLMs). Prompt engineering is the process of structuring inputs, and it has emerged as a crucial technique for maximizing the utility and accuracy of these models. Both foundational and advanced prompt engineering methodologies-including techniques such as self-consistency, chain of thought, and generated knowledge, which can significantly enhance the performance of models-are explored in this paper. Additionally, the prompt methods for vision language models (VLMs) are examined in detail. Prompt methods are evaluated with subjective and objective metrics, ensuring a robust analysis of their efficacy. Critical to this discussion is the role of prompt engineering in artificial intelligence (AI) security, particularly in terms of defending against adversarial attacks that exploit vulnerabilities in LLMs. Strategies for minimizing these risks and improving the robustness of models are thoroughly reviewed. Finally, we provide a perspective for future research and applications.
{"title":"Unleashing the potential of prompt engineering for large language models.","authors":"Banghao Chen, Zhaofeng Zhang, Nicolas Langrené, Shengxin Zhu","doi":"10.1016/j.patter.2025.101260","DOIUrl":"10.1016/j.patter.2025.101260","url":null,"abstract":"<p><p>This review explores the role of prompt engineering in unleashing the capabilities of large language models (LLMs). Prompt engineering is the process of structuring inputs, and it has emerged as a crucial technique for maximizing the utility and accuracy of these models. Both foundational and advanced prompt engineering methodologies-including techniques such as self-consistency, chain of thought, and generated knowledge, which can significantly enhance the performance of models-are explored in this paper. Additionally, the prompt methods for vision language models (VLMs) are examined in detail. Prompt methods are evaluated with subjective and objective metrics, ensuring a robust analysis of their efficacy. Critical to this discussion is the role of prompt engineering in artificial intelligence (AI) security, particularly in terms of defending against adversarial attacks that exploit vulnerabilities in LLMs. Strategies for minimizing these risks and improving the robustness of models are thoroughly reviewed. Finally, we provide a perspective for future research and applications.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 6","pages":"101260"},"PeriodicalIF":6.7,"publicationDate":"2025-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12191768/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144508676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-06eCollection Date: 2025-06-13DOI: 10.1016/j.patter.2025.101262
Yi-Fan Li, Xiaoyong Pan, Hong-Bin Shen
We describe NLSExplorer, an interpretable approach for nuclear localization signal (NLS) prediction. By utilizing the extracted information on nuclear-specific sites from the protein language model to assist in NLS detection, NLSExplorer achieves superior performance with greater than 10% improvement in the F1 score compared with existing methods on benchmark datasets and highlights other nuclear transport segments. We applied NLSExplorer to the nucleus-localized proteins in the Swiss-Prot database to extract valuable segments. A comprehensive analysis of these segments revealed a potential NLS landscape and uncovered features of nuclear transport segments across 416 species. This study introduces a powerful tool for exploring the NLS universe and provides a versatile network that can efficiently detect characteristic domains and motifs.
{"title":"Discovering the nuclear localization signal universe through a deep learning model with interpretable attention units.","authors":"Yi-Fan Li, Xiaoyong Pan, Hong-Bin Shen","doi":"10.1016/j.patter.2025.101262","DOIUrl":"10.1016/j.patter.2025.101262","url":null,"abstract":"<p><p>We describe NLSExplorer, an interpretable approach for nuclear localization signal (NLS) prediction. By utilizing the extracted information on nuclear-specific sites from the protein language model to assist in NLS detection, NLSExplorer achieves superior performance with greater than 10% improvement in the F1 score compared with existing methods on benchmark datasets and highlights other nuclear transport segments. We applied NLSExplorer to the nucleus-localized proteins in the Swiss-Prot database to extract valuable segments. A comprehensive analysis of these segments revealed a potential NLS landscape and uncovered features of nuclear transport segments across 416 species. This study introduces a powerful tool for exploring the NLS universe and provides a versatile network that can efficiently detect characteristic domains and motifs.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 6","pages":"101262"},"PeriodicalIF":6.7,"publicationDate":"2025-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12191761/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144508670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-02eCollection Date: 2025-06-13DOI: 10.1016/j.patter.2025.101257
Beatrice Savoldi, Jasmijn Bastings, Luisa Bentivogli, Eva Vanmassenhove
Gender bias in machine translation (MT) has been studied for over a decade, a time marked by societal, linguistic, and technological shifts. With the early optimism for a quick solution in mind, we review over 100 studies on the topic and uncover a more complex reality-one that resists a simple technical fix. While we identify key trends and advancements, persistent gaps remain. We argue that there is no simple technical solution to bias. Building on insights from our review, we examine the growing prominence of large language models and discuss the challenges and opportunities they present in the context of gender bias and translation. By doing so, we hope to inspire future work in the field to break with past limitations and to be less focused on a technical fix; more user-centric, multilingual, and multiculturally diverse; more personalized; and better grounded in real-world needs.
{"title":"A decade of gender bias in machine translation.","authors":"Beatrice Savoldi, Jasmijn Bastings, Luisa Bentivogli, Eva Vanmassenhove","doi":"10.1016/j.patter.2025.101257","DOIUrl":"10.1016/j.patter.2025.101257","url":null,"abstract":"<p><p>Gender bias in machine translation (MT) has been studied for over a decade, a time marked by societal, linguistic, and technological shifts. With the early optimism for a quick solution in mind, we review over 100 studies on the topic and uncover a more complex reality-one that resists a simple technical fix. While we identify key trends and advancements, persistent gaps remain. We argue that there is no simple technical solution to bias. Building on insights from our review, we examine the growing prominence of large language models and discuss the challenges and opportunities they present in the context of gender bias and translation. By doing so, we hope to inspire future work in the field to break with past limitations and to be less focused on a technical fix; more user-centric, multilingual, and multiculturally diverse; more personalized; and better grounded in real-world needs.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 6","pages":"101257"},"PeriodicalIF":6.7,"publicationDate":"2025-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12191736/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144508667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Natural products (NPs) play a vital role in drug discovery, with many FDA-approved drugs derived from these compounds. Despite their significance, the biosynthetic pathways of NPs remain poorly characterized due to their inherent complexity and the limitations of traditional retrosynthesis methods in predicting such intricate reactions. While template-free machine learning models have demonstrated promise in organic synthesis, their application to biosynthetic pathways is still in its infancy. Addressing this gap, we propose the graph-sequence enhanced transformer (GSETransformer), which leverages both graph structural information and sequential dependencies to achieve superior performance in addressing the complexity of biosynthetic data. When evaluated on benchmark datasets, GSETransformer achieves state-of-the-art performance in single- and multi-step retrosynthesis tasks. These results highlight its effectiveness in computational biosynthesis and its potential to facilitate the design of NP-based therapeutics.
{"title":"Graph-sequence enhanced transformer for template-free prediction of natural product biosynthesis.","authors":"Shan Cong, Meng Zhang, Yu Song, Sihao Chang, Jing Tian, Hongji Zeng, Hongchao Ji","doi":"10.1016/j.patter.2025.101259","DOIUrl":"10.1016/j.patter.2025.101259","url":null,"abstract":"<p><p>Natural products (NPs) play a vital role in drug discovery, with many FDA-approved drugs derived from these compounds. Despite their significance, the biosynthetic pathways of NPs remain poorly characterized due to their inherent complexity and the limitations of traditional retrosynthesis methods in predicting such intricate reactions. While template-free machine learning models have demonstrated promise in organic synthesis, their application to biosynthetic pathways is still in its infancy. Addressing this gap, we propose the graph-sequence enhanced transformer (GSETransformer), which leverages both graph structural information and sequential dependencies to achieve superior performance in addressing the complexity of biosynthetic data. When evaluated on benchmark datasets, GSETransformer achieves state-of-the-art performance in single- and multi-step retrosynthesis tasks. These results highlight its effectiveness in computational biosynthesis and its potential to facilitate the design of NP-based therapeutics.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 8","pages":"101259"},"PeriodicalIF":7.4,"publicationDate":"2025-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12365517/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144972454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-29eCollection Date: 2025-05-09DOI: 10.1016/j.patter.2025.101242
Joseph Bingham, Saman Zonouz, Dvir Aran
Neural networks have long strived to emulate the learning capabilities of the human brain. While deep neural networks (DNNs) draw inspiration from the brain in neuron design, their training methods diverge from biological foundations. Backpropagation, the primary training method for DNNs, requires substantial computational resources and fully labeled datasets, presenting major bottlenecks in development and application. This work demonstrates that by returning to biomimicry, specifically mimicking how the brain learns through pruning, we can solve various classical machine learning problems while utilizing orders of magnitude fewer computational resources and no labels. Our experiments successfully personalized multiple speech recognition and image classification models, including ResNet50 on ImageNet, resulting in an increased sparsity of approximately 70% while simultaneously improving model accuracy to around 90%, all without the limitations of backpropagation. This biologically inspired approach offers a promising avenue for efficient, personalized machine learning models in resource-constrained environments.
{"title":"Fine-Pruning: A biologically inspired algorithm for personalization of machine learning models.","authors":"Joseph Bingham, Saman Zonouz, Dvir Aran","doi":"10.1016/j.patter.2025.101242","DOIUrl":"10.1016/j.patter.2025.101242","url":null,"abstract":"<p><p>Neural networks have long strived to emulate the learning capabilities of the human brain. While deep neural networks (DNNs) draw inspiration from the brain in neuron design, their training methods diverge from biological foundations. Backpropagation, the primary training method for DNNs, requires substantial computational resources and fully labeled datasets, presenting major bottlenecks in development and application. This work demonstrates that by returning to biomimicry, specifically mimicking how the brain learns through pruning, we can solve various classical machine learning problems while utilizing orders of magnitude fewer computational resources and no labels. Our experiments successfully personalized multiple speech recognition and image classification models, including ResNet50 on ImageNet, resulting in an increased sparsity of approximately 70% while simultaneously improving model accuracy to around 90%, all without the limitations of backpropagation. This biologically inspired approach offers a promising avenue for efficient, personalized machine learning models in resource-constrained environments.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 5","pages":"101242"},"PeriodicalIF":6.7,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12142609/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144250067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-29eCollection Date: 2025-07-11DOI: 10.1016/j.patter.2025.101238
Antonio Bikić, Wolfram H P Pernice
Multisensory perception produces vast amounts of data requiring efficient processing. This paper focuses on the multisensory example of touch in biological and artificial systems. We integrate philosophical theories of multisensory perception with neuromorphic hardware and demonstrate how classical sensory integration concepts can enhance artificial sensory systems. This approach bridges theoretical neuroscience and computational applications using philosophical tools. We contrast human touch perception, involving feature binding, with artificial perception in neuromorphic computing, where such integration is absent. Two theoretical frameworks are of interest, feature binding and modalities as conventional kinds, in evaluating their relevance to artificial touch. Our findings suggest that a hardware-tailored adaptation of the conventional modalities approach accurately reflects artificial touch perception. Unlike human perception, artificial systems process sensory data separately, lacking binding mechanisms. We explore the implications of these differences, highlighting challenges in replicating human sensory experiences and the role of subjective experience in perception.
{"title":"Emulating sensation by bridging neuromorphic computing and multisensory integration.","authors":"Antonio Bikić, Wolfram H P Pernice","doi":"10.1016/j.patter.2025.101238","DOIUrl":"10.1016/j.patter.2025.101238","url":null,"abstract":"<p><p>Multisensory perception produces vast amounts of data requiring efficient processing. This paper focuses on the multisensory example of touch in biological and artificial systems. We integrate philosophical theories of multisensory perception with neuromorphic hardware and demonstrate how classical sensory integration concepts can enhance artificial sensory systems. This approach bridges theoretical neuroscience and computational applications using philosophical tools. We contrast human touch perception, involving feature binding, with artificial perception in neuromorphic computing, where such integration is absent. Two theoretical frameworks are of interest, feature binding and modalities as conventional kinds, in evaluating their relevance to artificial touch. Our findings suggest that a hardware-tailored adaptation of the conventional modalities approach accurately reflects artificial touch perception. Unlike human perception, artificial systems process sensory data separately, lacking binding mechanisms. We explore the implications of these differences, highlighting challenges in replicating human sensory experiences and the role of subjective experience in perception.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 7","pages":"101238"},"PeriodicalIF":7.4,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12416093/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145030930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-29eCollection Date: 2025-06-13DOI: 10.1016/j.patter.2025.101236
Frederik F Flöther, Daniel Blankenberg, Maria Demidik, Karl Jansen, Raga Krishnakumar, Rajiv Krishnakumar, Nouamane Laanait, Laxmi Parida, Carl Y Saab, Filippo Utro
Biomarkers play a central role in medicine's gradual progress toward proactive, personalized precision diagnostics and interventions. However, finding biomarkers that provide very early indicators of a change in health status, for example, for multifactorial diseases, has been challenging. The discovery of such biomarkers stands to benefit significantly from advanced information processing and means to detect complex correlations, which quantum computing offers. In this perspective, quantum algorithms, particularly in machine learning, are mapped to key applications in biomarker discovery. The opportunities and challenges associated with the algorithms and applications are discussed. The analysis is structured according to different data types-multidimensional, time series, and erroneous data-and covers key data modalities in healthcare-electronic health records, omics, and medical images. An outlook is provided concerning open research challenges.
{"title":"How quantum computing can enhance biomarker discovery.","authors":"Frederik F Flöther, Daniel Blankenberg, Maria Demidik, Karl Jansen, Raga Krishnakumar, Rajiv Krishnakumar, Nouamane Laanait, Laxmi Parida, Carl Y Saab, Filippo Utro","doi":"10.1016/j.patter.2025.101236","DOIUrl":"10.1016/j.patter.2025.101236","url":null,"abstract":"<p><p>Biomarkers play a central role in medicine's gradual progress toward proactive, personalized precision diagnostics and interventions. However, finding biomarkers that provide very early indicators of a change in health status, for example, for multifactorial diseases, has been challenging. The discovery of such biomarkers stands to benefit significantly from advanced information processing and means to detect complex correlations, which quantum computing offers. In this perspective, quantum algorithms, particularly in machine learning, are mapped to key applications in biomarker discovery. The opportunities and challenges associated with the algorithms and applications are discussed. The analysis is structured according to different data types-multidimensional, time series, and erroneous data-and covers key data modalities in healthcare-electronic health records, omics, and medical images. An outlook is provided concerning open research challenges.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 6","pages":"101236"},"PeriodicalIF":6.7,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12191739/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144508673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-04-28eCollection Date: 2025-08-08DOI: 10.1016/j.patter.2025.101240
Nina de Lacy, Michael Ramshaw, Wai Yin Lam
Many diseases are the end outcomes of multifactorial risks that interact and increment over months or years. Time-series AI methods have attracted increasing interest given their ability to operate on native time-series data to predict disease outcomes. Instantiating such models in risk stratification tools has proceeded more slowly, in part limited by factors such as structural complexity, model size, and explainability. Here, we present RiskPath, an explainable AI toolbox that offers advanced time-series methods and additional functionality relevant to risk stratification use cases in classic and emerging longitudinal cohorts. Theoretically informed optimization is integrated in prediction to specify optimal model topology or explore performance-complexity trade-offs. Accompanying modules allow the user to map the changing importance of predictors over the disease course, visualize the most important antecedent time epochs contributing to disease risk, or remove predictors to construct compact models for clinical applications with minimal performance impact.
{"title":"RiskPath: Explainable deep learning for multistep biomedical prediction in longitudinal data.","authors":"Nina de Lacy, Michael Ramshaw, Wai Yin Lam","doi":"10.1016/j.patter.2025.101240","DOIUrl":"10.1016/j.patter.2025.101240","url":null,"abstract":"<p><p>Many diseases are the end outcomes of multifactorial risks that interact and increment over months or years. Time-series AI methods have attracted increasing interest given their ability to operate on native time-series data to predict disease outcomes. Instantiating such models in risk stratification tools has proceeded more slowly, in part limited by factors such as structural complexity, model size, and explainability. Here, we present RiskPath, an explainable AI toolbox that offers advanced time-series methods and additional functionality relevant to risk stratification use cases in classic and emerging longitudinal cohorts. Theoretically informed optimization is integrated in prediction to specify optimal model topology or explore performance-complexity trade-offs. Accompanying modules allow the user to map the changing importance of predictors over the disease course, visualize the most important antecedent time epochs contributing to disease risk, or remove predictors to construct compact models for clinical applications with minimal performance impact.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 8","pages":"101240"},"PeriodicalIF":7.4,"publicationDate":"2025-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12365533/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144972465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}