Recent studies have demonstrated that high-quality annotated data are crucial for segmentation performance. However, incomplete or corrupted mask annotations remain common, limiting supervised learning. To address this, we introduce a mask-reconstruction task, referred to as masked segmentation label modeling (MSLM), which refines partially occluded labels by leveraging visible regions without manual annotations. We further propose the label masked autoencoder (L-MAE), which identifies erroneous regions and reconstructs them through contextual inference. The L-MAE fuses incomplete labels and corresponding images into a unified map for reconstruction, and an image patch supplement (IPS) algorithm restores missing image information, improving the average mean intersection over union (mIoU) by 4.1%. To validate the L-MAE, we train segmentation models on a degraded and L-MAE-enhanced Pascal VOC dataset, with the latter achieving a 13.5% mIoU improvement. The L-MAE attains predict area (PA)-mIoU scores of 91.0% on Pascal VOC 2012 and 86.4% on Cityscapes, outperforming state-of-the-art supervised segmentation models.
{"title":"A label masked autoencoder for image-guided segmentation label completion.","authors":"Jiaru Jia, Mingzhe Liu, Dongfen Li, Xin Chen, Ruili Wang, Linlin Zhuo, Keqin Li","doi":"10.1016/j.patter.2025.101455","DOIUrl":"https://doi.org/10.1016/j.patter.2025.101455","url":null,"abstract":"<p><p>Recent studies have demonstrated that high-quality annotated data are crucial for segmentation performance. However, incomplete or corrupted mask annotations remain common, limiting supervised learning. To address this, we introduce a mask-reconstruction task, referred to as masked segmentation label modeling (MSLM), which refines partially occluded labels by leveraging visible regions without manual annotations. We further propose the label masked autoencoder (L-MAE), which identifies erroneous regions and reconstructs them through contextual inference. The L-MAE fuses incomplete labels and corresponding images into a unified map for reconstruction, and an image patch supplement (IPS) algorithm restores missing image information, improving the average mean intersection over union (mIoU) by 4.1%. To validate the L-MAE, we train segmentation models on a degraded and L-MAE-enhanced Pascal VOC dataset, with the latter achieving a 13.5% mIoU improvement. The L-MAE attains predict area (PA)-mIoU scores of 91.0% on Pascal VOC 2012 and 86.4% on Cityscapes, outperforming state-of-the-art supervised segmentation models.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"7 2","pages":"101455"},"PeriodicalIF":7.4,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12998695/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147487457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Autonomous AI-to-AI creative systems promise new frontiers in machine creativity, yet we show that they systematically converge toward generic outputs. We built iterative feedback loops between Stable Diffusion XL (SDXL; image generation) and Large Language and Vision Assistant (LLaVA; image description), forming autonomous text → image → text → image cycles. Across 700 trajectories with diverse prompts and 7 temperature settings over 100 iterations, all runs converged to nearly identical visuals-what we term "visual elevator music." Quantitative analysis revealed just 12 dominant motifs with commercially safe aesthetics, such as stormy lighthouses and palatial interiors. This convergence persisted across model pairs, indicating structural limits in cross-modal AI creativity. The effect mirrors human cultural transmission, where iterated learning amplifies cognitive biases, but here, diversity collapses entirely as AI loops gravitate to high-probability attractors in training data. Our findings expose hidden homogenizing tendencies in current architectures and underscore the need for anti-convergence mechanisms and sustained human-AI interplay to preserve creative diversity.
自主的人工智能对人工智能的创造性系统为机器创造力提供了新的领域,但我们表明,它们系统地向通用输出收敛。我们在Stable Diffusion XL (SDXL;图像生成)和Large Language and Vision Assistant (LLaVA;图像描述)之间构建迭代反馈循环,形成自主的文本→图像→文本→图像循环。在700条轨道上,不同的提示和7种温度设置超过100次迭代,所有的运行都汇聚成几乎相同的视觉效果——我们称之为“视觉电梯音乐”。定量分析显示,只有12个占主导地位的主题具有商业安全的美学,如暴风雨般的灯塔和富丽堂皇的室内装饰。这种趋同在模型对中持续存在,表明跨模式人工智能创造力的结构性限制。这种效应反映了人类的文化传播,反复的学习放大了认知偏见,但在这里,多样性完全崩溃,因为人工智能循环被训练数据中的高概率吸引子所吸引。我们的研究结果揭示了当前架构中隐藏的同质化趋势,并强调了反收敛机制和持续的人类与人工智能相互作用的必要性,以保持创造性的多样性。
{"title":"Autonomous language-image generation loops converge to generic visual motifs.","authors":"Arend Hintze, Frida Proschinger Åström, Jory Schossau","doi":"10.1016/j.patter.2025.101451","DOIUrl":"10.1016/j.patter.2025.101451","url":null,"abstract":"<p><p>Autonomous AI-to-AI creative systems promise new frontiers in machine creativity, yet we show that they systematically converge toward generic outputs. We built iterative feedback loops between Stable Diffusion XL (SDXL; image generation) and Large Language and Vision Assistant (LLaVA; image description), forming autonomous text → image → text → image cycles. Across 700 trajectories with diverse prompts and 7 temperature settings over 100 iterations, all runs converged to nearly identical visuals-what we term \"visual elevator music.\" Quantitative analysis revealed just 12 dominant motifs with commercially safe aesthetics, such as stormy lighthouses and palatial interiors. This convergence persisted across model pairs, indicating structural limits in cross-modal AI creativity. The effect mirrors human cultural transmission, where iterated learning amplifies cognitive biases, but here, diversity collapses entirely as AI loops gravitate to high-probability attractors in training data. Our findings expose hidden homogenizing tendencies in current architectures and underscore the need for anti-convergence mechanisms and sustained human-AI interplay to preserve creative diversity.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"7 1","pages":"101451"},"PeriodicalIF":7.4,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12827715/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146053908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-17eCollection Date: 2026-01-09DOI: 10.1016/j.patter.2025.101430
Alex de Vries-Gao
Although there are ways to estimate the global power demand of artificial intelligence (AI) systems, it remains challenging to quantify the associated carbon and water footprints. The lack of distinction between AI and non-AI workloads in the environmental reports of data center operators makes it possible to assess the environmental impact of AI workloads only by approximating them through data centers' general performance metrics. The environmental disclosure of tech companies is, however, often insufficient to determine even the total data center performance of these companies. The shortcomings in the environmental disclosure of data center operators could be remedied with new policies mandating the disclosure of additional metrics. Because the environmental impact of data centers is growing rapidly, the urgency of transparency in the tech sector is also increasing. The carbon footprint of AI systems alone could be between 32.6 and 79.7 million tons of CO2 emissions in 2025, while the water footprint could reach 312.5-764.6 billion L.
{"title":"The carbon and water footprints of data centers and what this could mean for artificial intelligence.","authors":"Alex de Vries-Gao","doi":"10.1016/j.patter.2025.101430","DOIUrl":"10.1016/j.patter.2025.101430","url":null,"abstract":"<p><p>Although there are ways to estimate the global power demand of artificial intelligence (AI) systems, it remains challenging to quantify the associated carbon and water footprints. The lack of distinction between AI and non-AI workloads in the environmental reports of data center operators makes it possible to assess the environmental impact of AI workloads only by approximating them through data centers' general performance metrics. The environmental disclosure of tech companies is, however, often insufficient to determine even the total data center performance of these companies. The shortcomings in the environmental disclosure of data center operators could be remedied with new policies mandating the disclosure of additional metrics. Because the environmental impact of data centers is growing rapidly, the urgency of transparency in the tech sector is also increasing. The carbon footprint of AI systems alone could be between 32.6 and 79.7 million tons of CO<sub>2</sub> emissions in 2025, while the water footprint could reach 312.5-764.6 billion L.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"7 1","pages":"101430"},"PeriodicalIF":7.4,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12827721/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-16eCollection Date: 2026-01-09DOI: 10.1016/j.patter.2025.101432
Galina Tremper, Torben Brenner, Moanes Ben Amor, Tobias Kussel, Martin Lablans
Record linkage and pseudonymization are crucial tasks in collaborative biomedical research. Data for a patient are rarely stored in one place and therefore often need to be linked and integrated across multiple institutions. Mainzelliste is an open-source software solution designed to solve these challenges by providing a comprehensive and flexible toolkit for pseudonymization, record linkage, and consent management. It supports a variety of pseudonyms, record linkage methods, and modular, informed patient consents. A highly flexible REST application programming interface (API) allows tight integration into existing applications and workflows. Since its initial release in 2015, Mainzelliste has evolved into a vibrant open-source software solution "by researchers, for researchers" including a user-friendly graphical interface, support for HL7 FHIR for consent and patient data, and record linkage based on secure multi-party computation, thereby supporting secure and efficient biomedical research.
{"title":"Mainzelliste: Ten years of pseudonymization, record linkage, and informed consent management.","authors":"Galina Tremper, Torben Brenner, Moanes Ben Amor, Tobias Kussel, Martin Lablans","doi":"10.1016/j.patter.2025.101432","DOIUrl":"10.1016/j.patter.2025.101432","url":null,"abstract":"<p><p>Record linkage and pseudonymization are crucial tasks in collaborative biomedical research. Data for a patient are rarely stored in one place and therefore often need to be linked and integrated across multiple institutions. Mainzelliste is an open-source software solution designed to solve these challenges by providing a comprehensive and flexible toolkit for pseudonymization, record linkage, and consent management. It supports a variety of pseudonyms, record linkage methods, and modular, informed patient consents. A highly flexible REST application programming interface (API) allows tight integration into existing applications and workflows. Since its initial release in 2015, Mainzelliste has evolved into a vibrant open-source software solution \"by researchers, for researchers\" including a user-friendly graphical interface, support for HL7 FHIR for consent and patient data, and record linkage based on secure multi-party computation, thereby supporting secure and efficient biomedical research.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"7 1","pages":"101432"},"PeriodicalIF":7.4,"publicationDate":"2025-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12827741/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-12DOI: 10.1016/j.patter.2025.101424
Alexandros Christopoulos, Athina Tzovara
Sleep is one of the most essential parts of our daily lives. The gold standard for studying sleep is polysomnography (PSG) recordings. The first step of analyzing PSG recordings involves splitting them into sleep stages, which is performed manually. Machine learning algorithms have attempted to automate the tedious task of sleep scoring, mostly via supervised learning. A recent study in Patterns introduces AISleep, a novel algorithm approaching the task of sleep scoring in an unsupervised framework. This algorithm is based on humanly interpretable features and provides robust results across different datasets and age groups.
{"title":"Sleep staging through an unsupervised learning lens.","authors":"Alexandros Christopoulos, Athina Tzovara","doi":"10.1016/j.patter.2025.101424","DOIUrl":"10.1016/j.patter.2025.101424","url":null,"abstract":"<p><p>Sleep is one of the most essential parts of our daily lives. The gold standard for studying sleep is polysomnography (PSG) recordings. The first step of analyzing PSG recordings involves splitting them into sleep stages, which is performed manually. Machine learning algorithms have attempted to automate the tedious task of sleep scoring, mostly via supervised learning. A recent study in <i>Patterns</i> introduces AISleep, a novel algorithm approaching the task of sleep scoring in an unsupervised framework. This algorithm is based on humanly interpretable features and provides robust results across different datasets and age groups.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 12","pages":"101424"},"PeriodicalIF":7.4,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12745977/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145865523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-12DOI: 10.1016/j.patter.2025.101397
Angelina Wang, Daniel E Ho, Sanmi Koyejo
Standard offline evaluations for language models fail to capture how these models actually behave in practice, where personalization fundamentally alters model behavior. In this work, we provide empirical evidence showcasing this phenomenon by comparing offline evaluations to field evaluations conducted by having 800 real users of ChatGPT and Gemini pose benchmark and other questions to their chat interfaces.
{"title":"The inadequacy of offline large language model evaluations: A need to account for personalization in model behavior.","authors":"Angelina Wang, Daniel E Ho, Sanmi Koyejo","doi":"10.1016/j.patter.2025.101397","DOIUrl":"10.1016/j.patter.2025.101397","url":null,"abstract":"<p><p>Standard offline evaluations for language models fail to capture how these models actually behave in practice, where personalization fundamentally alters model behavior. In this work, we provide empirical evidence showcasing this phenomenon by comparing offline evaluations to field evaluations conducted by having 800 real users of ChatGPT and Gemini pose benchmark and other questions to their chat interfaces.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 12","pages":"101397"},"PeriodicalIF":7.4,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12745978/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145865746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-03eCollection Date: 2026-02-13DOI: 10.1016/j.patter.2025.101426
Xiang Zhang, Mingyue Zhao, Fei Yao, Wenxin Ma, Jin Zhang, Yueze Li, Xiuxiu Zhou, Yu Guan, Yi Xiao, Li Fan, Shaohua Kevin Zhou, Shiyuan Liu
Emphysema, a diffuse and heterogeneous phenotype of chronic obstructive pulmonary disease (COPD), carries substantial morbidity and elevates lung cancer risk. While computed tomography (CT) aids in detection and monitoring, current deep learning methods depend on large annotated datasets. Unsupervised anomaly detection (UAD) provides an alternative but faces challenges with emphysema anomalies and weak emphysema semantics. In this study, we propose a self-supervised framework trained exclusively on non-emphysema CT scans using synthetically generated lesions to guide pixel-level anomaly modeling. We introduce EDLNet, an encoder-decoder architecture with spatial-channel refinement and adaptive feature fusion for emphysema detection and localization, followed by an unsupervised manner for emphysema staging. Multi-center evaluations show that our framework outperforms existing UAD approaches in detection and localization, while achieving a mean staging accuracy of 93.13% and a macro AUROC of 99.08%. This approach bridges clinical knowledge and artificial intelligence, offering a scalable and interpretable solution for lung disease analysis.
{"title":"A self-supervised framework for emphysema anomaly detection and staging in computed tomography scans.","authors":"Xiang Zhang, Mingyue Zhao, Fei Yao, Wenxin Ma, Jin Zhang, Yueze Li, Xiuxiu Zhou, Yu Guan, Yi Xiao, Li Fan, Shaohua Kevin Zhou, Shiyuan Liu","doi":"10.1016/j.patter.2025.101426","DOIUrl":"https://doi.org/10.1016/j.patter.2025.101426","url":null,"abstract":"<p><p>Emphysema, a diffuse and heterogeneous phenotype of chronic obstructive pulmonary disease (COPD), carries substantial morbidity and elevates lung cancer risk. While computed tomography (CT) aids in detection and monitoring, current deep learning methods depend on large annotated datasets. Unsupervised anomaly detection (UAD) provides an alternative but faces challenges with emphysema anomalies and weak emphysema semantics. In this study, we propose a self-supervised framework trained exclusively on non-emphysema CT scans using synthetically generated lesions to guide pixel-level anomaly modeling. We introduce EDLNet, an encoder-decoder architecture with spatial-channel refinement and adaptive feature fusion for emphysema detection and localization, followed by an unsupervised manner for emphysema staging. Multi-center evaluations show that our framework outperforms existing UAD approaches in detection and localization, while achieving a mean staging accuracy of 93.13% and a macro AUROC of 99.08%. This approach bridges clinical knowledge and artificial intelligence, offering a scalable and interpretable solution for lung disease analysis.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"7 2","pages":"101426"},"PeriodicalIF":7.4,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921507/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147272177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01eCollection Date: 2025-12-12DOI: 10.1016/j.patter.2025.101428
David Fernandez Bonet, Johanna I Blumenthal, Shuai Lang, Simon K Dahlberg, Ian T Hoffecker
DNA barcode networks are the basis of sequencing-based microscopy, an emerging family of chemical imaging methods aiming to reconstruct spatial information, without optics, using sequencing technology. These methods capture microscopic spatial information by forming networks composed of many local chemical interactions, each marked by a unique, DNA-based barcode. However, the fundamental laws governing such networks are not yet understood, and spatial barcode networks are influenced by structural distortions such as false or shortcut edges. Current methods lack ground-truth-free tools to validate spatial quality, and we address this with a framework for topology-based quality control. We define a fundamental feature of spatial networks, spatial coherence, which quantifies geometric self-consistency in a network. By formalizing this relationship into quantitative metrics adapted from classical geometric rules, we could quantify spatial distortions by using only network data and show how these can be used as an optimization criterion to iteratively improve spatial reconstruction.
{"title":"Spatial coherence in DNA barcode networks.","authors":"David Fernandez Bonet, Johanna I Blumenthal, Shuai Lang, Simon K Dahlberg, Ian T Hoffecker","doi":"10.1016/j.patter.2025.101428","DOIUrl":"10.1016/j.patter.2025.101428","url":null,"abstract":"<p><p>DNA barcode networks are the basis of sequencing-based microscopy, an emerging family of chemical imaging methods aiming to reconstruct spatial information, without optics, using sequencing technology. These methods capture microscopic spatial information by forming networks composed of many local chemical interactions, each marked by a unique, DNA-based barcode. However, the fundamental laws governing such networks are not yet understood, and spatial barcode networks are influenced by structural distortions such as false or shortcut edges. Current methods lack ground-truth-free tools to validate spatial quality, and we address this with a framework for topology-based quality control. We define a fundamental feature of spatial networks, spatial coherence, which quantifies geometric self-consistency in a network. By formalizing this relationship into quantitative metrics adapted from classical geometric rules, we could quantify spatial distortions by using only network data and show how these can be used as an optimization criterion to iteratively improve spatial reconstruction.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"6 12","pages":"101428"},"PeriodicalIF":7.4,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12745985/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145865547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-26eCollection Date: 2026-02-13DOI: 10.1016/j.patter.2025.101425
Oriol Gracia Carmona, Vilde Leipart, Gro V Amdam, Christine Orengo, Franca Fraternali
Protein language models (PLMs) are increasingly used to assess the impact of genetic variants, achieving high accuracy and often outperforming traditional pathogenicity predictors. They enable zero-shot inference, making predictions without task-specific fine-tuning, though studying in-frame insertions and deletions (indels) remains challenging due to altered protein lengths and limited annotated datasets. Here, we present IndeLLM, a scoring approach for indel pathogenicity that accounts for sequence length differences. Our zero-shot method relies solely on sequence information, requires minimal computing resources, and achieves performance comparable to existing predictors. Building on this, we developed a Siamese network via transfer learning that outperformed all tested indel predictors (Matthews correlation coefficient = 0.77). To enhance accessibility, we provide a plug-and-play Google Colab notebook for using IndeLLM and visualizing the impact of indels on protein sequence and structure. The tool is freely available on GitHub and Google Colab.
{"title":"Leveraging protein language models and a scoring function for indel characterization and transfer learning.","authors":"Oriol Gracia Carmona, Vilde Leipart, Gro V Amdam, Christine Orengo, Franca Fraternali","doi":"10.1016/j.patter.2025.101425","DOIUrl":"https://doi.org/10.1016/j.patter.2025.101425","url":null,"abstract":"<p><p>Protein language models (PLMs) are increasingly used to assess the impact of genetic variants, achieving high accuracy and often outperforming traditional pathogenicity predictors. They enable zero-shot inference, making predictions without task-specific fine-tuning, though studying in-frame insertions and deletions (indels) remains challenging due to altered protein lengths and limited annotated datasets. Here, we present IndeLLM, a scoring approach for indel pathogenicity that accounts for sequence length differences. Our zero-shot method relies solely on sequence information, requires minimal computing resources, and achieves performance comparable to existing predictors. Building on this, we developed a Siamese network via transfer learning that outperformed all tested indel predictors (Matthews correlation coefficient = 0.77). To enhance accessibility, we provide a plug-and-play Google Colab notebook for using IndeLLM and visualizing the impact of indels on protein sequence and structure. The tool is freely available on GitHub and Google Colab.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"7 2","pages":"101425"},"PeriodicalIF":7.4,"publicationDate":"2025-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921505/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147272287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-20eCollection Date: 2026-02-13DOI: 10.1016/j.patter.2025.101423
Felipe Yáñez, Xiaoliang Luo, Omar Valerio Minero, Bradley C Love
Large language models (LLMs) can surpass humans in certain forecasting tasks. What role does this leave for humans in the overall decision process? One possibility is that humans, despite performing worse than LLMs, can still add value when teamed with them. A human and machine team can surpass each individual teammate when team members' confidence is well calibrated and team members diverge in which tasks they find difficult (i.e., calibration and diversity are needed). We simplified and extended a Bayesian approach to combining judgments using a logistic regression framework that integrates confidence-weighted judgments for any number of team members. Using this straightforward method, we demonstrated its effectiveness in both image classification and neuroscience forecasting tasks. Combining human judgments with one or more machines consistently improved overall team performance. Our hope is that this simple and effective strategy for integrating the judgments of humans and machines will lead to productive collaborations.
{"title":"Confidence-weighted integration of human and machine judgments for superior decision-making.","authors":"Felipe Yáñez, Xiaoliang Luo, Omar Valerio Minero, Bradley C Love","doi":"10.1016/j.patter.2025.101423","DOIUrl":"https://doi.org/10.1016/j.patter.2025.101423","url":null,"abstract":"<p><p>Large language models (LLMs) can surpass humans in certain forecasting tasks. What role does this leave for humans in the overall decision process? One possibility is that humans, despite performing worse than LLMs, can still add value when teamed with them. A human and machine team can surpass each individual teammate when team members' confidence is well calibrated and team members diverge in which tasks they find difficult (i.e., calibration and diversity are needed). We simplified and extended a Bayesian approach to combining judgments using a logistic regression framework that integrates confidence-weighted judgments for any number of team members. Using this straightforward method, we demonstrated its effectiveness in both image classification and neuroscience forecasting tasks. Combining human judgments with one or more machines consistently improved overall team performance. Our hope is that this simple and effective strategy for integrating the judgments of humans and machines will lead to productive collaborations.</p>","PeriodicalId":36242,"journal":{"name":"Patterns","volume":"7 2","pages":"101423"},"PeriodicalIF":7.4,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921503/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147272221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}