Pub Date : 2026-02-05DOI: 10.1093/bioinformatics/btag059
Weiqiang Lin, Xinyi Xiao, Chuan Qiu, Hui Shen, Hongwen Deng
Motivation: Understanding spatial organization, intercellular interactions and regulatory networks within the spatial context of tissues is crucial for uncovering complex biological processes and disease mechanisms. Spatial transcriptomics technologies have revolutionized this field by enabling the spatially resolved profiling of gene expression. 10X Genomics Visium has emerged as the predominant spatial technology, but its low resolution and the complexity of integrating multimodal datasets present significant analytical challenges, particularly for researchers with limited computational and statistical expertise. Current spatial transcriptomics analysis platforms generally fall short of effectively integrating multi-modal data and maximizing the utility of spatial information-such as uncovering complex cellular spatial dependencies, multimodal gradient patterns and spatial co-expression of ligand-receptor pairs and regulatory networks related to disease or biological states-thereby limiting their ability to provide comprehensive end-to-end analytical workflows when analyzing 10X Genomics Visium data.
Results: To address these limitations, we developed transFusion, a novel, advanced web-based platform specializing in the most comprehensive and effective integration analysis of scRNA-seq and 10X Visium spatial transcriptomics data. transFusion offers 12 key functions, from basic visualization to advanced analyses, including intercellular dependency analysis, ligand-receptor co-expression identification and visualization, and spatial multimodal gradient variation patterns. Two case studies were used to demonstrate transFusion's capabilities in exploring tissue architecture, intercellular communication, dependency networks and multimodal gradient variation patterns with minimal computational skills and statistical expertise. transFusion provides a flexible and powerful framework for multi-modal data integration analysis.
Availability: transFusion is freely available at https://github.com/WQLin8/transFusion.
Supplementary information: Supplementary data are available at Bioinformatics online.
{"title":"transFusion: a Novel Comprehensive Platform for integration Analysis of Single-Cell and Spatial Transcriptomics.","authors":"Weiqiang Lin, Xinyi Xiao, Chuan Qiu, Hui Shen, Hongwen Deng","doi":"10.1093/bioinformatics/btag059","DOIUrl":"https://doi.org/10.1093/bioinformatics/btag059","url":null,"abstract":"<p><strong>Motivation: </strong>Understanding spatial organization, intercellular interactions and regulatory networks within the spatial context of tissues is crucial for uncovering complex biological processes and disease mechanisms. Spatial transcriptomics technologies have revolutionized this field by enabling the spatially resolved profiling of gene expression. 10X Genomics Visium has emerged as the predominant spatial technology, but its low resolution and the complexity of integrating multimodal datasets present significant analytical challenges, particularly for researchers with limited computational and statistical expertise. Current spatial transcriptomics analysis platforms generally fall short of effectively integrating multi-modal data and maximizing the utility of spatial information-such as uncovering complex cellular spatial dependencies, multimodal gradient patterns and spatial co-expression of ligand-receptor pairs and regulatory networks related to disease or biological states-thereby limiting their ability to provide comprehensive end-to-end analytical workflows when analyzing 10X Genomics Visium data.</p><p><strong>Results: </strong>To address these limitations, we developed transFusion, a novel, advanced web-based platform specializing in the most comprehensive and effective integration analysis of scRNA-seq and 10X Visium spatial transcriptomics data. transFusion offers 12 key functions, from basic visualization to advanced analyses, including intercellular dependency analysis, ligand-receptor co-expression identification and visualization, and spatial multimodal gradient variation patterns. Two case studies were used to demonstrate transFusion's capabilities in exploring tissue architecture, intercellular communication, dependency networks and multimodal gradient variation patterns with minimal computational skills and statistical expertise. transFusion provides a flexible and powerful framework for multi-modal data integration analysis.</p><p><strong>Availability: </strong>transFusion is freely available at https://github.com/WQLin8/transFusion.</p><p><strong>Supplementary information: </strong>Supplementary data are available at Bioinformatics online.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146127726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-04DOI: 10.1093/bioinformatics/btag057
Chengye Li, Hongwei Ma, Mingyang Ren
Motivation: Heterogeneity is a hallmark of both macroscopic complex diseases and microscopic single-cell distribution. Gaussian Graphical Models (GGM)-based heterogeneity analysis highlights its important role in capturing the essential characteristics of biological regulatory networks, but faces instability with scarce samples from rare subgroups. Transfer learning offers promise by leveraging auxiliary data, yet existing approaches rely on unrealistic overall similarity between domains, requiring the same subgroup number and similar parameters. Numerous biological problems call for local similarities, where only some subgroups share statistical structures.
Results: In this article, we propose LtransHeteroGGM, a novel local transfer learning framework for GGM-based heterogeneity analysis. It can achieve powerful subgroup-level local knowledge transfer between target and informative auxiliary domains, despite unknown subgroup structures and numbers, while mitigating the negative interference of non-informative domains. The effectiveness and robustness of the proposed approach are demonstrated through comprehensive numerical simulations and real-world T cell heterogeneity analysis.
Availability and implementation: The R implementation of LtransHeteroGGM is available at https://github.com/Ren-Mingyang/LtransHeteroGGM.
{"title":"LtransHeteroGGM: Local transfer learning for Gaussian graphical model-based heterogeneity analysis.","authors":"Chengye Li, Hongwei Ma, Mingyang Ren","doi":"10.1093/bioinformatics/btag057","DOIUrl":"https://doi.org/10.1093/bioinformatics/btag057","url":null,"abstract":"<p><strong>Motivation: </strong>Heterogeneity is a hallmark of both macroscopic complex diseases and microscopic single-cell distribution. Gaussian Graphical Models (GGM)-based heterogeneity analysis highlights its important role in capturing the essential characteristics of biological regulatory networks, but faces instability with scarce samples from rare subgroups. Transfer learning offers promise by leveraging auxiliary data, yet existing approaches rely on unrealistic overall similarity between domains, requiring the same subgroup number and similar parameters. Numerous biological problems call for local similarities, where only some subgroups share statistical structures.</p><p><strong>Results: </strong>In this article, we propose LtransHeteroGGM, a novel local transfer learning framework for GGM-based heterogeneity analysis. It can achieve powerful subgroup-level local knowledge transfer between target and informative auxiliary domains, despite unknown subgroup structures and numbers, while mitigating the negative interference of non-informative domains. The effectiveness and robustness of the proposed approach are demonstrated through comprehensive numerical simulations and real-world T cell heterogeneity analysis.</p><p><strong>Availability and implementation: </strong>The R implementation of LtransHeteroGGM is available at https://github.com/Ren-Mingyang/LtransHeteroGGM.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146121286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-03DOI: 10.1093/bioinformatics/btag058
Joan Segura, Ruben Sanchez-Garcia, Sebastian Bittrich, Yana Rose, Stephen K Burley, Jose M Duarte
Motivation: The rapid expansion of three-dimensional (3D) biomolecular structure information, driven by breakthroughs in artificial intelligence/deep learning (AI/DL)-based structure predictions, has created an urgent need for scalable and efficient structure similarity search methods. Traditional alignment-based approaches, such as structural superposition tools, are computationally expensive and challenging to scale with the vast number of available macromolecular structures.
Results: Herein, we present a scalable structure similarity search strategy designed to navigate extensive repositories of experimentally determined structures and computed structure models predicted using AI/DL methods. Our approach leverages protein language models and a deep neural network architecture to transform 3D structures into fixed-length vectors, enabling efficient large-scale comparisons. Although trained to predict TM-scores between single-domain structures, our model generalizes beyond the domain level, accurately identifying 3D similarity for full-length polypeptide chains and multimeric assemblies. By integrating vector databases, our method facilitates efficient large-scale structure retrieval, addressing the growing challenges posed by the expanding volume of 3D biostructure information.
Availability: Source code available at https://github.com/bioinsilico/rcsb-embedding-search.Source code DOI: https://doi.org/10.6084/m9.figshare.30546698.v1.Benchmark datasets DOI: https://doi.org/10.6084/m9.figshare.30546650.v1.Web server prototype available at: http://embedding-search.rcsb.org/.
Supplementary information: Supplementary data are available at Bioinformatics online.
{"title":"Multi-scale structural similarity embedding search across entire proteomes.","authors":"Joan Segura, Ruben Sanchez-Garcia, Sebastian Bittrich, Yana Rose, Stephen K Burley, Jose M Duarte","doi":"10.1093/bioinformatics/btag058","DOIUrl":"10.1093/bioinformatics/btag058","url":null,"abstract":"<p><strong>Motivation: </strong>The rapid expansion of three-dimensional (3D) biomolecular structure information, driven by breakthroughs in artificial intelligence/deep learning (AI/DL)-based structure predictions, has created an urgent need for scalable and efficient structure similarity search methods. Traditional alignment-based approaches, such as structural superposition tools, are computationally expensive and challenging to scale with the vast number of available macromolecular structures.</p><p><strong>Results: </strong>Herein, we present a scalable structure similarity search strategy designed to navigate extensive repositories of experimentally determined structures and computed structure models predicted using AI/DL methods. Our approach leverages protein language models and a deep neural network architecture to transform 3D structures into fixed-length vectors, enabling efficient large-scale comparisons. Although trained to predict TM-scores between single-domain structures, our model generalizes beyond the domain level, accurately identifying 3D similarity for full-length polypeptide chains and multimeric assemblies. By integrating vector databases, our method facilitates efficient large-scale structure retrieval, addressing the growing challenges posed by the expanding volume of 3D biostructure information.</p><p><strong>Availability: </strong>Source code available at https://github.com/bioinsilico/rcsb-embedding-search.Source code DOI: https://doi.org/10.6084/m9.figshare.30546698.v1.Benchmark datasets DOI: https://doi.org/10.6084/m9.figshare.30546650.v1.Web server prototype available at: http://embedding-search.rcsb.org/.</p><p><strong>Supplementary information: </strong>Supplementary data are available at Bioinformatics online.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146115223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-03DOI: 10.1093/bioinformatics/btag018
Artem Ivanov, Vladimir Popov, Maxim Morozov, Evgenii Olekhnovich, Vladimir Ulyantsev
Motivation: Microbial communities consist of thousands of microorganisms and viruses and have a tight connection with an environment, such as gut microbiota modulation of host body metabolism. However, the direct relationship between the presence of certain microorganism and the host state often remains unknown. Toolkits using reference-based approaches are limited to microbes present in databases. Reference-free methods often require enormous resources for metagenomic assembly or results in many poorly interpretable features based on k-mers.
Results: Here we present MetaFX-an open-source library for feature extraction from whole-genome metagenomic sequencing data and classification of groups of samples. Using a large volume of metagenomic samples deposited in databases, MetaFX compares samples grouped by metadata criteria (e.g. disease, treatment, etc.) and constructs genomic features distinct for certain types of communities. Features constructed based on statistical k-mer analysis and de Bruijn graphs partition. Those features are used in machine learning models for classification of novel samples. Extracted features can be visualized on de Bruijn graphs and annotated for providing biological insights. We demonstrate the utility of MetaFX by building classification models for 590 human gut samples with inflammatory bowel disease. Our results outperform the previous research disease prediction accuracy up to 17%, and improves classification results compared to taxonomic analysis by 9±10% on average.
Availability and implementation: MetaFX is a feature extraction toolkit applicable for metagenomic datasets analysis and samples classification. The source code, test data, and relevant information for MetaFX are freely accessible at https://github.com/ctlab/metafx under the MIT License. Alternatively, MetaFX can be obtained via http://doi.org/10.5281/zenodo.16949369.
{"title":"MetaFX: feature extraction from whole-genome metagenomic sequencing data.","authors":"Artem Ivanov, Vladimir Popov, Maxim Morozov, Evgenii Olekhnovich, Vladimir Ulyantsev","doi":"10.1093/bioinformatics/btag018","DOIUrl":"10.1093/bioinformatics/btag018","url":null,"abstract":"<p><strong>Motivation: </strong>Microbial communities consist of thousands of microorganisms and viruses and have a tight connection with an environment, such as gut microbiota modulation of host body metabolism. However, the direct relationship between the presence of certain microorganism and the host state often remains unknown. Toolkits using reference-based approaches are limited to microbes present in databases. Reference-free methods often require enormous resources for metagenomic assembly or results in many poorly interpretable features based on k-mers.</p><p><strong>Results: </strong>Here we present MetaFX-an open-source library for feature extraction from whole-genome metagenomic sequencing data and classification of groups of samples. Using a large volume of metagenomic samples deposited in databases, MetaFX compares samples grouped by metadata criteria (e.g. disease, treatment, etc.) and constructs genomic features distinct for certain types of communities. Features constructed based on statistical k-mer analysis and de Bruijn graphs partition. Those features are used in machine learning models for classification of novel samples. Extracted features can be visualized on de Bruijn graphs and annotated for providing biological insights. We demonstrate the utility of MetaFX by building classification models for 590 human gut samples with inflammatory bowel disease. Our results outperform the previous research disease prediction accuracy up to 17%, and improves classification results compared to taxonomic analysis by 9±10% on average.</p><p><strong>Availability and implementation: </strong>MetaFX is a feature extraction toolkit applicable for metagenomic datasets analysis and samples classification. The source code, test data, and relevant information for MetaFX are freely accessible at https://github.com/ctlab/metafx under the MIT License. Alternatively, MetaFX can be obtained via http://doi.org/10.5281/zenodo.16949369.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12891910/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146013864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-03DOI: 10.1093/bioinformatics/btag035
David Kouřil, Trevor Manz, Tereza Clarence, Nils Gehlenborg
Summary: Uchimata is a toolkit for visualization of 3D structures of genomes. It consists of two packages: a Javascript library facilitating the rendering of 3D models of genomes, and a Python widget for visualization in Jupyter Notebooks. Main features include an expressive way to specify visual encodings, and filtering of 3D genome structures based on genomic semantics and spatial aspects. Uchimata is designed to be highly integratable with biological tooling available in Python.
Availability and implementation: Uchimata is released under the MIT License. The Javascript library is available on NPM, while the widget is available as a Python package hosted on PyPI. The source code for both is available publicly on Github (https://github.com/hms-dbmi/uchimata and https://github.com/hms-dbmi/uchimata-py) and Zenodo (https://doi.org/10.5281/zenodo.17831959 and https://doi.org/10.5281/zenodo.17832045). The documentation with examples is hosted at https://hms-dbmi.github.io/uchimata/.
{"title":"Uchimata: a toolkit for visualization of 3D genome structures on the web and in computational notebooks.","authors":"David Kouřil, Trevor Manz, Tereza Clarence, Nils Gehlenborg","doi":"10.1093/bioinformatics/btag035","DOIUrl":"10.1093/bioinformatics/btag035","url":null,"abstract":"<p><strong>Summary: </strong>Uchimata is a toolkit for visualization of 3D structures of genomes. It consists of two packages: a Javascript library facilitating the rendering of 3D models of genomes, and a Python widget for visualization in Jupyter Notebooks. Main features include an expressive way to specify visual encodings, and filtering of 3D genome structures based on genomic semantics and spatial aspects. Uchimata is designed to be highly integratable with biological tooling available in Python.</p><p><strong>Availability and implementation: </strong>Uchimata is released under the MIT License. The Javascript library is available on NPM, while the widget is available as a Python package hosted on PyPI. The source code for both is available publicly on Github (https://github.com/hms-dbmi/uchimata and https://github.com/hms-dbmi/uchimata-py) and Zenodo (https://doi.org/10.5281/zenodo.17831959 and https://doi.org/10.5281/zenodo.17832045). The documentation with examples is hosted at https://hms-dbmi.github.io/uchimata/.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12904833/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146020897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"ProteoGyver: a fast, user-friendly tool for routine QC and analysis of MS-based proteomics data.","authors":"Kari Salokas, Salla Keskitalo, Markku Varjosalo","doi":"10.1093/bioinformatics/btag050","DOIUrl":"10.1093/bioinformatics/btag050","url":null,"abstract":"<p><strong>Availability and implementation: </strong>PG image and source code are available in github and dockerhub under LGPL-2.1.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12910381/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146088319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-03DOI: 10.1093/bioinformatics/btag024
Soo Bin Kwon, Jason Ernst
Motivation: Identifying pairwise associations between genomic loci is an important challenge for which large and diverse collections of epigenomic and transcription factor (TF) binding data can potentially be informative.
Results: We developed Learning Evidence of Pairwise Association from Epigenomic and TF binding data (LEPAE). LEPAE uses neural networks to quantify evidence of association for pairs of genomic windows from large-scale epigenomic and TF binding data along with distance information. We applied LEPAE using thousands of human datasets. We show using additional data that LEPAE captures biologically meaningful pairwise relationships between genomic loci, and we expect LEPAE scores to be a resource.
Availability and implementation: The LEPAE scores and the software are available at https://github.com/ernstlab/LEPAE.
{"title":"Learning a pairwise epigenomic and transcription factor binding association score across the human genome.","authors":"Soo Bin Kwon, Jason Ernst","doi":"10.1093/bioinformatics/btag024","DOIUrl":"10.1093/bioinformatics/btag024","url":null,"abstract":"<p><strong>Motivation: </strong>Identifying pairwise associations between genomic loci is an important challenge for which large and diverse collections of epigenomic and transcription factor (TF) binding data can potentially be informative.</p><p><strong>Results: </strong>We developed Learning Evidence of Pairwise Association from Epigenomic and TF binding data (LEPAE). LEPAE uses neural networks to quantify evidence of association for pairs of genomic windows from large-scale epigenomic and TF binding data along with distance information. We applied LEPAE using thousands of human datasets. We show using additional data that LEPAE captures biologically meaningful pairwise relationships between genomic loci, and we expect LEPAE scores to be a resource.</p><p><strong>Availability and implementation: </strong>The LEPAE scores and the software are available at https://github.com/ernstlab/LEPAE.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12910503/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146013837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-03DOI: 10.1093/bioinformatics/btaf679
Nguyen Khoa Tran, My Ky Huynh, Alexander D Kotman, Martin Jürgens, Thomas Kurz, Sascha Dietrich, Gunnar W Klau, Nan Qin
Motivation: Live-cell imaging-based drug screening increases the likelihood of identifying effective and safe drugs by providing dynamic, high-content, and physiologically relevant data. As a result, it improves the success rate of drug development and facilitates the translation of benchside discoveries to bedside applications. Despite these advantages, no comprehensive metrics currently exist to evaluate dose-time-dependent drug responses. To address this gap, we established a systematic framework to assess drug effects across a range of concentrations and exposure durations simultaneously. This metric enables more accurate evaluation of drug responses measured by live-cell imaging.
Results: We employed treatment concentrations ranging from 0 to 10 μM and performed live-cell imaging-based measurements over a 120-h incubation period. To analyze the experimental data, we developed VUScope, a new mathematical model combining the 4-parameter logistic curve and a logistic function to characterize dose-time-dependent responses. This enabled us to calculate the Growth Rate Inhibition Volume Under the dose-time-response Surface (GRIVUS), which serves as a critical metric for assessing dynamic drug responses. Furthermore, our mathematical model allowed us to predict long-term treatment responses based on short-term drug responses. We validated the predictive capabilities of our model using independent datasets and observed that VUScope enhances prediction accuracy and offers deeper insights into drug effects than previously possible. By integrating VUScope into high-throughput drug screening platforms, we can further improve the efficacy of drug development and treatment selection.
Availability and implementation: We have made VUScope more accessible to users conducting pharmacological studies by uploading a detailed description, example datasets, and the source code to vuscope.albi.hhu.de, https://github.com/AlBi-HHU/VUScope, and https://doi.org/10.5281/zenodo.17610533.
{"title":"VUScope: a mathematical model for evaluating image-based drug response measurements and predicting long-term incubation outcomes.","authors":"Nguyen Khoa Tran, My Ky Huynh, Alexander D Kotman, Martin Jürgens, Thomas Kurz, Sascha Dietrich, Gunnar W Klau, Nan Qin","doi":"10.1093/bioinformatics/btaf679","DOIUrl":"10.1093/bioinformatics/btaf679","url":null,"abstract":"<p><strong>Motivation: </strong>Live-cell imaging-based drug screening increases the likelihood of identifying effective and safe drugs by providing dynamic, high-content, and physiologically relevant data. As a result, it improves the success rate of drug development and facilitates the translation of benchside discoveries to bedside applications. Despite these advantages, no comprehensive metrics currently exist to evaluate dose-time-dependent drug responses. To address this gap, we established a systematic framework to assess drug effects across a range of concentrations and exposure durations simultaneously. This metric enables more accurate evaluation of drug responses measured by live-cell imaging.</p><p><strong>Results: </strong>We employed treatment concentrations ranging from 0 to 10 μM and performed live-cell imaging-based measurements over a 120-h incubation period. To analyze the experimental data, we developed VUScope, a new mathematical model combining the 4-parameter logistic curve and a logistic function to characterize dose-time-dependent responses. This enabled us to calculate the Growth Rate Inhibition Volume Under the dose-time-response Surface (GRIVUS), which serves as a critical metric for assessing dynamic drug responses. Furthermore, our mathematical model allowed us to predict long-term treatment responses based on short-term drug responses. We validated the predictive capabilities of our model using independent datasets and observed that VUScope enhances prediction accuracy and offers deeper insights into drug effects than previously possible. By integrating VUScope into high-throughput drug screening platforms, we can further improve the efficacy of drug development and treatment selection.</p><p><strong>Availability and implementation: </strong>We have made VUScope more accessible to users conducting pharmacological studies by uploading a detailed description, example datasets, and the source code to vuscope.albi.hhu.de, https://github.com/AlBi-HHU/VUScope, and https://doi.org/10.5281/zenodo.17610533.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12904834/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146121336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-03DOI: 10.1093/bioinformatics/btag031
Favour James, Dexter Pratt, Christopher Churas, Augustin Luna
Motivation: Knowledge graphs (KGs) are powerful tools for structuring and analyzing biological information due to their ability to represent data and improve queries across heterogeneous datasets. However, constructing KGs from unstructured literature remains challenging due to the cost and expertise required for manual curation. Prior works have explored text-mining techniques to automate this process, but have limitations that impact their ability to capture complex relationships fully. Traditional text-mining methods struggle with understanding context across sentences. Additionally, these methods lack expert-level background knowledge, making it difficult to infer relationships that require awareness of concepts indirectly described in the text. Large Language Models (LLMs) present an opportunity to overcome these challenges. LLMs are trained on diverse literature, equipping them with contextual knowledge that enables more accurate information extraction.
Results: We present textToKnowledgeGraph, an artificial intelligence tool using LLMs to extract interactions from individual publications directly in Biological Expression Language (BEL). BEL was chosen for its compact, detailed representation of biological relationships, enabling structured, computationally accessible encoding. This work makes several contributions. (i) Development of the open-source Python textToKnowledgeGraph package (pypi.org/project/texttoknowledgegraph) for BEL extraction from scientific articles, usable from the command line and within other projects, (ii) an interactive application within Cytoscape Web to simplify extraction and exploration, (iii) a dataset of extractions that have been both computationally and manually reviewed to support future fine-tuning efforts.
Availability and implementation: https://github.com/ndexbio/llm-text-to-knowledge-graph.
{"title":"textToKnowledgeGraph: generation of molecular interaction knowledge graphs using large language models for exploration in Cytoscape.","authors":"Favour James, Dexter Pratt, Christopher Churas, Augustin Luna","doi":"10.1093/bioinformatics/btag031","DOIUrl":"10.1093/bioinformatics/btag031","url":null,"abstract":"<p><strong>Motivation: </strong>Knowledge graphs (KGs) are powerful tools for structuring and analyzing biological information due to their ability to represent data and improve queries across heterogeneous datasets. However, constructing KGs from unstructured literature remains challenging due to the cost and expertise required for manual curation. Prior works have explored text-mining techniques to automate this process, but have limitations that impact their ability to capture complex relationships fully. Traditional text-mining methods struggle with understanding context across sentences. Additionally, these methods lack expert-level background knowledge, making it difficult to infer relationships that require awareness of concepts indirectly described in the text. Large Language Models (LLMs) present an opportunity to overcome these challenges. LLMs are trained on diverse literature, equipping them with contextual knowledge that enables more accurate information extraction.</p><p><strong>Results: </strong>We present textToKnowledgeGraph, an artificial intelligence tool using LLMs to extract interactions from individual publications directly in Biological Expression Language (BEL). BEL was chosen for its compact, detailed representation of biological relationships, enabling structured, computationally accessible encoding. This work makes several contributions. (i) Development of the open-source Python textToKnowledgeGraph package (pypi.org/project/texttoknowledgegraph) for BEL extraction from scientific articles, usable from the command line and within other projects, (ii) an interactive application within Cytoscape Web to simplify extraction and exploration, (iii) a dataset of extractions that have been both computationally and manually reviewed to support future fine-tuning efforts.</p><p><strong>Availability and implementation: </strong>https://github.com/ndexbio/llm-text-to-knowledge-graph.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146004727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Motivation: Dendritic spines, postsynaptic structures characterized by their complex shapes, provide the essential structural foundation for synaptic function. Their shape is dynamic, undergoing alterations in various conditions, notably during neurodegenerative disorders like Alzheimer's disease. The dramatically increasing prevalence of such diseases highlights an urgent need for effective treatments. A key strategy in developing these treatments involves evaluating how dendritic spine morphology responds to potential therapeutic compounds. Although a link between spine shape and function is recognized, its precise nature is still not fully elucidated. Consequently, advancing our understanding of dendritic spines in both health and disease necessitates the urgent development of more effective methods for assessing their morphology.
Results: This study introduces qualitatively new 3D dendritic shape descriptors based on spherical harmonics and Zernike moments and proposes a bases on them clustering approach for grouping dendritic spines with similar shapes applied to 3D polygonal spines meshes acquired from Z-stack dendrite images. By integrating these methods, we achieve improved differentiation between normal and pathological spines represented by the Alzheimer's disease in vitro model, offering a more precise representation of morphological diversity. Additionally, the proposed spherical harmonics approach enables dendritic spine reconstruction from vector-based shape representations, providing a novel tool for studying structural changes associated with neurodegeneration and possibilities for synthetic dendritic spines dataset generation.
Availability and implementation: The software used for experiments is public and available at https://github.com/Biomed-imaging-lab/SpineTool with the DOI: 10.5281/zenodo.17359066. Descriptors codebase is available at https://github.com/Biomed-imaging-lab/Spine-Shape-Descriptors with the DOI: 10.5281/zenodo.17302859.
{"title":"3D dendritic spines shape descriptors for efficient classification and morphology analysis in control and Alzheimer's disease modeling neurons.","authors":"Daria Smirnova, Anita Ustinova, Viacheslav Chukanov, Ekaterina Pchitskaya","doi":"10.1093/bioinformatics/btag025","DOIUrl":"10.1093/bioinformatics/btag025","url":null,"abstract":"<p><strong>Motivation: </strong>Dendritic spines, postsynaptic structures characterized by their complex shapes, provide the essential structural foundation for synaptic function. Their shape is dynamic, undergoing alterations in various conditions, notably during neurodegenerative disorders like Alzheimer's disease. The dramatically increasing prevalence of such diseases highlights an urgent need for effective treatments. A key strategy in developing these treatments involves evaluating how dendritic spine morphology responds to potential therapeutic compounds. Although a link between spine shape and function is recognized, its precise nature is still not fully elucidated. Consequently, advancing our understanding of dendritic spines in both health and disease necessitates the urgent development of more effective methods for assessing their morphology.</p><p><strong>Results: </strong>This study introduces qualitatively new 3D dendritic shape descriptors based on spherical harmonics and Zernike moments and proposes a bases on them clustering approach for grouping dendritic spines with similar shapes applied to 3D polygonal spines meshes acquired from Z-stack dendrite images. By integrating these methods, we achieve improved differentiation between normal and pathological spines represented by the Alzheimer's disease in vitro model, offering a more precise representation of morphological diversity. Additionally, the proposed spherical harmonics approach enables dendritic spine reconstruction from vector-based shape representations, providing a novel tool for studying structural changes associated with neurodegeneration and possibilities for synthetic dendritic spines dataset generation.</p><p><strong>Availability and implementation: </strong>The software used for experiments is public and available at https://github.com/Biomed-imaging-lab/SpineTool with the DOI: 10.5281/zenodo.17359066. Descriptors codebase is available at https://github.com/Biomed-imaging-lab/Spine-Shape-Descriptors with the DOI: 10.5281/zenodo.17302859.</p>","PeriodicalId":93899,"journal":{"name":"Bioinformatics (Oxford, England)","volume":" ","pages":""},"PeriodicalIF":5.4,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12891915/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146013884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}