{"title":"Towards a robust multi-view information bottleneck using Cauchy–Schwarz divergence","authors":"Qi Zhang, Mingfei Lu, Jingmin Xin, Badong Chen","doi":"10.1016/j.inffus.2025.102934","DOIUrl":null,"url":null,"abstract":"Efficiently preserving task-relevant information while removing noise and redundancy in multi-view data remains a core challenge. The information bottleneck principle offers an information-theoretic framework to compress data while retaining essential information for the task. However, estimating mutual information in high-dimensional spaces is computationally intractable. Commonly used variational methods introduce uncertainty and risk performance degradation. To overcome these limitations, we propose a robust deterministic multi-view information bottleneck framework that circumvents the need for variational inference or distributional assumptions. Specifically, we present a non-parametric mutual information estimation based on the Cauchy–Schwarz divergence, eliminating the need for auxiliary neural estimators and significantly simplifying the optimization of the information bottleneck. Leveraging this mutual information measure, we design a neural network framework that robustly compresses high-dimensional multi-view data into a low-dimensional representation, extracting task-relevant features that adhere to both sufficiency and minimality. Additionally, attention mechanisms are employed to fuse compact features across different views, capturing interdependencies and enhancing the integration of complementary information. This fusion process improves the robustness of the overall representation. Statistical analysis using the Nemenyi test shows statistically significant differences in performance between our method and existing algorithms, with a critical distance (CD = 1.856, <mml:math altimg=\"si1.svg\" display=\"inline\"><mml:mi>p</mml:mi></mml:math>-value <mml:math altimg=\"si2.svg\" display=\"inline\"><mml:mrow><mml:mo><</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>05</mml:mn></mml:mrow></mml:math>), demonstrating the superiority of our approach. Experimental results on synthetic data highlight the framework’s robustness in handling noise and redundancy, demonstrating its effectiveness in challenging environments. Validation on eight real-world datasets, including electroencephalography and Alzheimer’s neuroimaging data, confirms its superior performance, particularly with limited training samples. The implementation is available at <ce:inter-ref xlink:href=\"https://github.com/archy666/CSMVIB\" xlink:type=\"simple\">https://github.com/archy666/CSMVIB</ce:inter-ref>.","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"22 1","pages":""},"PeriodicalIF":14.7000,"publicationDate":"2025-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1016/j.inffus.2025.102934","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Efficiently preserving task-relevant information while removing noise and redundancy in multi-view data remains a core challenge. The information bottleneck principle offers an information-theoretic framework to compress data while retaining essential information for the task. However, estimating mutual information in high-dimensional spaces is computationally intractable. Commonly used variational methods introduce uncertainty and risk performance degradation. To overcome these limitations, we propose a robust deterministic multi-view information bottleneck framework that circumvents the need for variational inference or distributional assumptions. Specifically, we present a non-parametric mutual information estimation based on the Cauchy–Schwarz divergence, eliminating the need for auxiliary neural estimators and significantly simplifying the optimization of the information bottleneck. Leveraging this mutual information measure, we design a neural network framework that robustly compresses high-dimensional multi-view data into a low-dimensional representation, extracting task-relevant features that adhere to both sufficiency and minimality. Additionally, attention mechanisms are employed to fuse compact features across different views, capturing interdependencies and enhancing the integration of complementary information. This fusion process improves the robustness of the overall representation. Statistical analysis using the Nemenyi test shows statistically significant differences in performance between our method and existing algorithms, with a critical distance (CD = 1.856, p-value <0.05), demonstrating the superiority of our approach. Experimental results on synthetic data highlight the framework’s robustness in handling noise and redundancy, demonstrating its effectiveness in challenging environments. Validation on eight real-world datasets, including electroencephalography and Alzheimer’s neuroimaging data, confirms its superior performance, particularly with limited training samples. The implementation is available at https://github.com/archy666/CSMVIB.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.