首页 > 最新文献

Information Processing & Management最新文献

英文 中文
DS_HURNSP: An effective method for mining high utility repeated negative sequential patterns from data streams DS_HURNSP:从数据流中挖掘高效用重复负序列模式的有效方法
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-22 DOI: 10.1016/j.ipm.2026.104637
Xiangjun Dong , Yicong Zhen , Ping Qiu , Jing Chi , Lei Guo , Wenpeng Lu , Long Zhao , Yongshun Gong , Yuhai Zhao
Mining high utility repeated negative sequential patterns (HURNSPs) from data streams is an important method for data stream analysis. However, the existing methods in this topic don’t consider negative events and repeated events, which results in weak decision-making effectiveness. So in this paper, we propose an effective algorithm DS_HURNSP for mining HURNSPs from data streams based on a sliding window model. First, we propose an effective utility-list prefix tree structure to store high utility repeated positive sequential patterns (HURPSPs). Second, we construct a utility mapping set based on a hash-table structure to enable rapid querying of HURPSPs information. Finally, we propose a two-stage computation method to compute the utility of high utility repeated negative sequential candidates (HURNSCs) by mapping them to the set of HURPSPs, avoiding rescanning database. Extensive experiments on six datasets show that the DS_HURNSP algorithm generates tens to thousands of times as many HURNSPs as the baseline method, and reduces the average runtime by more than half.
从数据流中挖掘高效用重复负序列模式(HURNSPs)是数据流分析的重要方法。然而,本课题现有的方法没有考虑负面事件和重复事件,导致决策有效性较弱。因此,本文提出了一种基于滑动窗口模型的有效算法DS_HURNSP从数据流中挖掘hurnsp。首先,我们提出了一种有效的实用列表前缀树结构来存储高实用重复正序列模式(hurpsp)。其次,基于散列表结构构造实用映射集,实现对hurpsp信息的快速查询。最后,我们提出了一种两阶段的计算方法,通过将高效用重复负序候选(HURNSCs)映射到hurpsp集合来计算它们的效用,从而避免重新扫描数据库。在6个数据集上的大量实验表明,DS_HURNSP算法产生的hurnsp数是基线方法的几十到几千倍,平均运行时间减少了一半以上。
{"title":"DS_HURNSP: An effective method for mining high utility repeated negative sequential patterns from data streams","authors":"Xiangjun Dong ,&nbsp;Yicong Zhen ,&nbsp;Ping Qiu ,&nbsp;Jing Chi ,&nbsp;Lei Guo ,&nbsp;Wenpeng Lu ,&nbsp;Long Zhao ,&nbsp;Yongshun Gong ,&nbsp;Yuhai Zhao","doi":"10.1016/j.ipm.2026.104637","DOIUrl":"10.1016/j.ipm.2026.104637","url":null,"abstract":"<div><div>Mining high utility repeated negative sequential patterns (HURNSPs) from data streams is an important method for data stream analysis. However, the existing methods in this topic don’t consider negative events and repeated events, which results in weak decision-making effectiveness. So in this paper, we propose an effective algorithm DS_HURNSP for mining HURNSPs from data streams based on a sliding window model. First, we propose an effective utility-list prefix tree structure to store high utility repeated positive sequential patterns (HURPSPs). Second, we construct a utility mapping set based on a hash-table structure to enable rapid querying of HURPSPs information. Finally, we propose a two-stage computation method to compute the utility of high utility repeated negative sequential candidates (HURNSCs) by mapping them to the set of HURPSPs, avoiding rescanning database. Extensive experiments on six datasets show that the DS_HURNSP algorithm generates tens to thousands of times as many HURNSPs as the baseline method, and reduces the average runtime by more than half.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104637"},"PeriodicalIF":6.9,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SEGA: Selective cross-lingual representation via sparse guided attention for low-resource multilingual named entity recognition SEGA:基于稀疏引导注意力的低资源多语言命名实体识别的选择性跨语言表示
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-22 DOI: 10.1016/j.ipm.2026.104627
Paerhati Tulajiang , Jinzhong Ning , Yuanyuan Sun , Liang Yang , Yuanyu Zhang , Kelaiti Xiao , Zhixing Lu , Yijia Zhang , Hongfei Lin
Multilingual named entity recognition (NER) is especially challenging in low-resource and typologically diverse languages, where translation drift, morphological variation, and noisy alignments degrade performance. Existing encoder-based methods often rely on dense attention or uniform alignment, which tends to propagate irrelevant signals across languages. We present SEGA, a lightweight and typology-aware framework that incorporates sparse guided attention to select auxiliary signals, alongside a weighted fusion layer that balances representations between cross-lingual and monolingual contexts. Unlike prior approaches, SEGA requires no parallel corpora and supports fully monolingual inference. We evaluate SEGA on six multilingual NER benchmarks spanning over 60 languages, including CoNLL, WikiANN, MasakhaNER 2.0, XTREME-40, WikiNEuRal, and MultiNERD. SEGA achieves new state-of-the-art results on five datasets, with absolute gains of up to +24.2 F1 over strong encoder baselines, and outperforming prompt-based large language models by up to +18.9 F1 in low-resource scenarios. Efficiency analyses show that SEGA adds only  ∼ 30M parameters beyond a standard dual encoder, making it lightweight and deployable on a single GPU. Comprehensive ablation, visualization, and error analyses confirm that SEGA is robust to alignment noise, morphological complexity, and boundary ambiguity, offering a practical and scalable solution for real-world multilingual NER.
多语言命名实体识别(NER)在低资源和类型多样化的语言中尤其具有挑战性,其中翻译漂移、形态变化和噪声对齐会降低性能。现有的基于编码器的方法通常依赖于密集关注或均匀对齐,这倾向于跨语言传播不相关的信号。我们提出了一种轻量级的类型感知框架SEGA,它结合了稀疏引导注意力来选择辅助信号,以及一个加权融合层,以平衡跨语言和单语言上下文之间的表示。与之前的方法不同,SEGA不需要并行语料库,并且完全支持单语言推理。我们对世嘉进行了六种多语言NER基准测试,涵盖60多种语言,包括CoNLL、WikiANN、MasakhaNER 2.0、XTREME-40、WikiNEuRal和multierd。世嘉在五个数据集上实现了新的最先进的结果,在强大的编码器基线上的绝对增益高达+24.2 F1,在低资源场景下优于基于提示的大型语言模型高达+18.9 F1。效率分析表明,世嘉在标准双编码器之外仅增加了 ~ 30M参数,使其重量轻,可部署在单个GPU上。综合消融、可视化和误差分析证实,SEGA对对齐噪声、形态复杂性和边界模糊具有鲁棒性,为现实世界的多语言NER提供了实用且可扩展的解决方案。
{"title":"SEGA: Selective cross-lingual representation via sparse guided attention for low-resource multilingual named entity recognition","authors":"Paerhati Tulajiang ,&nbsp;Jinzhong Ning ,&nbsp;Yuanyuan Sun ,&nbsp;Liang Yang ,&nbsp;Yuanyu Zhang ,&nbsp;Kelaiti Xiao ,&nbsp;Zhixing Lu ,&nbsp;Yijia Zhang ,&nbsp;Hongfei Lin","doi":"10.1016/j.ipm.2026.104627","DOIUrl":"10.1016/j.ipm.2026.104627","url":null,"abstract":"<div><div>Multilingual named entity recognition (NER) is especially challenging in low-resource and typologically diverse languages, where translation drift, morphological variation, and noisy alignments degrade performance. Existing encoder-based methods often rely on dense attention or uniform alignment, which tends to propagate irrelevant signals across languages. We present SEGA, a lightweight and typology-aware framework that incorporates sparse guided attention to select auxiliary signals, alongside a weighted fusion layer that balances representations between cross-lingual and monolingual contexts. Unlike prior approaches, SEGA requires no parallel corpora and supports fully monolingual inference. We evaluate SEGA on six multilingual NER benchmarks spanning over 60 languages, including CoNLL, WikiANN, MasakhaNER 2.0, XTREME-40, WikiNEuRal, and MultiNERD. SEGA achieves new state-of-the-art results on five datasets, with absolute gains of up to +24.2 F1 over strong encoder baselines, and outperforming prompt-based large language models by up to +18.9 F1 in low-resource scenarios. Efficiency analyses show that SEGA adds only  ∼ 30M parameters beyond a standard dual encoder, making it lightweight and deployable on a single GPU. Comprehensive ablation, visualization, and error analyses confirm that SEGA is robust to alignment noise, morphological complexity, and boundary ambiguity, offering a practical and scalable solution for real-world multilingual NER.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104627"},"PeriodicalIF":6.9,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TaNSP: An efficient target pattern mining algorithm based on negative sequential pattern TaNSP:一种基于负序模式的高效目标模式挖掘算法
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-22 DOI: 10.1016/j.ipm.2026.104643
Xiaowen Cui , Xue Dong , Ping Qiu , Chuanhou Sun , Yuhai Zhao , Wenpeng Lu , Xiangjun Dong
Target pattern mining (TPM) aims to return sets of target patterns related to a user-queried target sequence. However, existing TPM research is confined to positive sequential patterns, overlooking negative sequential patterns, which results in limited decision support capabilities. Moreover, introducing negative sequential patterns faces challenges of low mining efficiency and limited pruning techniques. To address these issues, we propose an efficient target pattern mining algorithm based on negative sequential pattern, called TaNSP, to achieve TPM for negative sequence as the user-queried target sequence and output negative sequential patterns containing the target query sequence, while also supporting positive sequential patterns. Specifically, we propose a pruning strategy based on a triple bitmap to guide pattern generation and improve mining efficiency. Then, we propose a pruning strategy to address the limitations of pruning techniques when the negative sequence is the target query sequence. The experimental results on six datasets demonstrate that, compared to the baseline method, TaNSP can increase operational efficiency by more than twice, demonstrating excellent scalability and practicality.
目标模式挖掘(TPM)旨在返回与用户查询的目标序列相关的目标模式集。然而,现有的TPM研究仅限于积极的顺序模式,忽视了消极的顺序模式,导致决策支持能力有限。此外,引入负序模式还面临着挖掘效率低和修剪技术有限的挑战。为了解决这些问题,我们提出了一种高效的基于负序列模式的目标模式挖掘算法TaNSP,将负序列作为用户查询的目标序列实现TPM,输出包含目标查询序列的负序列模式,同时也支持正序列模式。具体来说,我们提出了一种基于三重位图的剪枝策略来指导模式生成,提高挖掘效率。然后,我们提出了一种剪枝策略,以解决当负序列是目标查询序列时剪枝技术的局限性。在6个数据集上的实验结果表明,与基线方法相比,TaNSP的运算效率提高了2倍以上,具有良好的可扩展性和实用性。
{"title":"TaNSP: An efficient target pattern mining algorithm based on negative sequential pattern","authors":"Xiaowen Cui ,&nbsp;Xue Dong ,&nbsp;Ping Qiu ,&nbsp;Chuanhou Sun ,&nbsp;Yuhai Zhao ,&nbsp;Wenpeng Lu ,&nbsp;Xiangjun Dong","doi":"10.1016/j.ipm.2026.104643","DOIUrl":"10.1016/j.ipm.2026.104643","url":null,"abstract":"<div><div>Target pattern mining (TPM) aims to return sets of target patterns related to a user-queried target sequence. However, existing TPM research is confined to positive sequential patterns, overlooking negative sequential patterns, which results in limited decision support capabilities. Moreover, introducing negative sequential patterns faces challenges of low mining efficiency and limited pruning techniques. To address these issues, we propose an efficient <u>ta</u>rget pattern mining algorithm based on <u>n</u>egative <u>s</u>equential <u>p</u>attern, called TaNSP, to achieve TPM for negative sequence as the user-queried target sequence and output negative sequential patterns containing the target query sequence, while also supporting positive sequential patterns. Specifically, we propose a pruning strategy based on a triple bitmap to guide pattern generation and improve mining efficiency. Then, we propose a pruning strategy to address the limitations of pruning techniques when the negative sequence is the target query sequence. The experimental results on six datasets demonstrate that, compared to the baseline method, TaNSP can increase operational efficiency by more than twice, demonstrating excellent scalability and practicality.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104643"},"PeriodicalIF":6.9,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From personalized learning to explainable prediction: A data-driven framework for patient no-shows 从个性化学习到可解释的预测:患者缺席的数据驱动框架
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-21 DOI: 10.1016/j.ipm.2026.104633
Wenbo Zhang , Xi Chen , Mingwei Wang , Rui Hou , Ning Li
The prediction of patient no-shows has emerged as a critical research topic in healthcare service management. To address the challenges of behavioral heterogeneity and factor diversity in patient no-shows, this study proposes a data-driven three-stage decision analytics method integrating personalized preference learning and explainable prediction. First, to capture patient no-show behavioral heterogeneity, we applied community detection algorithms to identify distinct patient subgroups, enabling subgroup-specific exploration of no-show patterns. Second, within each subgroup, we employ a preference learning model to quantify the relative importance of the influencing factors tailored to each group’s characteristics. Finally, we developed an XGBoost predictive model based on SHAP to achieve highly accurate and interpretable no-show probability predictions. Experimental results based on 2431 real outpatient appointment records from the Department of Ophthalmology at General Medical Harbin 242 Hospital showed that air quality emerged as a critical factor influencing patient attendance, with poorer air quality significantly increasing the likelihood of no-shows. In addition, the proposed PL-XGBoost-SHAP model achieved an average accuracy of 90.08%, precision of 92.36%, recall of 95.65% and F1 score of 93.96% across the four subgroups. These results showed an accuracy improvement of 2–12% compared with the five baseline models, demonstrating the scientific validity and feasibility of the proposed approach. Sensitivity analyses and statistical tests confirmed the robustness and generalizability of the proposed method. Consequently, this study offers significant practical implications for healthcare providers, enabling the design of personalized appointment reminders and effective allocate resources.
患者缺席预测已成为医疗服务管理领域的一个重要研究课题。为了解决患者失诊的行为异质性和因素多样性的挑战,本研究提出了一种数据驱动的三阶段决策分析方法,该方法将个性化偏好学习和可解释预测相结合。首先,为了捕捉患者的失诊行为异质性,我们应用社区检测算法来识别不同的患者亚组,从而实现针对亚组的失诊模式探索。其次,在每个子群体中,我们采用偏好学习模型来量化针对每个群体特征的影响因素的相对重要性。最后,我们开发了一个基于SHAP的XGBoost预测模型,以实现高精度和可解释的缺席概率预测。基于哈尔滨242总医院眼科2431份真实门诊预约记录的实验结果表明,空气质量成为影响患者出诊的关键因素,空气质量越差,患者失诊的可能性显著增加。此外,所提出的PL-XGBoost-SHAP模型在4个亚组中的平均准确率为90.08%,精密度为92.36%,召回率为95.65%,F1得分为93.96%。结果表明,与5种基线模型相比,该方法的精度提高了2-12%,证明了该方法的科学有效性和可行性。敏感性分析和统计检验证实了该方法的稳健性和可推广性。因此,本研究为医疗保健提供者提供了重要的实际意义,使个性化预约提醒的设计和有效分配资源成为可能。
{"title":"From personalized learning to explainable prediction: A data-driven framework for patient no-shows","authors":"Wenbo Zhang ,&nbsp;Xi Chen ,&nbsp;Mingwei Wang ,&nbsp;Rui Hou ,&nbsp;Ning Li","doi":"10.1016/j.ipm.2026.104633","DOIUrl":"10.1016/j.ipm.2026.104633","url":null,"abstract":"<div><div>The prediction of patient no-shows has emerged as a critical research topic in healthcare service management. To address the challenges of behavioral heterogeneity and factor diversity in patient no-shows, this study proposes a data-driven three-stage decision analytics method integrating personalized preference learning and explainable prediction. First, to capture patient no-show behavioral heterogeneity, we applied community detection algorithms to identify distinct patient subgroups, enabling subgroup-specific exploration of no-show patterns. Second, within each subgroup, we employ a preference learning model to quantify the relative importance of the influencing factors tailored to each group’s characteristics. Finally, we developed an XGBoost predictive model based on SHAP to achieve highly accurate and interpretable no-show probability predictions. Experimental results based on 2431 real outpatient appointment records from the Department of Ophthalmology at General Medical Harbin 242 Hospital showed that air quality emerged as a critical factor influencing patient attendance, with poorer air quality significantly increasing the likelihood of no-shows. In addition, the proposed PL-XGBoost-SHAP model achieved an average accuracy of 90.08%, precision of 92.36%, recall of 95.65% and F1 score of 93.96% across the four subgroups. These results showed an accuracy improvement of 2–12% compared with the five baseline models, demonstrating the scientific validity and feasibility of the proposed approach. Sensitivity analyses and statistical tests confirmed the robustness and generalizability of the proposed method. Consequently, this study offers significant practical implications for healthcare providers, enabling the design of personalized appointment reminders and effective allocate resources.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104633"},"PeriodicalIF":6.9,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CausalLog: Log parsing using LLMs with causal intervention for bias mitigation CausalLog:使用带有因果干预的llm进行日志解析,以减轻偏差
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-21 DOI: 10.1016/j.ipm.2026.104609
Yuan Tian, Shi Ying, Tiangang Li
Log parsing transforms unstructured log messages into structured formats, serving as a critical step for various log analysis tasks. Large language models (LLMs) have recently shown strong performance in this task. However, they tend to rely on their experiential knowledge as shortcuts, introducing bias and reducing parsing accuracy. To address this issue, we propose CausalLog, a lightweight and flexible debiasing framework for log parsing. CausalLog is inspired by the Structural Causal Model and the front-door adjustment principle. On this basis, counterfactual rewriting is implemented through tailored prompt engineering, aiming to mitigate biases without accessing LLM internals. In addition, n-gram statistics of log data are integrated as a bias-free reference for an adjustment score, which helps improve both parsing accuracy and interpretability. Experiments on public log datasets show that CausalLog outperforms state-of-the-art methods, providing observational evidence that it improves both log grouping and template extraction accuracy.
日志解析将非结构化的日志消息转换为结构化的格式,是执行各种日志分析任务的关键步骤。大型语言模型(llm)最近在这项任务中表现出了强大的性能。然而,他们往往依赖于他们的经验知识作为捷径,引入偏见和降低解析的准确性。为了解决这个问题,我们提出了CausalLog,这是一个轻量级的、灵活的日志解析去偏框架。CausalLog的灵感来源于结构因果模型和前门调整原则。在此基础上,通过定制的提示工程实现反事实重写,旨在在不访问LLM内部的情况下减轻偏见。此外,将日志数据的n-gram统计信息集成为调整分数的无偏差参考,有助于提高解析精度和可解释性。在公共日志数据集上的实验表明,CausalLog优于最先进的方法,提供了观察证据,表明它提高了日志分组和模板提取的准确性。
{"title":"CausalLog: Log parsing using LLMs with causal intervention for bias mitigation","authors":"Yuan Tian,&nbsp;Shi Ying,&nbsp;Tiangang Li","doi":"10.1016/j.ipm.2026.104609","DOIUrl":"10.1016/j.ipm.2026.104609","url":null,"abstract":"<div><div>Log parsing transforms unstructured log messages into structured formats, serving as a critical step for various log analysis tasks. Large language models (LLMs) have recently shown strong performance in this task. However, they tend to rely on their experiential knowledge as shortcuts, introducing bias and reducing parsing accuracy. To address this issue, we propose CausalLog, a lightweight and flexible debiasing framework for log parsing. CausalLog is inspired by the Structural Causal Model and the front-door adjustment principle. On this basis, counterfactual rewriting is implemented through tailored prompt engineering, aiming to mitigate biases without accessing LLM internals. In addition, n-gram statistics of log data are integrated as a bias-free reference for an adjustment score, which helps improve both parsing accuracy and interpretability. Experiments on public log datasets show that CausalLog outperforms state-of-the-art methods, providing observational evidence that it improves both log grouping and template extraction accuracy.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104609"},"PeriodicalIF":6.9,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural embeddings of collaboration networks predict citation impact and innovation 协作网络的神经嵌入预测引文影响和创新
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-20 DOI: 10.1016/j.ipm.2026.104634
Chengyu Li , Wenlong Yang , Meiling Li , Yang Wang
Scientific collaboration is a key driver of breakthroughs in contemporary science. While prior research has primarily focused on team diversity based on individual attributes, quantifying structural diversity based on collaborative relationships at the individual paper level remains underexplored. In this study, we apply neural embedding techniques to quantify structural diversity at the paper level and examine its association with research outcomes. Leveraging the node2vec and GraphSAGE algorithms, we embed each scientist into a dense vector space using the Microsoft Academic Graph dataset, covering 4,772,781 papers written by 5,162,397 scientists from 1995 to 2016. We show that our embedding model effectively classifies scientists across scientific domains and accurately predicts their primary scientific domains. Crucially, our analysis reveals that structural diversity is strongly associated with citation impact, novelty, and disruption at the individual paper level. Specifically, a 1SD increase in structural diversity is associated with a 7.6%, 24.7%, and 2.9% increase in citation impact, and the odds of publishing novel and disruptive papers, respectively. These findings are generalizable across multiple model specifications, time spans, different team sizes, and scientific domains. Finally, our analysis reveals that structural diversity exhibits the strongest correlation with citation impact, novelty, and disruption among all examined diversity measures. Our study highlights the importance of fostering structurally diverse collaborations and has policy implications for institutions, funders, and governments aiming to support impactful and innovative research.
科学合作是当代科学取得突破的关键驱动力。虽然先前的研究主要集中在基于个人属性的团队多样性上,但基于个人论文层面的合作关系的量化结构多样性仍未得到充分探索。在本研究中,我们应用神经嵌入技术在论文层面量化结构多样性,并检验其与研究成果的关系。利用node2vec和GraphSAGE算法,我们使用微软学术图数据集将每位科学家嵌入到密集的向量空间中,该数据集涵盖了1995年至2016年由5,162,397名科学家撰写的4,772,781篇论文。我们的嵌入模型有效地对科学家进行了跨科学领域的分类,并准确地预测了他们的主要科学领域。至关重要的是,我们的分析表明,结构多样性与单个论文的引用影响、新颖性和颠覆性密切相关。具体来说,结构多样性每增加1SD,引用影响力和发表新颖和颠覆性论文的几率分别增加7.6%、24.7%和2.9%。这些发现可以推广到多个模型规范、时间跨度、不同的团队规模和科学领域。最后,我们的分析表明,结构多样性与被引影响、新颖性和颠覆性的相关性最强。我们的研究强调了促进结构多样化合作的重要性,并对旨在支持有影响力和创新研究的机构、资助者和政府具有政策意义。
{"title":"Neural embeddings of collaboration networks predict citation impact and innovation","authors":"Chengyu Li ,&nbsp;Wenlong Yang ,&nbsp;Meiling Li ,&nbsp;Yang Wang","doi":"10.1016/j.ipm.2026.104634","DOIUrl":"10.1016/j.ipm.2026.104634","url":null,"abstract":"<div><div>Scientific collaboration is a key driver of breakthroughs in contemporary science. While prior research has primarily focused on team diversity based on individual attributes, quantifying structural diversity based on collaborative relationships at the individual paper level remains underexplored. In this study, we apply neural embedding techniques to quantify structural diversity at the paper level and examine its association with research outcomes. Leveraging the node2vec and GraphSAGE algorithms, we embed each scientist into a dense vector space using the Microsoft Academic Graph dataset, covering 4,772,781 papers written by 5,162,397 scientists from 1995 to 2016. We show that our embedding model effectively classifies scientists across scientific domains and accurately predicts their primary scientific domains. Crucially, our analysis reveals that structural diversity is strongly associated with citation impact, novelty, and disruption at the individual paper level. Specifically, a 1SD increase in structural diversity is associated with a 7.6%, 24.7%, and 2.9% increase in citation impact, and the odds of publishing novel and disruptive papers, respectively. These findings are generalizable across multiple model specifications, time spans, different team sizes, and scientific domains. Finally, our analysis reveals that structural diversity exhibits the strongest correlation with citation impact, novelty, and disruption among all examined diversity measures. Our study highlights the importance of fostering structurally diverse collaborations and has policy implications for institutions, funders, and governments aiming to support impactful and innovative research.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104634"},"PeriodicalIF":6.9,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial intelligence applications and innovation ambidexterity: The “inverted U-shaped” regulating effect of risk-taking 人工智能应用与创新双重性:风险承担的“倒u型”调节效应
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-20 DOI: 10.1016/j.ipm.2026.104635
Lihua Fu , Luo Jin , Yaokuang Li
Artificial intelligence (AI) is driving technological and industrial transformation, reshaping enterprise structure and innovation. Despite its importance, research lacks insight into AI’s impact on innovation ambidexterity and the underexplored role of risk-taking. Drawing on information processing theory, this study constructs a theoretical model to examine the impact of AI applications on enterprise exploration, exploitation, and innovation ambidexterity, as well as incorporates the level of risk-taking as a moderating variable. Using Stata 17 to perform panel data regression and a series of robustness tests, we analyze 1,908 firm-year observations from the information transmission, software, and information technology service industries, as well as the scientific research and technology service industries listed on the Shanghai and Shenzhen Stock Exchanges from 2012 to 2022. The empirical analyses propose that AI applications remarkably enhance innovation ambidexterity at the 1 % level. This positive effect is the most pronounced when risk-taking is moderate. This study extends information processing theory to the AI-enabled innovation context and further enriches its boundary conditions by introducing risk-taking as a nonlinear moderator. Managerially, the findings suggest that enterprises should calibrate risk-taking levels to complement AI deployment, enabling a balanced approach to exploration and exploitation.
人工智能(AI)正在推动技术和产业变革,重塑企业结构和创新。尽管它很重要,但研究缺乏对人工智能对创新的影响的洞察力,以及对风险承担作用的探索不足。利用信息处理理论,构建了人工智能应用对企业探索、开发和创新双元性影响的理论模型,并将风险承担水平作为调节变量。利用Stata 17进行面板数据回归和一系列稳健性检验,对2012 - 2022年沪深两市上市的信息传输、软件、信息技术服务行业以及科研和技术服务行业的1908个企业年观测数据进行了分析。实证分析表明,人工智能应用在1%的水平上显著提高了创新的双元性。当冒险适度时,这种积极影响最为明显。本研究将信息处理理论扩展到人工智能驱动的创新环境中,并通过引入风险承担作为非线性调节因子进一步丰富了其边界条件。在管理方面,研究结果表明,企业应调整风险承担水平,以补充人工智能的部署,从而实现勘探和开发的平衡方法。
{"title":"Artificial intelligence applications and innovation ambidexterity: The “inverted U-shaped” regulating effect of risk-taking","authors":"Lihua Fu ,&nbsp;Luo Jin ,&nbsp;Yaokuang Li","doi":"10.1016/j.ipm.2026.104635","DOIUrl":"10.1016/j.ipm.2026.104635","url":null,"abstract":"<div><div>Artificial intelligence (AI) is driving technological and industrial transformation, reshaping enterprise structure and innovation. Despite its importance, research lacks insight into AI’s impact on innovation ambidexterity and the underexplored role of risk-taking. Drawing on information processing theory, this study constructs a theoretical model to examine the impact of AI applications on enterprise exploration, exploitation, and innovation ambidexterity, as well as incorporates the level of risk-taking as a moderating variable. Using Stata 17 to perform panel data regression and a series of robustness tests, we analyze 1,908 firm-year observations from the information transmission, software, and information technology service industries, as well as the scientific research and technology service industries listed on the Shanghai and Shenzhen Stock Exchanges from 2012 to 2022. The empirical analyses propose that AI applications remarkably enhance innovation ambidexterity at the 1 % level. This positive effect is the most pronounced when risk-taking is moderate. This study extends information processing theory to the AI-enabled innovation context and further enriches its boundary conditions by introducing risk-taking as a nonlinear moderator. Managerially, the findings suggest that enterprises should calibrate risk-taking levels to complement AI deployment, enabling a balanced approach to exploration and exploitation.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104635"},"PeriodicalIF":6.9,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PRISM-X: Progressive semi-supervised threat detection in X-ray scans with self-guided multimodal refinement PRISM-X:自引导多模态改进的x射线扫描渐进半监督威胁检测
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-20 DOI: 10.1016/j.ipm.2026.104640
Abdelfatah Ahmed , Mohammad Irshaid , Mohamad Alansari , Divya Velayudhan , Mohammed Tarnini , Mohammed El-Amine Azz , Naser Al-Khalayleh , Taimur Hassan , Ernesto Damiani , Naoufel Werghi
X-ray baggage screening demands robust automated detection systems that can perform reliably under limited annotation. We introduce PRISM-X, a progressive semi-supervised framework that integrates pseudo-label generation, region-level multimodal grounding, and contrastive refinement into a unified detection pipeline. Evaluated on the SIXray and CLCXray benchmarks, PRISM-X consistently outperforms both vision-only and vision-language baselines. On SIXray, it achieves 66.0% mAP and 87.4% AP50, improving over the strongest vision-only method by +6.8% mAP and the leading vision-language model by +6.7% grounding accuracy. On CLCXray, PRISM-X reaches 45.7% mAP, 64.7% AP50, and 51.6% AP75, surpassing Cascade R-CNN by +3.3% mAP and GDINO by +3.4% AP75. These results demonstrate the effectiveness of PRISM-X in low-label, cluttered X-ray scenarios and its superiority over existing weakly and semi-supervised approaches.
x射线行李检查需要强大的自动检测系统,可以在有限的注释下可靠地执行。我们介绍了PRISM-X,这是一个渐进式半监督框架,将伪标签生成、区域级多模态接地和对比细化集成到统一的检测管道中。在SIXray和clxray基准测试中,PRISM-X始终优于纯视觉和视觉语言基线。在SIXray上,它的mAP值达到66.0%,AP50值达到87.4%,比最强的纯视觉方法mAP值提高了6.8%,比领先的视觉语言模型接地精度提高了6.7%。在CLCXray上,PRISM-X达到45.7%的mAP, 64.7%的AP50和51.6%的AP75,超过Cascade R-CNN +3.3%的mAP和GDINO +3.4%的AP75。这些结果证明了PRISM-X在低标签、杂乱x射线场景中的有效性,以及它比现有的弱监督和半监督方法的优越性。
{"title":"PRISM-X: Progressive semi-supervised threat detection in X-ray scans with self-guided multimodal refinement","authors":"Abdelfatah Ahmed ,&nbsp;Mohammad Irshaid ,&nbsp;Mohamad Alansari ,&nbsp;Divya Velayudhan ,&nbsp;Mohammed Tarnini ,&nbsp;Mohammed El-Amine Azz ,&nbsp;Naser Al-Khalayleh ,&nbsp;Taimur Hassan ,&nbsp;Ernesto Damiani ,&nbsp;Naoufel Werghi","doi":"10.1016/j.ipm.2026.104640","DOIUrl":"10.1016/j.ipm.2026.104640","url":null,"abstract":"<div><div>X-ray baggage screening demands robust automated detection systems that can perform reliably under limited annotation. We introduce PRISM-X, a progressive semi-supervised framework that integrates pseudo-label generation, region-level multimodal grounding, and contrastive refinement into a unified detection pipeline. Evaluated on the SIXray and CLCXray benchmarks, PRISM-X consistently outperforms both vision-only and vision-language baselines. On SIXray, it achieves 66.0% mAP and 87.4% AP<sub>50</sub>, improving over the strongest vision-only method by +6.8% mAP and the leading vision-language model by +6.7% grounding accuracy. On CLCXray, PRISM-X reaches 45.7% mAP, 64.7% AP<sub>50</sub>, and 51.6% AP<sub>75</sub>, surpassing Cascade R-CNN by +3.3% mAP and GDINO by +3.4% AP<sub>75</sub>. These results demonstrate the effectiveness of PRISM-X in low-label, cluttered X-ray scenarios and its superiority over existing weakly and semi-supervised approaches.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104640"},"PeriodicalIF":6.9,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SEHLP: A summary-enhanced large language model for financial report sentiment analysis via hybrid LoRA and dynamic prefix tuning SEHLP:通过混合LoRA和动态前缀调优,用于财务报告情感分析的摘要增强大型语言模型
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-20 DOI: 10.1016/j.ipm.2026.104639
Haozhou Li, Qinke Peng, Xu Mou, Zeyuan Zeng, Ruimeng Li, Jinzhi Wang, Wentong Sun
Financial sentiment analysis (FSA) has garnered considerable attention for its potential to detect bullish and bearish sentiments that drive stock market fluctuations. Nonetheless, extracting salient sentiments from analyst reports encounters two main challenges. First, the highly specialized terms and expressions prevalent in these reports make it difficult for general Large Language Models (LLMs) to interpret financial expertise. Second, sentiment cues are implicit and dispersed across long-range dependencies, whereas existing LLM-based FSA methods relying on a single fine-tuning strategy lack fine-grained control during adaptation, thus leading to key information loss. To tackle these issues, we propose SEHLP, the first LLM that integrates summary information with a hybrid adaptation strategy that combines Low-rank Adaptation (LoRA) and dynamic Prefix Tuning to enhance FSA. Specifically, we employ prompt engineering on Qwen-2.5-14B to generate concise summaries that distill salient insights of each report, and construct FinLLaMA as SEHLP’s backbone through Supervised Fine-tuning (SFT) on extensive domain-specific instructions, enhancing financial knowledge comprehension. To inject summary information and enable fine-grained control during fine-tuning, we propose a hybrid adaptation strategy that concatenates LoRA-updated attention projections with dynamic summary-enhanced key-value prefixes, thereby fully utilizing sentiment cues in analyst reports and their summaries. Moreover, we construct a large-scale LCFR-Instruct corpus with 16,912 samples to address the lack of high-quality FSA instruction data. Comprehensive experiments on the LCFR-Instruct and FinTHUC-Instruct benchmark datasets indicate that SEHLP, with only 1.3B parameters, consistently surpasses competing LLMs, exhibiting ACC gains of 1.89% and 1.59% over the larger FinGPT-7B model on both datasets while maintaining superior efficiency. Our code is publicly accessible at https://github.com/lhz9999/SEHLP.
金融情绪分析(Financial sentiment analysis,简称FSA)因能够发现驱动股市波动的看涨和看跌情绪而备受关注。然而,从分析师报告中提取重要情绪面临着两个主要挑战。首先,这些报告中普遍存在的高度专业化的术语和表达使得一般的大型语言模型(llm)很难解释金融专业知识。其次,情绪线索是隐式的,分散在长期依赖关系中,而现有的基于llm的FSA方法依赖于单一的微调策略,在适应过程中缺乏细粒度控制,从而导致关键信息丢失。为了解决这些问题,我们提出了SEHLP,这是第一个将摘要信息与混合自适应策略相结合的LLM,该策略结合了低秩自适应(LoRA)和动态前缀调优来增强FSA。具体而言,我们在Qwen-2.5-14B上采用即时工程生成简明摘要,从中提炼出每篇报告的突出见解,并在广泛的领域特定指令上通过监督微调(SFT)构建FinLLaMA作为SEHLP的主干,增强金融知识理解。为了注入摘要信息并在微调过程中实现细粒度控制,我们提出了一种混合适应策略,该策略将lora更新的注意力预测与动态摘要增强的键值前缀连接起来,从而充分利用分析师报告及其摘要中的情绪线索。此外,我们构建了一个包含16,912个样本的大规模lcfr - instruction语料库,以解决缺乏高质量的FSA指令数据的问题。在lcfr - directive和finthuc - directive基准数据集上进行的综合实验表明,仅使用13亿个参数的SEHLP始终优于竞争对手的llm,在这两个数据集上,SEHLP比更大的FinGPT-7B模型的ACC增益分别为1.89%和1.59%,同时保持了优越的效率。我们的代码可以在https://github.com/lhz9999/SEHLP上公开访问。
{"title":"SEHLP: A summary-enhanced large language model for financial report sentiment analysis via hybrid LoRA and dynamic prefix tuning","authors":"Haozhou Li,&nbsp;Qinke Peng,&nbsp;Xu Mou,&nbsp;Zeyuan Zeng,&nbsp;Ruimeng Li,&nbsp;Jinzhi Wang,&nbsp;Wentong Sun","doi":"10.1016/j.ipm.2026.104639","DOIUrl":"10.1016/j.ipm.2026.104639","url":null,"abstract":"<div><div>Financial sentiment analysis (FSA) has garnered considerable attention for its potential to detect bullish and bearish sentiments that drive stock market fluctuations. Nonetheless, extracting salient sentiments from analyst reports encounters two main challenges. First, the highly specialized terms and expressions prevalent in these reports make it difficult for general Large Language Models (LLMs) to interpret financial expertise. Second, sentiment cues are implicit and dispersed across long-range dependencies, whereas existing LLM-based FSA methods relying on a single fine-tuning strategy lack fine-grained control during adaptation, thus leading to key information loss. To tackle these issues, we propose SEHLP, the first LLM that integrates summary information with a hybrid adaptation strategy that combines Low-rank Adaptation (LoRA) and dynamic Prefix Tuning to enhance FSA. Specifically, we employ prompt engineering on Qwen-2.5-14B to generate concise summaries that distill salient insights of each report, and construct FinLLaMA as SEHLP’s backbone through Supervised Fine-tuning (SFT) on extensive domain-specific instructions, enhancing financial knowledge comprehension. To inject summary information and enable fine-grained control during fine-tuning, we propose a hybrid adaptation strategy that concatenates LoRA-updated attention projections with dynamic summary-enhanced key-value prefixes, thereby fully utilizing sentiment cues in analyst reports and their summaries. Moreover, we construct a large-scale LCFR-Instruct corpus with 16,912 samples to address the lack of high-quality FSA instruction data. Comprehensive experiments on the LCFR-Instruct and FinTHUC-Instruct benchmark datasets indicate that SEHLP, with only 1.3B parameters, consistently surpasses competing LLMs, exhibiting ACC gains of 1.89% and 1.59% over the larger FinGPT-7B model on both datasets while maintaining superior efficiency. Our code is publicly accessible at <span><span>https://github.com/lhz9999/SEHLP</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104639"},"PeriodicalIF":6.9,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond efficient fine-tuning: Efficient hybrid fine-tuning of CLIP models guided by explainable ViT attention 超越高效的微调:由可解释的ViT注意力指导的CLIP模型的高效混合微调
IF 6.9 1区 管理学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-20 DOI: 10.1016/j.ipm.2026.104628
Hui Ye , Xuri Ge , Junqi Wang , Junchen Fu , Xin Xin , Jiao Xue , Yao Chen , Pengjie Ren , Zhumin Chen
To address the inefficiency of full fine-tuning of Contrastive Language-Image Pre-training (CLIP) models and the performance loss of adapter-based methods, we propose a novel efficient hybrid fine-tuning strategy (called HFLIP) to achieve a balance of efficiency and performance. HFLIP fine-tunes the key selected ViT blocks with interpretable semantic attention supervision on selected transformer heads via machine-learning methods’ selection, while keeping other blocks adapter-based for efficiency. Specifically, HFLIP introduces two key components: (1) a Dynamic Block-selection Genetic Algorithm (DBGA) that automatically selects a small subset of critical blocks in the ViT for full tuning, while keeping the rest adapter-tuned, ensuring a proper trade-off between fine-tuning effectiveness and efficiency; and (2) a Clustering-based Head-selection with Explainable-attention Guidance (CHEG), where hierarchical clustering is employed to identify representative attention heads, which are then fine-tuned under guidance from explainable attention maps, encouraging semantically consistent and globally diverse attention patterns. Extensive experiments on multiple downstream tasks show that HFLIP achieves comparable or even better performance than full fine-tuning by updating only 30% of the training parameters, while reducing GPU memory consumption by about 16%. In addition, HFLIP makes the CLIP-based ViT attention mechanism more interpretable compared to both the pretrained CLIP and other fine-tuned variants. We release our code at https://github.com/huiye8870/HFLIP.
为了解决对比语言-图像预训练(CLIP)模型完全微调的低效率和基于适配器的方法的性能损失问题,我们提出了一种新的高效混合微调策略(称为HFLIP)来实现效率和性能的平衡。HFLIP通过机器学习方法的选择对选定的变压器磁头进行可解释语义注意力监督的关键选定ViT块进行微调,同时保持其他块基于适配器的效率。具体来说,HFLIP引入了两个关键组件:(1)动态块选择遗传算法(DBGA),该算法自动选择ViT中的一小部分关键块进行全面调优,同时保持其余适配器调优,确保在微调效果和效率之间进行适当的权衡;(2)基于聚类的可解释注意指导(CHEG)头选择,其中采用分层聚类来识别具有代表性的注意头,然后在可解释注意图的指导下对其进行微调,从而鼓励语义一致和全局多样化的注意模式。在多个下游任务上进行的大量实验表明,HFLIP通过仅更新30%的训练参数,实现了与完全微调相当甚至更好的性能,同时减少了约16%的GPU内存消耗。此外,与预训练的CLIP和其他微调的变体相比,HFLIP使基于CLIP的ViT注意机制更具可解释性。我们在https://github.com/huiye8870/HFLIP上发布我们的代码。
{"title":"Beyond efficient fine-tuning: Efficient hybrid fine-tuning of CLIP models guided by explainable ViT attention","authors":"Hui Ye ,&nbsp;Xuri Ge ,&nbsp;Junqi Wang ,&nbsp;Junchen Fu ,&nbsp;Xin Xin ,&nbsp;Jiao Xue ,&nbsp;Yao Chen ,&nbsp;Pengjie Ren ,&nbsp;Zhumin Chen","doi":"10.1016/j.ipm.2026.104628","DOIUrl":"10.1016/j.ipm.2026.104628","url":null,"abstract":"<div><div>To address the inefficiency of full fine-tuning of Contrastive Language-Image Pre-training (CLIP) models and the performance loss of adapter-based methods, we propose a novel efficient hybrid fine-tuning strategy (called HFLIP) to achieve a balance of efficiency and performance. HFLIP fine-tunes the key selected ViT blocks with interpretable semantic attention supervision on selected transformer heads via machine-learning methods’ selection, while keeping other blocks adapter-based for efficiency. Specifically, HFLIP introduces two key components: (1) a Dynamic Block-selection Genetic Algorithm (DBGA) that automatically selects a small subset of critical blocks in the ViT for full tuning, while keeping the rest adapter-tuned, ensuring a proper trade-off between fine-tuning effectiveness and efficiency; and (2) a Clustering-based Head-selection with Explainable-attention Guidance (CHEG), where hierarchical clustering is employed to identify representative attention heads, which are then fine-tuned under guidance from explainable attention maps, encouraging semantically consistent and globally diverse attention patterns. Extensive experiments on multiple downstream tasks show that HFLIP achieves comparable or even better performance than full fine-tuning by updating only 30% of the training parameters, while reducing GPU memory consumption by about 16%. In addition, HFLIP makes the CLIP-based ViT attention mechanism more interpretable compared to both the pretrained CLIP and other fine-tuned variants. We release our code at <span><span>https://github.com/huiye8870/HFLIP</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 4","pages":"Article 104628"},"PeriodicalIF":6.9,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146023316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Information Processing & Management
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1