首页 > 最新文献

Applied Intelligence最新文献

英文 中文
Automatic rib segmentation and sequential labeling via multi-axial slicing and 3D reconstruction 通过多轴切片和三维重建实现肋骨自动分割和顺序标记
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-26 DOI: 10.1007/s10489-024-05785-4
Hyunsung Kim, Seonghyeon Ko, Junghyun Bum, Duc-Tai Le, Hyunseung Choo

Radiologists often inspect hundreds of two-dimensional computed-tomography (CT) images to accurately locate lesions and make diagnoses, by classifying and labeling the ribs. However, this task is repetitive and time consuming. To effectively address this problem, we propose a multi-axial rib segmentation and sequential labeling (MARSS) method. First, we slice the CT volume into sagittal, frontal, and transverse planes for segmentation. The segmentation masks generated for each plane are then reconstructed into a single 3D segmentation mask using binarization techniques. After separating the left and right rib volumes from the entire CT volume, we cluster the connected components identified as bones and sequentially assign labels to each rib. The segmentation and sequential labeling performance of this method outperformed existing methods by up to 4.2%. The proposed automatic rib sequential labeling method enhances the efficiency of radiologists. In addition, this method provides an extended opportunity for advancements not only in rib segmentation but also in bone-fracture detection and lesion-diagnosis research.

放射科医生经常要检查数百张二维计算机断层扫描(CT)图像,通过对肋骨进行分类和标记,准确定位病灶并做出诊断。然而,这项工作既重复又耗时。为有效解决这一问题,我们提出了一种多轴肋骨分割和连续标记(MARSS)方法。首先,我们将 CT 体切成矢状面、额状面和横向面进行分割。然后,利用二值化技术将每个平面生成的分割掩膜重建为一个单独的三维分割掩膜。将左右肋骨卷从整个 CT 卷中分离出来后,我们将被识别为骨骼的连接组件聚类,并按顺序为每根肋骨分配标签。该方法的分割和顺序标签性能比现有方法高出 4.2%。所提出的自动肋骨顺序标记方法提高了放射科医生的工作效率。此外,这种方法不仅为肋骨分割,还为骨骨折检测和病变诊断研究提供了更多进步的机会。
{"title":"Automatic rib segmentation and sequential labeling via multi-axial slicing and 3D reconstruction","authors":"Hyunsung Kim,&nbsp;Seonghyeon Ko,&nbsp;Junghyun Bum,&nbsp;Duc-Tai Le,&nbsp;Hyunseung Choo","doi":"10.1007/s10489-024-05785-4","DOIUrl":"10.1007/s10489-024-05785-4","url":null,"abstract":"<div><p>Radiologists often inspect hundreds of two-dimensional computed-tomography (CT) images to accurately locate lesions and make diagnoses, by classifying and labeling the ribs. However, this task is repetitive and time consuming. To effectively address this problem, we propose a multi-axial rib segmentation and sequential labeling (MARSS) method. First, we slice the CT volume into sagittal, frontal, and transverse planes for segmentation. The segmentation masks generated for each plane are then reconstructed into a single 3D segmentation mask using binarization techniques. After separating the left and right rib volumes from the entire CT volume, we cluster the connected components identified as bones and sequentially assign labels to each rib. The segmentation and sequential labeling performance of this method outperformed existing methods by up to 4.2%. The proposed automatic rib sequential labeling method enhances the efficiency of radiologists. In addition, this method provides an extended opportunity for advancements not only in rib segmentation but also in bone-fracture detection and lesion-diagnosis research.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12644 - 12660"},"PeriodicalIF":3.4,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A framework based on physics-informed graph neural ODE: for continuous spatial-temporal pandemic prediction 基于物理信息图神经 ODE 的框架:用于连续时空流行病预测
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-26 DOI: 10.1007/s10489-024-05834-y
Haodong Cheng, Yingchi Mao, Xiao Jia

Physics-informed spatial-temporal discrete sequence learning networks have great potential in solving partial differential equations and time series prediction compared to traditional fully connected PINN algorithms, and can serve as the foundation for data-driven sequence prediction modeling and inverse problem analysis. However, such existing models are unable to deal with inverse problem scenarios in which the parameters of the physical process are time-varying and unknown, while usually failing to make predictions in continuous time. In this paper, we propose a continuous time series prediction algorithm constructed by the physics-informed graph neural ordinary differential equation (PGNODE). Proposed parameterized GNODE-GRU and physics-informed loss constraints are used to explicitly characterize and solve unknown time-varying hyperparameters. The GNODE solver integrates this physical parameter to predict the sequence value at any time. This paper uses epidemic prediction tasks as a case study, and experimental results demonstrate that the proposed algorithm can effectively improve the prediction accuracy of the spread of epidemics in the future continuous time.

与传统的全连接 PINN 算法相比,物理信息时空离散序列学习网络在求解偏微分方程和时间序列预测方面具有巨大潜力,可作为数据驱动序列预测建模和逆问题分析的基础。然而,现有模型无法处理物理过程参数时变和未知的逆问题场景,同时通常无法在连续时间内进行预测。本文提出了一种由物理信息图神经常微分方程(PGNODE)构建的连续时间序列预测算法。提出的参数化 GNODE-GRU 和物理信息损失约束用于明确描述和求解未知的时变超参数。GNODE 求解器整合了这一物理参数,以预测任何时间的序列值。本文以流行病预测任务为案例进行研究,实验结果表明,所提出的算法能有效提高流行病在未来连续时间内传播的预测精度。
{"title":"A framework based on physics-informed graph neural ODE: for continuous spatial-temporal pandemic prediction","authors":"Haodong Cheng,&nbsp;Yingchi Mao,&nbsp;Xiao Jia","doi":"10.1007/s10489-024-05834-y","DOIUrl":"10.1007/s10489-024-05834-y","url":null,"abstract":"<div><p>Physics-informed spatial-temporal discrete sequence learning networks have great potential in solving partial differential equations and time series prediction compared to traditional fully connected PINN algorithms, and can serve as the foundation for data-driven sequence prediction modeling and inverse problem analysis. However, such existing models are unable to deal with inverse problem scenarios in which the parameters of the physical process are time-varying and unknown, while usually failing to make predictions in continuous time. In this paper, we propose a continuous time series prediction algorithm constructed by the physics-informed graph neural ordinary differential equation (PGNODE). Proposed parameterized GNODE-GRU and physics-informed loss constraints are used to explicitly characterize and solve unknown time-varying hyperparameters. The GNODE solver integrates this physical parameter to predict the sequence value at any time. This paper uses epidemic prediction tasks as a case study, and experimental results demonstrate that the proposed algorithm can effectively improve the prediction accuracy of the spread of epidemics in the future continuous time.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12661 - 12675"},"PeriodicalIF":3.4,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FMCF: Few-shot Multimodal aspect-based sentiment analysis framework based on Contrastive Finetuning FMCF:基于对比微调的少镜头多模态方面情感分析框架
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-25 DOI: 10.1007/s10489-024-05841-z
Yongping Du, Runfeng Xie, Bochao Zhang, Zihao Yin

Multimodal aspect-based sentiment analysis (MABSA) aims to predict the sentiment of aspect by the fusion of different modalities such as image, text and so on. However, the availability of high-quality multimodal data remains limited. Therefore, few-shot MABSA is a new challenge. Previous works are rarely able to cope with low-resource and few-shot scenarios. In order to address the above problems, we design a Few-shot Multimodal aspect-based sentiment analysis framework based on Contrastive Finetuning (FMCF). Initially, the image modality is transformed to the corresponding textual caption to achieve the entailed semantic information and a contrastive dataset is constructed based on similarity retrieval for finetuning in the following stage. Further, a sentence encoder is trained based on SBERT, which combines supervised contrastive learning and sentence-level multi-feature fusion to complete MABSA. The experiments demonstrate that our framework achieves excellent performance in the few-shot scenarios. Importantly, with only 256 training samples and limited computational resources, the proposed method outperforms fine-tuned models that use all available data on the Twitter dataset.

基于多模态方面的情感分析(MABSA)旨在通过融合图像、文本等不同模态来预测方面的情感。然而,高质量多模态数据的可用性仍然有限。因此,少镜头 MABSA 是一个新的挑战。以往的研究很少能应对低资源和少镜头场景。为了解决上述问题,我们设计了一种基于对比微调(FMCF)的少镜头多模态情感分析框架。首先,将图像模态转换为相应的文字说明,以获得所包含的语义信息,然后根据相似性检索构建对比数据集,以便在下一阶段进行微调。然后,基于 SBERT 训练句子编码器,将有监督的对比学习和句子级多特征融合结合起来,完成 MABSA。实验证明,我们的框架在少拍场景中取得了优异的性能。重要的是,在只有 256 个训练样本和有限计算资源的情况下,所提出的方法优于使用 Twitter 数据集上所有可用数据的微调模型。
{"title":"FMCF: Few-shot Multimodal aspect-based sentiment analysis framework based on Contrastive Finetuning","authors":"Yongping Du,&nbsp;Runfeng Xie,&nbsp;Bochao Zhang,&nbsp;Zihao Yin","doi":"10.1007/s10489-024-05841-z","DOIUrl":"10.1007/s10489-024-05841-z","url":null,"abstract":"<div><p>Multimodal aspect-based sentiment analysis (MABSA) aims to predict the sentiment of aspect by the fusion of different modalities such as image, text and so on. However, the availability of high-quality multimodal data remains limited. Therefore, few-shot MABSA is a new challenge. Previous works are rarely able to cope with low-resource and few-shot scenarios. In order to address the above problems, we design a <b>F</b>ew-shot <b>M</b>ultimodal aspect-based sentiment analysis framework based on <b>C</b>ontrastive <b>F</b>inetuning (FMCF). Initially, the image modality is transformed to the corresponding textual caption to achieve the entailed semantic information and a contrastive dataset is constructed based on similarity retrieval for finetuning in the following stage. Further, a sentence encoder is trained based on SBERT, which combines supervised contrastive learning and sentence-level multi-feature fusion to complete MABSA. The experiments demonstrate that our framework achieves excellent performance in the few-shot scenarios. Importantly, with only 256 training samples and limited computational resources, the proposed method outperforms fine-tuned models that use all available data on the Twitter dataset.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12629 - 12643"},"PeriodicalIF":3.4,"publicationDate":"2024-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable cognitive decline detection in free dialogues with a Machine Learning approach based on pre-trained Large Language Models 利用基于预训练大型语言模型的机器学习方法,在自由对话中检测可解释的认知能力下降情况
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-24 DOI: 10.1007/s10489-024-05808-0
Francisco de Arriba-Pérez, Silvia García-Méndez, Javier Otero-Mosquera, Francisco J. González-Castaño

Cognitive and neurological impairments are very common, but only a small proportion of affected individuals are diagnosed and treated, partly because of the high costs associated with frequent screening. Detecting pre-illness stages and analyzing the progression of neurological disorders through effective and efficient intelligent systems can be beneficial for timely diagnosis and early intervention. We propose using Large Language Models to extract features from free dialogues to detect cognitive decline. These features comprise high-level reasoning content-independent features (such as comprehension, decreased awareness, increased distraction, and memory problems). Our solution comprises (i) preprocessing, (ii) feature engineering via Natural Language Processing techniques and prompt engineering, (iii) feature analysis and selection to optimize performance, and (iv) classification, supported by automatic explainability. We also explore how to improve Chatgpt’s direct cognitive impairment prediction capabilities using the best features in our models. Evaluation metrics obtained endorse the effectiveness of a mixed approach combining feature extraction with Chatgpt and a specialized Machine Learning model to detect cognitive decline within free-form conversational dialogues with older adults. Ultimately, our work may facilitate the development of an inexpensive, non-invasive, and rapid means of detecting and explaining cognitive decline.

认知和神经系统损伤非常常见,但只有一小部分患者得到诊断和治疗,部分原因是频繁筛查所需的高昂费用。通过有效和高效的智能系统检测疾病的前期阶段并分析神经系统疾病的进展情况,有利于及时诊断和早期干预。我们建议使用大型语言模型从自由对话中提取特征来检测认知能力衰退。这些特征包括与内容无关的高级推理特征(如理解能力、意识下降、注意力分散和记忆问题)。我们的解决方案包括:(i) 预处理;(ii) 通过自然语言处理技术和提示工程进行特征工程;(iii) 特征分析和选择以优化性能;(iv) 在自动可解释性的支持下进行分类。我们还探索了如何利用模型中的最佳特征来提高 Chatgpt 直接预测认知障碍的能力。所获得的评估指标证明了将 Chatgpt 的特征提取与专门的机器学习模型相结合的混合方法在检测老年人自由形式对话中的认知能力下降方面的有效性。最终,我们的工作将有助于开发一种廉价、非侵入性和快速的认知衰退检测和解释方法。
{"title":"Explainable cognitive decline detection in free dialogues with a Machine Learning approach based on pre-trained Large Language Models","authors":"Francisco de Arriba-Pérez,&nbsp;Silvia García-Méndez,&nbsp;Javier Otero-Mosquera,&nbsp;Francisco J. González-Castaño","doi":"10.1007/s10489-024-05808-0","DOIUrl":"10.1007/s10489-024-05808-0","url":null,"abstract":"<div><p>Cognitive and neurological impairments are very common, but only a small proportion of affected individuals are diagnosed and treated, partly because of the high costs associated with frequent screening. Detecting pre-illness stages and analyzing the progression of neurological disorders through effective and efficient intelligent systems can be beneficial for timely diagnosis and early intervention. We propose using Large Language Models to extract features from free dialogues to detect cognitive decline. These features comprise high-level reasoning content-independent features (such as comprehension, decreased awareness, increased distraction, and memory problems). Our solution comprises (<i>i</i>) preprocessing, (<i>ii</i>) feature engineering via Natural Language Processing techniques and prompt engineering, (<i>iii</i>) feature analysis and selection to optimize performance, and (<i>iv</i>) classification, supported by automatic explainability. We also explore how to improve Chat<span>gpt</span>’s direct cognitive impairment prediction capabilities using the best features in our models. Evaluation metrics obtained endorse the effectiveness of a mixed approach combining feature extraction with Chat<span>gpt</span> and a specialized Machine Learning model to detect cognitive decline within free-form conversational dialogues with older adults. Ultimately, our work may facilitate the development of an inexpensive, non-invasive, and rapid means of detecting and explaining cognitive decline.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12613 - 12628"},"PeriodicalIF":3.4,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-label feature selection for missing labels by granular-ball based mutual information 基于颗粒球互信息的缺失标签多标签特征选择
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-23 DOI: 10.1007/s10489-024-05809-z
Wenhao Shu, Yichen Hu, Wenbin Qian

Multi-label feature selection serves an effective dimensionality reduction technique in the high-dimensional multi-label data. However, most feature selection methods regard the label as complete. In fact, in real-world applications, labels in a multi-label dataset may be missing due to various difficulties in collecting sufficient labels, which enables some valuable information to be overlooked and leads to an inaccurate prediction in the classification. To address these issues, a feature selection algorithm based on the granular-ball based mutual information is proposed for the multi-label data with missing labels in this paper. At first, to improve the classification ability, a label recovery model is proposed to calculate some labels, which utilizes the correlation between labels, the properties of label specific features and global common features. Secondly, to avoid computing the neighborhood radius, a granular-ball based mutual information metric for evaluating candidate features is proposed, which well fits the data distribution. Finally, the corresponding feature selection algorithm is developed for selecting a subset from the multi-label data with missing labels. Experiments on the different datasets demonstrate that compared with the state-of-the-art algorithms the proposed algorithm considerably improves the classification accuracy. The code is publicly available online at https://github.com/skylark-leo/MLMLFS.git

在高维多标签数据中,多标签特征选择是一种有效的降维技术。然而,大多数特征选择方法都认为标签是完整的。事实上,在实际应用中,由于收集足够标签的各种困难,多标签数据集中的标签可能会缺失,这使得一些有价值的信息被忽视,导致分类预测不准确。针对这些问题,本文提出了一种基于颗粒球互信息的特征选择算法,用于处理标签缺失的多标签数据。首先,为了提高分类能力,本文提出了一种标签恢复模型,利用标签之间的相关性、标签特定特征的属性和全局公共特征,计算出一些标签。其次,为了避免计算邻域半径,本文提出了一种基于颗粒球的互信息指标来评估候选特征,该指标能很好地贴合数据分布。最后,开发了相应的特征选择算法,用于从缺失标签的多标签数据中选择子集。在不同数据集上的实验表明,与最先进的算法相比,所提出的算法大大提高了分类准确率。代码可在 https://github.com/skylark-leo/MLMLFS.git 在线公开获取。
{"title":"Multi-label feature selection for missing labels by granular-ball based mutual information","authors":"Wenhao Shu,&nbsp;Yichen Hu,&nbsp;Wenbin Qian","doi":"10.1007/s10489-024-05809-z","DOIUrl":"10.1007/s10489-024-05809-z","url":null,"abstract":"<p>Multi-label feature selection serves an effective dimensionality reduction technique in the high-dimensional multi-label data. However, most feature selection methods regard the label as complete. In fact, in real-world applications, labels in a multi-label dataset may be missing due to various difficulties in collecting sufficient labels, which enables some valuable information to be overlooked and leads to an inaccurate prediction in the classification. To address these issues, a feature selection algorithm based on the granular-ball based mutual information is proposed for the multi-label data with missing labels in this paper. At first, to improve the classification ability, a label recovery model is proposed to calculate some labels, which utilizes the correlation between labels, the properties of label specific features and global common features. Secondly, to avoid computing the neighborhood radius, a granular-ball based mutual information metric for evaluating candidate features is proposed, which well fits the data distribution. Finally, the corresponding feature selection algorithm is developed for selecting a subset from the multi-label data with missing labels. Experiments on the different datasets demonstrate that compared with the state-of-the-art algorithms the proposed algorithm considerably improves the classification accuracy. The code is publicly available online at https://github.com/skylark-leo/MLMLFS.git</p>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 23","pages":"12589 - 12612"},"PeriodicalIF":3.4,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142413386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain adaptation of time series via contrastive learning with task-specific consistency 通过具有特定任务一致性的对比学习实现时间序列的领域适应性
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-20 DOI: 10.1007/s10489-024-05799-y
Tao Wu, Qiushu Chen, Dongfang Zhao, Jinhua Wang, Linhua Jiang

Unsupervised domain adaptation (UDA) for time series analysis remains challenging due to the lack of labeled data in target domains. Existing methods rely heavily on auxiliary data yet often fail to fully exploit the intrinsic task consistency between different domains. To address this limitation, we propose a novel time series UDA framework called CLTC that enhances feature transferability by capturing semantic context and reconstructing class-wise representations. Specifically, contrastive learning is first utilized to capture contextual representations that enable label transfer across domains. Dual reconstruction on samples from the same class then refines the task-specific features to improve consistency. To align the cross-domain distributions without target labels, we leverage Sinkhorn divergence which can handle non-overlapping supports. Consequently, our CLTC reduces the domain gap while retaining task-specific consistency for effective knowledge transfer. Extensive experiments on four time series benchmarks demonstrate state-of-the-art performance improvements of 0.7-3.6% over existing methods, and ablation study validates the efficacy of each component.

由于缺乏目标领域的标注数据,用于时间序列分析的无监督领域适应(UDA)仍然具有挑战性。现有方法严重依赖辅助数据,但往往无法充分利用不同领域之间任务的内在一致性。为了解决这一局限性,我们提出了一种名为 CLTC 的新型时间序列 UDA 框架,该框架通过捕捉语义上下文和重建类别表征来增强特征的可转移性。具体来说,首先利用对比学习来捕捉上下文表征,从而实现跨领域的标签转移。然后,对同一类别的样本进行双重重构,完善特定任务的特征,以提高一致性。为了对齐没有目标标签的跨域分布,我们利用了可以处理非重叠支持的 Sinkhorn 分歧。因此,我们的 CLTC 在减少领域差距的同时,还保持了特定任务的一致性,从而实现了有效的知识转移。在四种时间序列基准上进行的广泛实验表明,与现有方法相比,我们的 CLTC 性能提高了 0.7-3.6%,而消融研究则验证了每个组件的功效。
{"title":"Domain adaptation of time series via contrastive learning with task-specific consistency","authors":"Tao Wu,&nbsp;Qiushu Chen,&nbsp;Dongfang Zhao,&nbsp;Jinhua Wang,&nbsp;Linhua Jiang","doi":"10.1007/s10489-024-05799-y","DOIUrl":"10.1007/s10489-024-05799-y","url":null,"abstract":"<div><p>Unsupervised domain adaptation (UDA) for time series analysis remains challenging due to the lack of labeled data in target domains. Existing methods rely heavily on auxiliary data yet often fail to fully exploit the intrinsic task consistency between different domains. To address this limitation, we propose a novel time series UDA framework called CLTC that enhances feature transferability by capturing semantic context and reconstructing class-wise representations. Specifically, contrastive learning is first utilized to capture contextual representations that enable label transfer across domains. Dual reconstruction on samples from the same class then refines the task-specific features to improve consistency. To align the cross-domain distributions without target labels, we leverage Sinkhorn divergence which can handle non-overlapping supports. Consequently, our CLTC reduces the domain gap while retaining task-specific consistency for effective knowledge transfer. Extensive experiments on four time series benchmarks demonstrate state-of-the-art performance improvements of 0.7-3.6% over existing methods, and ablation study validates the efficacy of each component.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 23","pages":"12576 - 12588"},"PeriodicalIF":3.4,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142412770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DAGCN: hybrid model for efficiently handling joint node and link prediction in cloud workflows DAGCN:高效处理云工作流中节点和链路联合预测的混合模型
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1007/s10489-024-05828-w
Ruimin Ma, Junqi Gao, Li Cheng, Yuyi Zhang, Ovanes Petrosian

In the cloud computing domain, significant strides have been made in performance prediction for cloud workflows, yet link prediction for cloud workflows remains largely unexplored. This paper introduces a novel challenge: joint node and link prediction in cloud workflows, with the aim of increasing the efficiency and overall performance of cloud computing resources. GNN-based methods have gained traction in handling graph-related tasks. The unique format of the DAG presents an underexplored area for GNNs effectiveness. To enhance comprehension of intricate graph structures and interrelationships, this paper introduces two novel models under the DAGCN framework: DAG-ConvGCN and DAG-AttGCN. The former synergizes the local receptive fields of the CNN with the global interpretive power of the GCN, whereas the latter integrates an attention mechanism to dynamically weigh the significance of node adjacencies. Through rigorous experimentation on a meticulously crafted joint node and link prediction task utilizing the Cluster-trace-v2018 dataset, both DAG-ConvGCN and DAG-AttGCN demonstrate superior performance over a spectrum of established machine learning and deep learning benchmarks. Moreover, the application of similarity measures such as the propagation kernel and the innovative GRBF kernel-which merges the graphlet kernel with the radial basis function kernel to accentuate graph topology and node features-reinforces the superiority of DAGCN models over graph-level prediction accuracy conventional baselines. This paper offers a fresh vantage point for advancing predictive methodologies within graph theory.

在云计算领域,云工作流的性能预测取得了长足进步,但云工作流的链接预测在很大程度上仍未得到探索。本文提出了一个新的挑战:云工作流中的节点和链接联合预测,旨在提高云计算资源的效率和整体性能。基于 GNN 的方法在处理与图相关的任务时受到了广泛关注。DAG 的独特格式为 GNN 的有效性提供了一个尚未充分开发的领域。为了加强对复杂图结构和相互关系的理解,本文在 DAGCN 框架下引入了两个新模型:DAG-ConvGCN 和 DAG-AttGCN。前者将 CNN 的局部感受野与 GCN 的全局解释能力协同起来,后者则整合了一种注意力机制,以动态权衡节点邻接关系的重要性。通过对利用 Cluster-trace-v2018 数据集精心设计的联合节点和链接预测任务进行严格实验,DAG-ConvGCN 和 DAG-AttGCN 在一系列成熟的机器学习和深度学习基准测试中都表现出了卓越的性能。此外,传播内核和创新 GRBF 内核等相似性度量的应用(GRBF 内核将小图内核与径向基函数内核合并,以突出图的拓扑结构和节点特征)加强了 DAGCN 模型在图级预测准确性传统基准上的优越性。本文为推进图论中的预测方法提供了一个全新的视角。
{"title":"DAGCN: hybrid model for efficiently handling joint node and link prediction in cloud workflows","authors":"Ruimin Ma,&nbsp;Junqi Gao,&nbsp;Li Cheng,&nbsp;Yuyi Zhang,&nbsp;Ovanes Petrosian","doi":"10.1007/s10489-024-05828-w","DOIUrl":"10.1007/s10489-024-05828-w","url":null,"abstract":"<div><p>In the cloud computing domain, significant strides have been made in performance prediction for cloud workflows, yet link prediction for cloud workflows remains largely unexplored. This paper introduces a novel challenge: joint node and link prediction in cloud workflows, with the aim of increasing the efficiency and overall performance of cloud computing resources. GNN-based methods have gained traction in handling graph-related tasks. The unique format of the DAG presents an underexplored area for GNNs effectiveness. To enhance comprehension of intricate graph structures and interrelationships, this paper introduces two novel models under the DAGCN framework: DAG-ConvGCN and DAG-AttGCN. The former synergizes the local receptive fields of the CNN with the global interpretive power of the GCN, whereas the latter integrates an attention mechanism to dynamically weigh the significance of node adjacencies. Through rigorous experimentation on a meticulously crafted joint node and link prediction task utilizing the Cluster-trace-v2018 dataset, both DAG-ConvGCN and DAG-AttGCN demonstrate superior performance over a spectrum of established machine learning and deep learning benchmarks. Moreover, the application of similarity measures such as the propagation kernel and the innovative GRBF kernel-which merges the graphlet kernel with the radial basis function kernel to accentuate graph topology and node features-reinforces the superiority of DAGCN models over graph-level prediction accuracy conventional baselines. This paper offers a fresh vantage point for advancing predictive methodologies within graph theory.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 23","pages":"12505 - 12530"},"PeriodicalIF":3.4,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142265436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive multimodal prompt for human-object interaction with local feature enhanced transformer 利用局部特征增强变换器实现人机交互的自适应多模态提示
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1007/s10489-024-05774-7
Kejun Xue, Yongbin Gao, Zhijun Fang, Xiaoyan Jiang, Wenjun Yu, Mingxuan Chen, Chenmou Wu

Human-object interaction (HOI) detection is an important computer vision task for recognizing the interaction between humans and surrounding objects in an image or video. The HOI datasets have a serious long-tailed data distribution problem because it is challenging to have a dataset that contains all potential interactions. Many HOI detectors have addressed this issue by utilizing visual-language models. However, due to the calculation mechanism of the Transformer, the visual-language model is not good at extracting the local features of input samples. Therefore, we propose a novel local feature enhanced Transformer to motivate encoders to extract multi-modal features that contain more information. Moreover, it is worth noting that the application of prompt learning in HOI detection is still in preliminary stages. Consequently, we propose a multi-modal adaptive prompt module, which uses an adaptive learning strategy to facilitate the interaction of language and visual prompts. In the HICO-DET and SWIG-HOI datasets, the proposed model achieves full interaction with 24.21% mAP and 14.29% mAP, respectively. Our code is available at https://github.com/small-code-cat/AMP-HOI.

人-物互动(HOI)检测是一项重要的计算机视觉任务,用于识别图像或视频中人与周围物体之间的互动。HOI 数据集存在严重的长尾数据分布问题,因为拥有一个包含所有潜在交互的数据集是一项挑战。许多 HOI 检测器利用视觉语言模型解决了这一问题。然而,由于变换器的计算机制,视觉语言模型并不能很好地提取输入样本的局部特征。因此,我们提出了一种新颖的局部特征增强变换器,以激励编码器提取包含更多信息的多模态特征。此外,值得注意的是,及时学习在 HOI 检测中的应用仍处于初级阶段。因此,我们提出了多模态自适应提示模块,该模块使用自适应学习策略来促进语言和视觉提示的交互。在 HICO-DET 和 SWIG-HOI 数据集中,所提出的模型分别以 24.21% 的 mAP 和 14.29% 的 mAP 实现了完全交互。我们的代码见 https://github.com/small-code-cat/AMP-HOI。
{"title":"Adaptive multimodal prompt for human-object interaction with local feature enhanced transformer","authors":"Kejun Xue,&nbsp;Yongbin Gao,&nbsp;Zhijun Fang,&nbsp;Xiaoyan Jiang,&nbsp;Wenjun Yu,&nbsp;Mingxuan Chen,&nbsp;Chenmou Wu","doi":"10.1007/s10489-024-05774-7","DOIUrl":"10.1007/s10489-024-05774-7","url":null,"abstract":"<div><p>Human-object interaction (HOI) detection is an important computer vision task for recognizing the interaction between humans and surrounding objects in an image or video. The HOI datasets have a serious long-tailed data distribution problem because it is challenging to have a dataset that contains all potential interactions. Many HOI detectors have addressed this issue by utilizing visual-language models. However, due to the calculation mechanism of the Transformer, the visual-language model is not good at extracting the local features of input samples. Therefore, we propose a novel local feature enhanced Transformer to motivate encoders to extract multi-modal features that contain more information. Moreover, it is worth noting that the application of prompt learning in HOI detection is still in preliminary stages. Consequently, we propose a multi-modal adaptive prompt module, which uses an adaptive learning strategy to facilitate the interaction of language and visual prompts. In the HICO-DET and SWIG-HOI datasets, the proposed model achieves full interaction with 24.21% mAP and 14.29% mAP, respectively. Our code is available at https://github.com/small-code-cat/AMP-HOI.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 23","pages":"12492 - 12504"},"PeriodicalIF":3.4,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142265435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved KD-tree based imbalanced big data classification and oversampling for MapReduce platforms 为 MapReduce 平台改进基于 KD 树的不平衡大数据分类和超采样
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1007/s10489-024-05763-w
William C. Sleeman IV, Martha Roseberry, Preetam Ghosh, Alberto Cano, Bartosz Krawczyk

In the era of big data, it is necessary to provide novel and efficient platforms for training machine learning models over large volumes of data. The MapReduce approach and its Apache Spark implementation are among the most popular methods that provide high-performance computing for classification algorithms. However, they require dedicated implementations that will take advantage of such architectures. Additionally, many real-world big data problems are plagued by class imbalance, posing challenges to the classifier training step. Existing solutions for alleviating skewed distributions do not work well in the MapReduce environment. In this paper, we propose a novel KD-tree based classifier, together with a variation of the SMOTE algorithm dedicated to the Spark platform. Our algorithms offer excellent predictive power and can work simultaneously with binary and multi-class imbalanced data. Exhaustive experiments conducted using the Amazon Web Service platform showcase the high efficiency and flexibility of our proposed algorithms.

在大数据时代,有必要为在大量数据中训练机器学习模型提供新颖而高效的平台。MapReduce 方法及其 Apache Spark 实现是为分类算法提供高性能计算的最流行方法之一。不过,它们需要专门的实现,以利用此类架构的优势。此外,现实世界中的许多大数据问题都受到类不平衡的困扰,这给分类器训练步骤带来了挑战。现有的缓解偏斜分布的解决方案在 MapReduce 环境中效果不佳。在本文中,我们提出了一种基于 KD 树的新型分类器,以及一种专用于 Spark 平台的 SMOTE 算法变体。我们的算法具有出色的预测能力,可同时处理二元和多类不平衡数据。使用亚马逊网络服务平台进行的详尽实验展示了我们提出的算法的高效性和灵活性。
{"title":"Improved KD-tree based imbalanced big data classification and oversampling for MapReduce platforms","authors":"William C. Sleeman IV,&nbsp;Martha Roseberry,&nbsp;Preetam Ghosh,&nbsp;Alberto Cano,&nbsp;Bartosz Krawczyk","doi":"10.1007/s10489-024-05763-w","DOIUrl":"10.1007/s10489-024-05763-w","url":null,"abstract":"<div><p>In the era of big data, it is necessary to provide novel and efficient platforms for training machine learning models over large volumes of data. The MapReduce approach and its Apache Spark implementation are among the most popular methods that provide high-performance computing for classification algorithms. However, they require dedicated implementations that will take advantage of such architectures. Additionally, many real-world big data problems are plagued by class imbalance, posing challenges to the classifier training step. Existing solutions for alleviating skewed distributions do not work well in the MapReduce environment. In this paper, we propose a novel KD-tree based classifier, together with a variation of the SMOTE algorithm dedicated to the Spark platform. Our algorithms offer excellent predictive power and can work simultaneously with binary and multi-class imbalanced data. Exhaustive experiments conducted using the Amazon Web Service platform showcase the high efficiency and flexibility of our proposed algorithms.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 23","pages":"12558 - 12575"},"PeriodicalIF":3.4,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142265136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-objective optimization enabling CFRP energy-efficient milling based on deep reinforcement learning 基于深度强化学习的多目标优化,实现 CFRP 节能铣削
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1007/s10489-024-05800-8
Meihang Zhang, Hua Zhang, Wei Yan, Lin Zhang, Zhigang Jiang

The expanding application of Carbon Fiber Reinforced Polymer (CFRP) in industries is drawing increasing attention to energy efficiency improvement and cost reducing during the secondary processing, particularly in milling. Machining parameter optimization is a practical and economical way to achieve this goal. However, the unclear milling mechanism and dynamic machining conditions of CFRP make it challenging. To fill this gap, this paper proposes a DRL-based approach that integrates physics-guided Transformer networks with Twin Delayed Deep Deterministic Policy Gradient (PGTTD3) to optimize CFRP milling parameters with multi-objectives. Firstly, a PG-Transformer-based CFRP milling energy consumption model is proposed, which modifies the existing De-stationary Attention module by integrating external physical variables to enhance modeling accuracy and efficiency. Secondly, a multi-objective optimization model considering energy consumption, milling time and machining cost for CFRP milling is formulated and mapped to a Markov Decision Process, and a reward function is designed. Thirdly, a PGTTD3 approach is proposed for dynamic parameter decision-making, incorporating a time difference strategy to enhance agent training stability and online adjustment reliability. The experimental results show that the proposed method reduces energy consumption, milling time and machining cost by 10.98%, 3.012%, and 14.56% in CFRP milling respectively, compared to the actual averages. The proposed algorithm exhibits excellent performance metrics when compared to state-of-the-art optimization algorithms, with an average improvement in optimization efficiency of over 20% and a maximum enhancement of 88.66%.

随着碳纤维增强聚合物(CFRP)在工业中应用的不断扩大,在二次加工(尤其是铣削加工)过程中提高能效和降低成本日益受到关注。加工参数优化是实现这一目标的实用而经济的方法。然而,CFRP 的铣削机理和动态加工条件并不明确,这给优化工作带来了挑战。为了填补这一空白,本文提出了一种基于 DRL 的方法,该方法将物理引导变压器网络与双延迟深度确定性策略梯度(PGTTD3)相结合,以优化 CFRP 的多目标铣削参数。首先,提出了基于 PG-Transformer 的 CFRP 铣削能耗模型,该模型通过集成外部物理变量对现有的去静态注意模块进行了修改,以提高建模精度和效率。其次,建立了考虑 CFRP 铣削能耗、铣削时间和加工成本的多目标优化模型,并将其映射为马尔可夫决策过程,设计了奖励函数。第三,提出了一种用于动态参数决策的 PGTTD3 方法,并结合时差策略提高了代理训练的稳定性和在线调整的可靠性。实验结果表明,与实际平均值相比,所提出的方法在 CFRP 铣削中分别降低了 10.98%、3.012% 和 14.56%的能耗、铣削时间和加工成本。与最先进的优化算法相比,所提出的算法表现出优异的性能指标,优化效率平均提高了 20% 以上,最高提高了 88.66%。
{"title":"Multi-objective optimization enabling CFRP energy-efficient milling based on deep reinforcement learning","authors":"Meihang Zhang,&nbsp;Hua Zhang,&nbsp;Wei Yan,&nbsp;Lin Zhang,&nbsp;Zhigang Jiang","doi":"10.1007/s10489-024-05800-8","DOIUrl":"10.1007/s10489-024-05800-8","url":null,"abstract":"<div><p>The expanding application of Carbon Fiber Reinforced Polymer (CFRP) in industries is drawing increasing attention to energy efficiency improvement and cost reducing during the secondary processing, particularly in milling. Machining parameter optimization is a practical and economical way to achieve this goal. However, the unclear milling mechanism and dynamic machining conditions of CFRP make it challenging. To fill this gap, this paper proposes a DRL-based approach that integrates physics-guided Transformer networks with Twin Delayed Deep Deterministic Policy Gradient (PGTTD3) to optimize CFRP milling parameters with multi-objectives. Firstly, a PG-Transformer-based CFRP milling energy consumption model is proposed, which modifies the existing De-stationary Attention module by integrating external physical variables to enhance modeling accuracy and efficiency. Secondly, a multi-objective optimization model considering energy consumption, milling time and machining cost for CFRP milling is formulated and mapped to a Markov Decision Process, and a reward function is designed. Thirdly, a PGTTD3 approach is proposed for dynamic parameter decision-making, incorporating a time difference strategy to enhance agent training stability and online adjustment reliability. The experimental results show that the proposed method reduces energy consumption, milling time and machining cost by 10.98%, 3.012%, and 14.56% in CFRP milling respectively, compared to the actual averages. The proposed algorithm exhibits excellent performance metrics when compared to state-of-the-art optimization algorithms, with an average improvement in optimization efficiency of over 20% and a maximum enhancement of 88.66%.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 23","pages":"12531 - 12557"},"PeriodicalIF":3.4,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142265135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1