首页 > 最新文献

IJCAI : proceedings of the conference最新文献

英文 中文
Predictive Modeling with Temporal Graphical Representation on Electronic Health Records. 利用电子健康记录的时态图形表示进行预测建模。
Pub Date : 2024-08-01 DOI: 10.24963/ijcai.2024/637
Jiayuan Chen, Changchang Yin, Yuanlong Wang, Ping Zhang

Deep learning-based predictive models, leveraging Electronic Health Records (EHR), are receiving increasing attention in healthcare. An effective representation of a patient's EHR should hierarchically encompass both the temporal relationships between historical visits and medical events, and the inherent structural information within these elements. Existing patient representation methods can be roughly categorized into sequential representation and graphical representation. The sequential representation methods focus only on the temporal relationships among longitudinal visits. On the other hand, the graphical representation approaches, while adept at extracting the graph-structured relationships between various medical events, fall short in effectively integrate temporal information. To capture both types of information, we model a patient's EHR as a novel temporal heterogeneous graph. This graph includes historical visits nodes and medical events nodes. It propagates structured information from medical event nodes to visit nodes and utilizes time-aware visit nodes to capture changes in the patient's health status. Furthermore, we introduce a novel temporal graph transformer (TRANS) that integrates temporal edge features, global positional encoding, and local structural encoding into heterogeneous graph convolution, capturing both temporal and structural information. We validate the effectiveness of TRANS through extensive experiments on three real-world datasets. The results show that our proposed approach achieves state-of-the-art performance.

利用电子健康记录(EHR)的基于深度学习的预测模型在医疗保健领域受到越来越多的关注。患者电子健康记录的有效表示方法应分层包含历史就诊和医疗事件之间的时间关系,以及这些元素中的固有结构信息。现有的患者表示方法可大致分为顺序表示法和图形表示法。顺序表示法只关注纵向就诊之间的时间关系。另一方面,图形表示法虽然善于提取各种医疗事件之间的图形结构关系,但却无法有效整合时间信息。为了捕捉这两类信息,我们将患者的电子病历建模为一个新颖的时间异构图。该图包括历史就诊节点和医疗事件节点。它将结构化信息从医疗事件节点传播到就诊节点,并利用时间感知就诊节点来捕捉患者健康状况的变化。此外,我们还引入了一种新型时序图转换器(TRANS),它将时序边缘特征、全局位置编码和局部结构编码整合到异构图卷积中,同时捕捉时序和结构信息。我们在三个真实世界数据集上进行了大量实验,验证了 TRANS 的有效性。结果表明,我们提出的方法达到了最先进的性能。
{"title":"Predictive Modeling with Temporal Graphical Representation on Electronic Health Records.","authors":"Jiayuan Chen, Changchang Yin, Yuanlong Wang, Ping Zhang","doi":"10.24963/ijcai.2024/637","DOIUrl":"10.24963/ijcai.2024/637","url":null,"abstract":"<p><p>Deep learning-based predictive models, leveraging Electronic Health Records (EHR), are receiving increasing attention in healthcare. An effective representation of a patient's EHR should hierarchically encompass both the temporal relationships between historical visits and medical events, and the inherent structural information within these elements. Existing patient representation methods can be roughly categorized into sequential representation and graphical representation. The sequential representation methods focus only on the temporal relationships among longitudinal visits. On the other hand, the graphical representation approaches, while adept at extracting the graph-structured relationships between various medical events, fall short in effectively integrate temporal information. To capture both types of information, we model a patient's EHR as a novel temporal heterogeneous graph. This graph includes historical visits nodes and medical events nodes. It propagates structured information from medical event nodes to visit nodes and utilizes time-aware visit nodes to capture changes in the patient's health status. Furthermore, we introduce a novel temporal graph transformer (TRANS) that integrates temporal edge features, global positional encoding, and local structural encoding into heterogeneous graph convolution, capturing both temporal and structural information. We validate the effectiveness of TRANS through extensive experiments on three real-world datasets. The results show that our proposed approach achieves state-of-the-art performance.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2024 ","pages":"5763-5771"},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11446542/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142367720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ReBandit: Random Effects Based Online RL Algorithm for Reducing Cannabis Use. ReBandit:基于随机效应的在线RL算法减少大麻使用。
Susobhan Ghosh, Yongyi Guo, Pei-Yao Hung, Lara Coughlin, Erin Bonar, Inbal Nahum-Shani, Maureen Walton, Susan Murphy

The escalating prevalence of cannabis use, and associated cannabis-use disorder (CUD), poses a significant public health challenge globally. With a notably wide treatment gap, especially among emerging adults (EAs; ages 18-25), addressing cannabis use and CUD remains a pivotal objective within the 2030 United Nations Agenda for Sustainable Development Goals (SDG). In this work, we develop an online reinforcement learning (RL) algorithm called reBandit which will be utilized in a mobile health study to deliver personalized mobile health interventions aimed at reducing cannabis use among EAs. reBandit utilizes random effects and informative Bayesian priors to learn quickly and efficiently in noisy mobile health environments. Moreover, reBandit employs Empirical Bayes and optimization techniques to autonomously update its hyper-parameters online. To evaluate the performance of our algorithm, we construct a simulation testbed using data from a prior study, and compare against commonly used algorithms in mobile health studies. We show that reBandit performs equally well or better than all the baseline algorithms, and the performance gap widens as population heterogeneity increases in the simulation environment, proving its adeptness to adapt to diverse population of study participants.

大麻使用和相关大麻使用障碍(CUD)的流行率不断攀升,对全球公共卫生构成了重大挑战。由于治疗差距明显,尤其是在新兴成年人(EAs;18-25 岁)中,解决大麻使用和 CUD 问题仍然是 2030 年联合国可持续发展目标(SDG)议程中的一个关键目标。在这项工作中,我们开发了一种名为 reBandit 的在线强化学习(RL)算法,该算法将用于一项移动健康研究,以提供个性化的移动健康干预措施,从而减少 EAs 中的大麻使用。此外,reBandit 还采用了经验贝叶斯和优化技术来在线自主更新其超参数。为了评估我们算法的性能,我们利用先前研究的数据构建了一个模拟测试平台,并与移动健康研究中常用的算法进行了比较。我们的结果表明,reBandit 的性能与所有基线算法相当,甚至更好,而且随着模拟环境中人群异质性的增加,性能差距也在扩大,这证明它能很好地适应多样化的研究参与者人群。
{"title":"ReBandit: Random Effects Based Online RL Algorithm for Reducing Cannabis Use.","authors":"Susobhan Ghosh, Yongyi Guo, Pei-Yao Hung, Lara Coughlin, Erin Bonar, Inbal Nahum-Shani, Maureen Walton, Susan Murphy","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>The escalating prevalence of cannabis use, and associated cannabis-use disorder (CUD), poses a significant public health challenge globally. With a notably wide treatment gap, especially among emerging adults (EAs; ages 18-25), addressing cannabis use and CUD remains a pivotal objective within the 2030 United Nations Agenda for Sustainable Development Goals (SDG). In this work, we develop an online reinforcement learning (RL) algorithm called reBandit which will be utilized in a mobile health study to deliver personalized mobile health interventions aimed at reducing cannabis use among EAs. reBandit utilizes <i>random effects</i> and <i>informative Bayesian priors</i> to learn quickly and efficiently in noisy mobile health environments. Moreover, reBandit employs Empirical Bayes and optimization techniques to autonomously update its hyper-parameters online. To evaluate the performance of our algorithm, we construct a simulation testbed using data from a prior study, and compare against commonly used algorithms in mobile health studies. We show that reBandit performs equally well or better than all the baseline algorithms, and the performance gap widens as population heterogeneity increases in the simulation environment, proving its adeptness to adapt to diverse population of study participants.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2024 ","pages":"7278-7286"},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11671148/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142904166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adapt to Adaptation: Learning Personalization for Cross-Silo Federated Learning. 适应适应:跨筒仓联邦学习的学习个性化。
Pub Date : 2022-07-01 DOI: 10.24963/ijcai.2022/301
Jun Luo, Shandong Wu

Conventional federated learning (FL) trains one global model for a federation of clients with decentralized data, reducing the privacy risk of centralized training. However, the distribution shift across non-IID datasets, often poses a challenge to this one-model-fits-all solution. Personalized FL aims to mitigate this issue systematically. In this work, we propose APPLE, a personalized cross-silo FL framework that adaptively learns how much each client can benefit from other clients' models. We also introduce a method to flexibly control the focus of training APPLE between global and local objectives. We empirically evaluate our method's convergence and generalization behaviors, and perform extensive experiments on two benchmark datasets and two medical imaging datasets under two non-IID settings. The results show that the proposed personalized FL framework, APPLE, achieves state-of-the-art performance compared to several other personalized FL approaches in the literature. The code is publicly available at https://github.com/ljaiverson/pFL-APPLE.

传统的联邦学习(FL)为具有分散数据的客户联邦训练一个全局模型,从而降低了集中训练的隐私风险。然而,跨非iid数据集的分布变化通常对这种“一模通万”的解决方案提出了挑战。个性化FL旨在系统地缓解这一问题。在这项工作中,我们提出了APPLE,这是一个个性化的跨竖井FL框架,可以自适应地学习每个客户端可以从其他客户端的模型中受益多少。本文还提出了一种在全局目标和局部目标之间灵活控制训练重点的方法。我们对该方法的收敛性和泛化性进行了实证评估,并在两个基准数据集和两个非iid设置下的医学成像数据集上进行了大量实验。结果表明,与文献中其他几种个性化FL方法相比,所提出的个性化FL框架APPLE实现了最先进的性能。该代码可在https://github.com/ljaiverson/pFL-APPLE上公开获得。
{"title":"Adapt to Adaptation: Learning Personalization for Cross-Silo Federated Learning.","authors":"Jun Luo,&nbsp;Shandong Wu","doi":"10.24963/ijcai.2022/301","DOIUrl":"https://doi.org/10.24963/ijcai.2022/301","url":null,"abstract":"<p><p>Conventional federated learning (FL) trains one global model for a federation of clients with decentralized data, reducing the privacy risk of centralized training. However, the distribution shift across non-IID datasets, often poses a challenge to this one-model-fits-all solution. Personalized FL aims to mitigate this issue systematically. In this work, we propose APPLE, a personalized cross-silo FL framework that adaptively learns how much each client can benefit from other clients' models. We also introduce a method to flexibly control the focus of training APPLE between global and local objectives. We empirically evaluate our method's convergence and generalization behaviors, and perform extensive experiments on two benchmark datasets and two medical imaging datasets under two non-IID settings. The results show that the proposed personalized FL framework, APPLE, achieves state-of-the-art performance compared to several other personalized FL approaches in the literature. The code is publicly available at https://github.com/ljaiverson/pFL-APPLE.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2022 ","pages":"2166-2173"},"PeriodicalIF":0.0,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10184830/pdf/nihms-1891606.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9857387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Stabilizing and Enhancing Link Prediction through Deepened Graph Auto-Encoders. 通过深化图自动编码器稳定并增强链接预测。
Pub Date : 2022-07-01 DOI: 10.24963/ijcai.2022/498
Xinxing Wu, Qiang Cheng

Graph neural networks have been widely used for a variety of learning tasks. Link prediction is a relatively under-studied graph learning task, with current state-of-the-art models based on one- or two-layer shallow graph auto-encoder (GAE) architectures. In this paper, we overcome the limitation of current methods for link prediction of non-Euclidean network data, which can only use shallow GAEs and variational GAEs. Our proposed methods innovatively incorporate standard auto-encoders (AEs) into the architectures of GAEs to capitalize on the intimate coupling of node and edge information in complex network data. Empirically, extensive experiments on various datasets demonstrate the competitive performance of our proposed approach. Theoretically, we prove that our deep extensions can inclusively express multiple polynomial filters with different orders. The codes of this paper are available at https://github.com/xinxingwu-uk/DGAE.

图神经网络已被广泛用于各种学习任务。链接预测是一项研究相对较少的图学习任务,目前最先进的模型基于单层或双层浅层图自动编码器(GAE)架构。在本文中,我们克服了目前对非欧几里得网络数据进行链接预测的方法只能使用浅层 GAE 和变异 GAE 的局限性。我们提出的方法创新性地将标准自动编码器(AE)融入到 GAE 的架构中,以利用复杂网络数据中节点和边缘信息的紧密耦合。在各种数据集上进行的大量实验证明了我们提出的方法具有竞争力的性能。从理论上讲,我们证明了我们的深度扩展可以包容地表达多个不同阶数的多项式滤波器。本文代码见 https://github.com/xinxingwu-uk/DGAE。
{"title":"Stabilizing and Enhancing Link Prediction through Deepened Graph Auto-Encoders.","authors":"Xinxing Wu, Qiang Cheng","doi":"10.24963/ijcai.2022/498","DOIUrl":"10.24963/ijcai.2022/498","url":null,"abstract":"<p><p>Graph neural networks have been widely used for a variety of learning tasks. Link prediction is a relatively under-studied graph learning task, with current state-of-the-art models based on one- or two-layer shallow graph auto-encoder (GAE) architectures. In this paper, we overcome the limitation of current methods for link prediction of non-Euclidean network data, which can only use shallow GAEs and variational GAEs. Our proposed methods innovatively incorporate standard auto-encoders (AEs) into the architectures of GAEs to capitalize on the intimate coupling of node and edge information in complex network data. Empirically, extensive experiments on various datasets demonstrate the competitive performance of our proposed approach. Theoretically, we prove that our deep extensions can inclusively express multiple polynomial filters with different orders. The codes of this paper are available at https://github.com/xinxingwu-uk/DGAE.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2022 ","pages":"3587-3593"},"PeriodicalIF":0.0,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9754798/pdf/nihms-1834388.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10767141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RCA: A Deep Collaborative Autoencoder Approach for Anomaly Detection RCA:一种深度协同自编码器异常检测方法
Pub Date : 2021-08-01 DOI: 10.24963/ijcai.2021/208
Boyang Liu, Ding Wang, Kaixiang Lin, P. Tan, Jiayu Zhou
Unsupervised anomaly detection (AD) plays a crucial role in many critical applications. Driven by the success of deep learning, recent years have witnessed growing interest in applying deep neural networks (DNNs) to AD problems. A common approach is using autoencoders to learn a feature representation for the normal observations in the data. The reconstruction error of the autoencoder is then used as outlier score to detect the anomalies. However, due to the high complexity brought upon by over-parameterization of DNNs, the reconstruction error of the anomalies could also be small, which hampers the effectiveness of these methods. To alleviate this problem, we propose a robust framework using collaborative autoencoders to jointly identify normal observations from the data while learning its feature representation. We investigate the theoretical properties of the framework and empirically show its outstanding performance as compared to other DNN-based methods. Empirical results also show resiliency of the framework to missing values compared to other baseline methods.
无监督异常检测(AD)在许多关键应用中起着至关重要的作用。在深度学习成功的推动下,近年来人们对将深度神经网络(dnn)应用于AD问题越来越感兴趣。一种常见的方法是使用自动编码器来学习数据中正常观测值的特征表示。然后将自编码器的重建误差作为异常值来检测异常。然而,由于深度神经网络的过度参数化带来的高度复杂性,异常的重建误差也可能很小,从而影响了这些方法的有效性。为了缓解这个问题,我们提出了一个使用协作自编码器的鲁棒框架,在学习其特征表示的同时,共同识别数据中的正常观测值。我们研究了该框架的理论性质,并通过经验证明了与其他基于dnn的方法相比,该框架具有出色的性能。实证结果还表明,与其他基线方法相比,该框架对缺失值具有弹性。
{"title":"RCA: A Deep Collaborative Autoencoder Approach for Anomaly Detection","authors":"Boyang Liu, Ding Wang, Kaixiang Lin, P. Tan, Jiayu Zhou","doi":"10.24963/ijcai.2021/208","DOIUrl":"https://doi.org/10.24963/ijcai.2021/208","url":null,"abstract":"Unsupervised anomaly detection (AD) plays a crucial role in many critical applications. Driven by the success of deep learning, recent years have witnessed growing interest in applying deep neural networks (DNNs) to AD problems. A common approach is using autoencoders to learn a feature representation for the normal observations in the data. The reconstruction error of the autoencoder is then used as outlier score to detect the anomalies. However, due to the high complexity brought upon by over-parameterization of DNNs, the reconstruction error of the anomalies could also be small, which hampers the effectiveness of these methods. To alleviate this problem, we propose a robust framework using collaborative autoencoders to jointly identify normal observations from the data while learning its feature representation. We investigate the theoretical properties of the framework and empirically show its outstanding performance as compared to other DNN-based methods. Empirical results also show resiliency of the framework to missing values compared to other baseline methods.","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"9 1","pages":"1505-1511"},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79839905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Human Gaze Assisted Artificial Intelligence: A Review. 人类凝视辅助人工智能:综述。
Pub Date : 2020-07-01 DOI: 10.24963/ijcai.2020/689
Ruohan Zhang, Akanksha Saran, Bo Liu, Yifeng Zhu, Sihang Guo, Scott Niekum, Dana Ballard, Mary Hayhoe

Human gaze reveals a wealth of information about internal cognitive state. Thus, gaze-related research has significantly increased in computer vision, natural language processing, decision learning, and robotics in recent years. We provide a high-level overview of the research efforts in these fields, including collecting human gaze data sets, modeling gaze behaviors, and utilizing gaze information in various applications, with the goal of enhancing communication between these research areas. We discuss future challenges and potential applications that work towards a common goal of human-centered artificial intelligence.

人类的凝视揭示了关于内部认知状态的丰富信息。因此,近年来,与凝视相关的研究在计算机视觉、自然语言处理、决策学习和机器人领域显著增加。我们对这些领域的研究工作进行了高层次的概述,包括收集人类凝视数据集、建模凝视行为以及在各种应用中利用凝视信息,目的是加强这些研究领域之间的沟通。我们讨论了未来的挑战和潜在的应用,以实现以人为中心的人工智能的共同目标。
{"title":"Human Gaze Assisted Artificial Intelligence: A Review.","authors":"Ruohan Zhang,&nbsp;Akanksha Saran,&nbsp;Bo Liu,&nbsp;Yifeng Zhu,&nbsp;Sihang Guo,&nbsp;Scott Niekum,&nbsp;Dana Ballard,&nbsp;Mary Hayhoe","doi":"10.24963/ijcai.2020/689","DOIUrl":"10.24963/ijcai.2020/689","url":null,"abstract":"<p><p>Human gaze reveals a wealth of information about internal cognitive state. Thus, gaze-related research has significantly increased in computer vision, natural language processing, decision learning, and robotics in recent years. We provide a high-level overview of the research efforts in these fields, including collecting human gaze data sets, modeling gaze behaviors, and utilizing gaze information in various applications, with the goal of enhancing communication between these research areas. We discuss future challenges and potential applications that work towards a common goal of human-centered artificial intelligence.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2020 ","pages":"4951-4958"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7476326/pdf/nihms-1622212.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38356030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Improving Attention Mechanism in Graph Neural Networks via Cardinality Preservation. 基于基数保留的图神经网络注意机制改进。
Pub Date : 2020-07-01 DOI: 10.24963/ijcai.2020/194
Shuo Zhang, Lei Xie

Graph Neural Networks (GNNs) are powerful for the representation learning of graph-structured data. Most of the GNNs use a message-passing scheme, where the embedding of a node is iteratively updated by aggregating the information from its neighbors. To achieve a better expressive capability of node influences, attention mechanism has grown to be popular to assign trainable weights to the nodes in aggregation. Though the attention-based GNNs have achieved remarkable results in various tasks, a clear understanding of their discriminative capacities is missing. In this work, we present a theoretical analysis of the representational properties of the GNN that adopts the attention mechanism as an aggregator. Our analysis determines all cases when those attention-based GNNs can always fail to distinguish certain distinct structures. Those cases appear due to the ignorance of cardinality information in attention-based aggregation. To improve the performance of attention-based GNNs, we propose cardinality preserved attention (CPA) models that can be applied to any kind of attention mechanisms. Our experiments on node and graph classification confirm our theoretical analysis and show the competitive performance of our CPA models. The code is available online: https://github.com/zetayue/CPA.

图神经网络(gnn)对于图结构数据的表示学习具有强大的功能。大多数gnn使用消息传递方案,其中节点的嵌入通过聚合其邻居的信息来迭代更新。为了更好地表达节点的影响,关注机制在聚合中为节点分配可训练的权值。尽管基于注意力的gnn在各种任务中取得了显著的成果,但对其判别能力的清晰认识仍然缺失。在这项工作中,我们对采用注意力机制作为聚合器的GNN的表征特性进行了理论分析。我们的分析确定了那些基于注意力的gnn总是无法区分某些不同结构的所有情况。这些情况的出现是由于在基于注意力的聚合中忽略了基数信息。为了提高基于注意的gnn的性能,我们提出了基数保持注意(CPA)模型,该模型可以应用于任何类型的注意机制。我们在节点和图分类上的实验证实了我们的理论分析,并展示了我们的CPA模型的竞争性能。代码可在网上获得:https://github.com/zetayue/CPA。
{"title":"Improving Attention Mechanism in Graph Neural Networks via Cardinality Preservation.","authors":"Shuo Zhang,&nbsp;Lei Xie","doi":"10.24963/ijcai.2020/194","DOIUrl":"https://doi.org/10.24963/ijcai.2020/194","url":null,"abstract":"<p><p>Graph Neural Networks (GNNs) are powerful for the representation learning of graph-structured data. Most of the GNNs use a message-passing scheme, where the embedding of a node is iteratively updated by aggregating the information from its neighbors. To achieve a better expressive capability of node influences, attention mechanism has grown to be popular to assign trainable weights to the nodes in aggregation. Though the attention-based GNNs have achieved remarkable results in various tasks, a clear understanding of their discriminative capacities is missing. In this work, we present a theoretical analysis of the representational properties of the GNN that adopts the attention mechanism as an aggregator. Our analysis determines all cases when those attention-based GNNs can always fail to distinguish certain distinct structures. Those cases appear due to the ignorance of cardinality information in attention-based aggregation. To improve the performance of attention-based GNNs, we propose cardinality preserved attention (CPA) models that can be applied to any kind of attention mechanisms. Our experiments on node and graph classification confirm our theoretical analysis and show the competitive performance of our CPA models. The code is available online: https://github.com/zetayue/CPA.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2020 ","pages":"1395-1402"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7416665/pdf/nihms-1615275.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38262582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Mixed-Variable Bayesian Optimization 混合变量贝叶斯优化
Pub Date : 2020-01-01 DOI: 10.24963/ijcai.2020/365
Erik A. Daxberger, Anastasia Makarova, M. Turchetta, Andreas Krause
The optimization of expensive to evaluate, black-box, mixed-variable functions, i.e. functions that have continuous and discrete inputs, is a difficult and yet pervasive problem in science and engineering. In Bayesian optimization (BO), special cases of this problem that consider fully continuous or fully discrete domains have been widely studied. However, few methods exist for mixed-variable domains and none of them can handle discrete constraints that arise in many real-world applications. In this paper, we introduce MiVaBo, a novel BO algorithm for the efficient optimization of mixed-variable functions combining a linear surrogate model based on expressive feature representations with Thompson sampling. We propose an effective method to optimize its acquisition function, a challenging problem for mixed-variable domains, making MiVaBo the first BO method that can handle complex constraints over the discrete variables. Moreover, we provide the first convergence analysis of a mixed-variable BO algorithm. Finally, we show that MiVaBo is significantly more sample efficient than state-of-the-art mixed-variable BO algorithms on several hyperparameter tuning tasks, including the tuning of deep generative models.
对具有连续和离散输入的昂贵的黑盒混合变量函数(即具有连续和离散输入的函数)的优化是科学和工程中一个困难但普遍存在的问题。在贝叶斯优化(BO)中,考虑完全连续或完全离散域的贝叶斯优化问题的特殊情况已经得到了广泛的研究。然而,针对混合变量域的方法很少,而且没有一种方法可以处理许多实际应用中出现的离散约束。在本文中,我们介绍了一种将基于表达特征表示的线性代理模型与汤普森采样相结合的混合变量函数高效优化的新型BO算法MiVaBo。我们提出了一种有效的方法来优化其获取函数,这是混合变量领域的一个具有挑战性的问题,使MiVaBo成为第一个可以处理离散变量上的复杂约束的BO方法。此外,我们提供了混合变量BO算法的第一个收敛性分析。最后,我们证明了MiVaBo在几个超参数调优任务上比最先进的混合变量BO算法具有更高的样本效率,包括深度生成模型的调优。
{"title":"Mixed-Variable Bayesian Optimization","authors":"Erik A. Daxberger, Anastasia Makarova, M. Turchetta, Andreas Krause","doi":"10.24963/ijcai.2020/365","DOIUrl":"https://doi.org/10.24963/ijcai.2020/365","url":null,"abstract":"The optimization of expensive to evaluate, black-box, mixed-variable functions, i.e. functions that have continuous and discrete inputs, is a difficult and yet pervasive problem in science and engineering. In Bayesian optimization (BO), special cases of this problem that consider fully continuous or fully discrete domains have been widely studied. However, few methods exist for mixed-variable domains and none of them can handle discrete constraints that arise in many real-world applications. In this paper, we introduce MiVaBo, a novel BO algorithm for the efficient optimization of mixed-variable functions combining a linear surrogate model based on expressive feature representations with Thompson sampling. We propose an effective method to optimize its acquisition function, a challenging problem for mixed-variable domains, making MiVaBo the first BO method that can handle complex constraints over the discrete variables. Moreover, we provide the first convergence analysis of a mixed-variable BO algorithm. Finally, we show that MiVaBo is significantly more sample efficient than state-of-the-art mixed-variable BO algorithms on several hyperparameter tuning tasks, including the tuning of deep generative models.","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"25 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86161869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
What Does the Evidence Say? Models to Help Make Sense of the Biomedical Literature. 证据怎么说?帮助理解生物医学文献的模型。
Pub Date : 2019-08-01 DOI: 10.24963/ijcai.2019/899
Byron C Wallace

Ideally decisions regarding medical treatments would be informed by the totality of the available evidence. The best evidence we currently have is in published natural language articles describing the conduct and results of clinical trials. Because these are unstructured, it is difficult for domain experts (e.g., physicians) to sort through and appraise the evidence pertaining to a given clinical question. Natural language technologies have the potential to improve access to the evidence via semi-automated processing of the biomedical literature. In this brief paper I highlight work on developing tasks, corpora, and models to support semi-automated evidence retrieval and extraction. The aim is to design models that can consume articles describing clinical trials and automatically extract from these key clinical variables and findings, and estimate their reliability. Completely automating 'machine reading' of evidence remains a distant aim given current technologies; the more immediate hope is to use such technologies to help domain experts access and make sense of unstructured biomedical evidence more efficiently, with the ultimate aim of improving patient care. Aside from their practical importance, these tasks pose core NLP challenges that directly motivate methodological innovation.

理想情况下,有关医疗的决定应以现有的全部证据为依据。我们目前拥有的最佳证据是发表在描述临床试验的行为和结果的自然语言文章中。因为这些是非结构化的,领域专家(例如,医生)很难整理和评估与给定临床问题有关的证据。自然语言技术有可能通过对生物医学文献的半自动化处理来改善对证据的获取。在这篇简短的文章中,我重点介绍了开发任务、语料库和模型以支持半自动证据检索和提取的工作。目的是设计模型,可以使用描述临床试验的文章,并自动从这些关键的临床变量和发现中提取,并估计其可靠性。鉴于目前的技术,完全自动化的证据“机器阅读”仍然是一个遥远的目标;更直接的希望是利用这些技术帮助领域专家更有效地获取和理解非结构化的生物医学证据,最终目的是改善患者护理。除了它们的实际重要性外,这些任务还构成了直接激发方法创新的核心NLP挑战。
{"title":"What Does the Evidence Say? Models to Help Make Sense of the Biomedical Literature.","authors":"Byron C Wallace","doi":"10.24963/ijcai.2019/899","DOIUrl":"https://doi.org/10.24963/ijcai.2019/899","url":null,"abstract":"<p><p>Ideally decisions regarding medical treatments would be informed by the totality of the available evidence. The best evidence we currently have is in published natural language articles describing the conduct and results of clinical trials. Because these are unstructured, it is difficult for domain experts (e.g., physicians) to sort through and appraise the evidence pertaining to a given clinical question. Natural language technologies have the potential to improve access to the evidence via semi-automated processing of the biomedical literature. In this brief paper I highlight work on developing tasks, corpora, and models to support semi-automated evidence retrieval and extraction. The aim is to design models that can consume articles describing clinical trials and automatically extract from these key clinical variables and findings, and estimate their reliability. Completely automating 'machine reading' of evidence remains a distant aim given current technologies; the more immediate hope is to use such technologies to help domain experts access and make sense of unstructured biomedical evidence more efficiently, with the ultimate aim of improving patient care. Aside from their practical importance, these tasks pose core NLP challenges that directly motivate methodological innovation.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2019 ","pages":"6416-6420"},"PeriodicalIF":0.0,"publicationDate":"2019-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8136417/pdf/nihms-1066622.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39023271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
DDL: Deep Dictionary Learning for Predictive Phenotyping. DDL:用于预测表型的深度字典学习。
Pub Date : 2019-08-01 DOI: 10.24963/ijcai.2019/812
Tianfan Fu, Trong Nghia Hoang, Cao Xiao, Jimeng Sun

Predictive phenotyping is about accurately predicting what phenotypes will occur in the next clinical visit based on longitudinal Electronic Health Record (EHR) data. While deep learning (DL) models have recently demonstrated strong performance in predictive phenotyping, they require access to a large amount of labeled data, which are expensive to acquire. To address this label-insufficient challenge, we propose a deep dictionary learning framework (DDL) for phenotyping, which utilizes unlabeled data as a complementary source of information to generate a better, more succinct data representation. Our empirical evaluations on multiple EHR datasets demonstrated that DDL outperforms the existing predictive phenotyping methods on a wide variety of clinical tasks that require patient phenotyping. The results also show that unlabeled data can be used to generate better data representation that helps improve DDL's phenotyping performance over existing methods that only uses labeled data.

预测性表型是指根据纵向电子健康记录(EHR)数据,准确预测下一次临床就诊中会出现哪些表型。虽然深度学习(DL)模型最近在预测表型方面表现出了强劲的性能,但它们需要访问大量的标记数据,而获取这些数据的成本很高。为了解决这一标签不足的难题,我们提出了一种用于表型分析的深度字典学习框架(DDL),该框架利用无标签数据作为补充信息源,生成更好、更简洁的数据表示。我们在多个电子病历数据集上进行的实证评估表明,在需要对患者进行表型分析的各种临床任务中,DDL 的表现优于现有的预测性表型分析方法。结果还表明,非标记数据可用于生成更好的数据表示,从而有助于提高 DDL 的表型分析性能,超过仅使用标记数据的现有方法。
{"title":"DDL: Deep Dictionary Learning for Predictive Phenotyping.","authors":"Tianfan Fu, Trong Nghia Hoang, Cao Xiao, Jimeng Sun","doi":"10.24963/ijcai.2019/812","DOIUrl":"10.24963/ijcai.2019/812","url":null,"abstract":"<p><p>Predictive phenotyping is about accurately predicting what phenotypes will occur in the next clinical visit based on longitudinal Electronic Health Record (EHR) data. While deep learning (DL) models have recently demonstrated strong performance in predictive phenotyping, they require access to a large amount of labeled data, which are expensive to acquire. To address this label-insufficient challenge, we propose a deep dictionary learning framework (DDL) for phenotyping, which utilizes unlabeled data as a complementary source of information to generate a better, more succinct data representation. Our empirical evaluations on multiple EHR datasets demonstrated that DDL outperforms the existing predictive phenotyping methods on a wide variety of clinical tasks that require patient phenotyping. The results also show that unlabeled data can be used to generate better data representation that helps improve DDL's phenotyping performance over existing methods that only uses labeled data.</p>","PeriodicalId":73334,"journal":{"name":"IJCAI : proceedings of the conference","volume":"2019 ","pages":"5857-5863"},"PeriodicalIF":0.0,"publicationDate":"2019-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7990269/pdf/nihms-1675238.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25517126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IJCAI : proceedings of the conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1