首页 > 最新文献

2021 IEEE International Conference on Big Knowledge (ICBK)最新文献

英文 中文
Intervention Prediction for Patients with Pressure Injury Using Random Forest 基于随机森林的压力性损伤干预预测
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00072
Liuqi Jin, Yan Pan, Jiaoyun Yang, Lin Han, Lin Lv, Miki Raviv, Ning An
Pressure injury (PI) is one of the major causes of short-term death. Early intervention for patients at risk plays an essential role in PI. However, many nurses may ignore risks. This paper aims to establish a model to predict interventions according to the patient's physical signs, which can help nurses develop care plans. We used data from 1,483 patients with 25 characteristics and 17 interventions. We use the Random Forest and Particle Swarm Optimization (PSO) to optimize model parameters. Then we compared it with KNN, SVM, and Decision Tree. The 10-fold cross-validation result showed that the Random Forest has better accuracy than other methods, with an f1 score of 0.84. This finding proved the feasibility of using machine learning to help formulate care plans according to the classification of index prediction results. Our model shows that hemoglobin, Braden PI score, and age are the three most influential risk factors.
压伤(PI)是短期死亡的主要原因之一。高危患者的早期干预在PI中起着至关重要的作用。然而,许多护士可能会忽视风险。本文旨在建立一个模型,根据患者的身体体征预测干预措施,以帮助护士制定护理计划。我们使用了来自1483名患者的数据,这些患者有25个特征和17种干预措施。我们使用随机森林和粒子群优化(PSO)来优化模型参数。然后将其与KNN、SVM和Decision Tree进行比较。10倍交叉验证结果表明Random Forest的准确率优于其他方法,f1得分为0.84。这一发现证明了根据指标预测结果的分类,利用机器学习帮助制定护理计划的可行性。我们的模型显示血红蛋白、Braden PI评分和年龄是三个最具影响的危险因素。
{"title":"Intervention Prediction for Patients with Pressure Injury Using Random Forest","authors":"Liuqi Jin, Yan Pan, Jiaoyun Yang, Lin Han, Lin Lv, Miki Raviv, Ning An","doi":"10.1109/ICKG52313.2021.00072","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00072","url":null,"abstract":"Pressure injury (PI) is one of the major causes of short-term death. Early intervention for patients at risk plays an essential role in PI. However, many nurses may ignore risks. This paper aims to establish a model to predict interventions according to the patient's physical signs, which can help nurses develop care plans. We used data from 1,483 patients with 25 characteristics and 17 interventions. We use the Random Forest and Particle Swarm Optimization (PSO) to optimize model parameters. Then we compared it with KNN, SVM, and Decision Tree. The 10-fold cross-validation result showed that the Random Forest has better accuracy than other methods, with an f1 score of 0.84. This finding proved the feasibility of using machine learning to help formulate care plans according to the classification of index prediction results. Our model shows that hemoglobin, Braden PI score, and age are the three most influential risk factors.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133057650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Topic-Guided Knowledge Graph Construction for Argument Mining 基于主题的知识图谱构建
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00049
Weichen Li, Patrick Abels, Zahra Ahmadi, Sophie Burkhardt, Benjamin Schiller, Iryna Gurevych, S. Kramer
Decision-making tasks usually follow five steps: identifying the problem, collecting data, extracting evidence, iden-tifying arguments, and making the decision. This paper focuses on two steps of decision-making: extracting evidence by building knowledge graphs (KGs) of specialized topics and identifying sentences' arguments through sentence-level argument mining. We present a hybrid model that combines topic modeling using latent Dirichlet allocation (LDA) and word embeddings to obtain external knowledge from structured and unstructured data. We use a topic model to extract topic- and sentence-specific evidence from the structured knowledge base Wikidata. A knowledge graph is constructed based on the cosine similarity between the entity word vectors of Wikidata and the vector of the given sentence. A second graph based on topic-specific articles found via Google supplements the general incompleteness of the structured knowledge base. Combining these graphs, we obtain a graph-based model that, as our evaluation shows, successfully capitalizes on both structured and unstructured data.
决策任务通常遵循五个步骤:识别问题,收集数据,提取证据,识别论点,做出决定。本文重点研究了决策的两个步骤:通过构建专业主题的知识图来提取证据,通过句子级的论据挖掘来识别句子的论点。我们提出了一个混合模型,该模型结合了使用潜在狄利克雷分配(LDA)和词嵌入的主题建模,从结构化和非结构化数据中获取外部知识。我们使用主题模型从结构化知识库Wikidata中提取主题和句子特定的证据。基于维基数据实体词向量与给定句子向量之间的余弦相似度构建知识图。第二个图基于通过谷歌找到的特定主题文章,补充了结构化知识库的一般不完备性。结合这些图,我们得到了一个基于图的模型,正如我们的评估所示,它成功地利用了结构化和非结构化数据。
{"title":"Topic-Guided Knowledge Graph Construction for Argument Mining","authors":"Weichen Li, Patrick Abels, Zahra Ahmadi, Sophie Burkhardt, Benjamin Schiller, Iryna Gurevych, S. Kramer","doi":"10.1109/ICKG52313.2021.00049","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00049","url":null,"abstract":"Decision-making tasks usually follow five steps: identifying the problem, collecting data, extracting evidence, iden-tifying arguments, and making the decision. This paper focuses on two steps of decision-making: extracting evidence by building knowledge graphs (KGs) of specialized topics and identifying sentences' arguments through sentence-level argument mining. We present a hybrid model that combines topic modeling using latent Dirichlet allocation (LDA) and word embeddings to obtain external knowledge from structured and unstructured data. We use a topic model to extract topic- and sentence-specific evidence from the structured knowledge base Wikidata. A knowledge graph is constructed based on the cosine similarity between the entity word vectors of Wikidata and the vector of the given sentence. A second graph based on topic-specific articles found via Google supplements the general incompleteness of the structured knowledge base. Combining these graphs, we obtain a graph-based model that, as our evaluation shows, successfully capitalizes on both structured and unstructured data.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117125095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Jointly Modeling Fact Triples and Text Information for Knowledge Base Completion 面向知识库补全的事实三元组和文本信息联合建模
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00037
Xiuxing Li, Zhenyu Li, Zhichao Duan, Jiacheng Xu, Ning Liu, Jianyong Wang
Knowledge bases become essential resources for many data mining and information retrieval tasks, but they remain far from complete. Knowledge base completion has attracted extensive research efforts from researchers and prac-titioners in diverse areas, which aims to infer missing facts from existing ones in a knowledge base. Quantities of knowledge base completion methods have been developed by regarding each relation as a translation from head entity to tail entity. However, existing methods merely concentrate on fact triples in the knowledge base or co-occurrence of words in the text, while supplementary semantic information expressed via related entities in the text has not been fully exploited. Meanwhile, the representation ability of current methods encounters bottlenecks due to the structure sparseness of knowledge base. In this paper, we propose a novel knowledge base representation learning method by taking advantage of the rich semantic information expressed via related entities in the textual corpus to expand the semantic structure of knowledge base. In this way, our model can break through the limitation of structure sparseness and promote the performance of knowledge base completion. Extensive experiments on two real-world datasets show that the proposed method successfully addresses the above issues and significantly outperforms the state-of-the-art methods on the benchmark task of link prediction.
知识库已成为许多数据挖掘和信息检索任务的必要资源,但知识库还远远不够完善。知识库补全吸引了各个领域的研究人员和实践者的广泛研究,其目的是从知识库中现有的事实中推断出缺失的事实。大量的知识库补全方法将每个关系视为从头部实体到尾部实体的转换。然而,现有的方法只关注知识库中的事实三元组或文本中词的共现,而没有充分利用文本中相关实体所表达的补充语义信息。同时,由于知识库的结构稀疏性,现有方法的表示能力遇到瓶颈。本文提出了一种新的知识库表示学习方法,利用文本语料库中相关实体表达的丰富语义信息来扩展知识库的语义结构。这样,我们的模型可以突破结构稀疏性的限制,提高知识库完成的性能。在两个真实数据集上的大量实验表明,该方法成功地解决了上述问题,并且在链路预测的基准任务上显著优于目前最先进的方法。
{"title":"Jointly Modeling Fact Triples and Text Information for Knowledge Base Completion","authors":"Xiuxing Li, Zhenyu Li, Zhichao Duan, Jiacheng Xu, Ning Liu, Jianyong Wang","doi":"10.1109/ICKG52313.2021.00037","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00037","url":null,"abstract":"Knowledge bases become essential resources for many data mining and information retrieval tasks, but they remain far from complete. Knowledge base completion has attracted extensive research efforts from researchers and prac-titioners in diverse areas, which aims to infer missing facts from existing ones in a knowledge base. Quantities of knowledge base completion methods have been developed by regarding each relation as a translation from head entity to tail entity. However, existing methods merely concentrate on fact triples in the knowledge base or co-occurrence of words in the text, while supplementary semantic information expressed via related entities in the text has not been fully exploited. Meanwhile, the representation ability of current methods encounters bottlenecks due to the structure sparseness of knowledge base. In this paper, we propose a novel knowledge base representation learning method by taking advantage of the rich semantic information expressed via related entities in the textual corpus to expand the semantic structure of knowledge base. In this way, our model can break through the limitation of structure sparseness and promote the performance of knowledge base completion. Extensive experiments on two real-world datasets show that the proposed method successfully addresses the above issues and significantly outperforms the state-of-the-art methods on the benchmark task of link prediction.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129932419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Robust Mathematical Model for Blood Supply Chain Network using Game Theory 基于博弈论的血液供应链网络鲁棒数学模型
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00066
Jaber Valizadeh, U. Aickelin, H. A. Khorshidi
No alternative to human blood has been found so far, and the only source is blood donation by donors. This study presents a blood supply chain optimization model focusing on the location and inventory management of different centers. The main purpose of this model is to reduce total costs, including hospital construction costs, patient allocation costs, patient service costs, expected time-out fines, non-absorbed blood fines, and outsourcing process costs. We then calculate the cost savings of collaborating in each hospital coalition to calculate the fair allocation of cost savings across hospitals. The proposed model is developed based on the data for the city of Tehran and previous studies in the field of the blood supply chain as well as using four Cooperative Game Theory (CGT) methods such as Shapley value, τ- Value, core-center and least core, to reduce the total cost and the fair profit sharing between hospitals have been evaluated.
到目前为止,还没有发现人类血液的替代品,唯一的来源是献血者的献血。本文提出了一个以不同中心的选址和库存管理为重点的血液供应链优化模型。该模型的主要目的是降低总成本,包括医院建设成本、患者分配成本、患者服务成本、预期超时罚款、未吸收血液罚款和外包流程成本。然后,我们计算每个医院联盟合作的成本节约,以计算成本节约在医院之间的公平分配。该模型基于德黑兰市的数据和先前在血液供应链领域的研究,并使用Shapley值、τ值、核心中心和最小核心等四种合作博弈论(CGT)方法来降低总成本,并评估了医院之间的公平利润分享。
{"title":"A Robust Mathematical Model for Blood Supply Chain Network using Game Theory","authors":"Jaber Valizadeh, U. Aickelin, H. A. Khorshidi","doi":"10.1109/ICKG52313.2021.00066","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00066","url":null,"abstract":"No alternative to human blood has been found so far, and the only source is blood donation by donors. This study presents a blood supply chain optimization model focusing on the location and inventory management of different centers. The main purpose of this model is to reduce total costs, including hospital construction costs, patient allocation costs, patient service costs, expected time-out fines, non-absorbed blood fines, and outsourcing process costs. We then calculate the cost savings of collaborating in each hospital coalition to calculate the fair allocation of cost savings across hospitals. The proposed model is developed based on the data for the city of Tehran and previous studies in the field of the blood supply chain as well as using four Cooperative Game Theory (CGT) methods such as Shapley value, τ- Value, core-center and least core, to reduce the total cost and the fair profit sharing between hospitals have been evaluated.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114601001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A divide-and-conquer method for computing preferred extensions of argumentation frameworks 一种计算论证框架优选扩展的分治法
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00039
Huan Zhang, Songmao Zhang
In this paper, we propose a divide-and-conquer method for solving the preferred extensions enumeration prob-lem, which is computationally intractable in argumentation frameworks. The rationale is to take advantage of the fact that for acyclic argumentation frameworks the computation becomes tractable with polynomial time. Concretely, we identify sufficient conditions for decomposing an argumentation framework into sub-frameworks based on certain cycles, where the soundness and completeness in computing preferred extensions are proved. Based on this conclusion, we devise the partitioning algorithm and carry out an evaluation on the International Competition on Computational Models of Argumentation (ICCMA) 2019 dataset. The results show that for the complex, time-consuming tasks our method could reduce running time when compared with the state-of-the-art solver in ICCMA. This is our first attempt in tackling the complex argumentative knowledge and many directions are yet to be explored, both theoretical and empirical.
本文提出了一种分而治之的方法来解决在论证框架中难以计算的首选扩展枚举问题。其基本原理是利用了这样一个事实,即对于非循环论证框架,计算在多项式时间内变得易于处理。具体地说,我们确定了基于一定循环将论证框架分解为子框架的充分条件,并证明了计算优选扩展的完备性和完全性。基于这一结论,我们设计了分区算法,并对2019年国际论证计算模型竞赛(ICCMA)数据集进行了评估。结果表明,对于复杂、耗时的任务,与ICCMA中最先进的求解器相比,我们的方法可以减少运行时间。这是我们在处理复杂的论证性知识方面的第一次尝试,许多方向还有待探索,无论是理论还是经验。
{"title":"A divide-and-conquer method for computing preferred extensions of argumentation frameworks","authors":"Huan Zhang, Songmao Zhang","doi":"10.1109/ICKG52313.2021.00039","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00039","url":null,"abstract":"In this paper, we propose a divide-and-conquer method for solving the preferred extensions enumeration prob-lem, which is computationally intractable in argumentation frameworks. The rationale is to take advantage of the fact that for acyclic argumentation frameworks the computation becomes tractable with polynomial time. Concretely, we identify sufficient conditions for decomposing an argumentation framework into sub-frameworks based on certain cycles, where the soundness and completeness in computing preferred extensions are proved. Based on this conclusion, we devise the partitioning algorithm and carry out an evaluation on the International Competition on Computational Models of Argumentation (ICCMA) 2019 dataset. The results show that for the complex, time-consuming tasks our method could reduce running time when compared with the state-of-the-art solver in ICCMA. This is our first attempt in tackling the complex argumentative knowledge and many directions are yet to be explored, both theoretical and empirical.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"2016 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114693912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Treatment Recommendation with Preference-based Reinforcement Learning 基于偏好的强化学习治疗推荐
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00025
Nan Xu, Nitin Kamra, Yan Liu
Treatment recommendation is a complex multi-faceted problem with many treatment goals considered by clini-cians and patients, e.g., optimizing the survival rate, mitigating negative impacts, reducing financial expenses, avoiding over-treatment, etc. Recently, deep reinforcement learning (RL) approaches have gained popularity for treatment recommendation. In this paper, we investigate preference-based reinforcement learning approaches for treatment recommendation, where the reward function is itself learned based on treatment goals, without requiring either expert demonstrations in advance or human involvement during policy learning. We first present an open sim-ulation platform11https://sites.google.com/view/tr-with-prl/ to model the evolution of two diseases, namely Cancer and Sepsis, and individuals' reactions to the received treatment. Secondly, we systematically examine preference-based RL for treatment recommendation via simulated experiments and observe high utility in the learned policy in terms of high survival rate and low side effects, with inferred rewards highly correlated to treatment goals. We further explore the transferability of inferred reward functions and guidelines for agent design to provide insights in achieving the right trade-off among various human objectives with preference-based RL approaches for treatment recommendation in the real world.
治疗推荐是一个复杂的多面性问题,临床医生和患者考虑的治疗目标很多,如提高生存率、减轻负面影响、减少经济支出、避免过度治疗等。最近,深度强化学习(RL)方法在治疗推荐中得到了广泛的应用。在本文中,我们研究了基于偏好的强化学习方法,用于治疗推荐,其中奖励函数本身是根据治疗目标学习的,不需要事先的专家演示或在政策学习过程中人类的参与。我们首先提出了一个开放的模拟平台11https://sites.google.com/view/tr-with-prl/来模拟两种疾病,即癌症和败血症的演变,以及个体对所接受治疗的反应。其次,我们通过模拟实验系统地检验了基于偏好的强化学习在治疗推荐中的应用,并观察到学习策略在高存活率和低副作用方面的高效用,推断的奖励与治疗目标高度相关。我们进一步探讨了推断奖励函数的可转移性和智能体设计指南,以提供在现实世界中使用基于偏好的强化学习方法在各种人类目标之间实现正确权衡的见解。
{"title":"Treatment Recommendation with Preference-based Reinforcement Learning","authors":"Nan Xu, Nitin Kamra, Yan Liu","doi":"10.1109/ICKG52313.2021.00025","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00025","url":null,"abstract":"Treatment recommendation is a complex multi-faceted problem with many treatment goals considered by clini-cians and patients, e.g., optimizing the survival rate, mitigating negative impacts, reducing financial expenses, avoiding over-treatment, etc. Recently, deep reinforcement learning (RL) approaches have gained popularity for treatment recommendation. In this paper, we investigate preference-based reinforcement learning approaches for treatment recommendation, where the reward function is itself learned based on treatment goals, without requiring either expert demonstrations in advance or human involvement during policy learning. We first present an open sim-ulation platform11https://sites.google.com/view/tr-with-prl/ to model the evolution of two diseases, namely Cancer and Sepsis, and individuals' reactions to the received treatment. Secondly, we systematically examine preference-based RL for treatment recommendation via simulated experiments and observe high utility in the learned policy in terms of high survival rate and low side effects, with inferred rewards highly correlated to treatment goals. We further explore the transferability of inferred reward functions and guidelines for agent design to provide insights in achieving the right trade-off among various human objectives with preference-based RL approaches for treatment recommendation in the real world.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128241281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Empirical Study of Deep Learning Frameworks for Melanoma Cancer Detection using Transfer Learning and Data Augmentation 基于迁移学习和数据增强的黑色素瘤癌症检测深度学习框架实证研究
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00015
Divya Gangwani, Qianxin Liang, Shuwen Wang, Xingquan Zhu
Melanoma is a type of skin cancer that usually develops rapidly and can spread to other parts of the body, causing death or complication of treatment to a large population. Early detection of Melanoma is the key to increase the chances of patients' survival. While accurate Melanoma diagnosis can be performed through clinical examination and dermatology tests, such procedures are usually time-consuming, costly, and severely delayed due to patients' emotional aspects or other obstacles. Recently, machine learning approaches, deep learning, in particular, have shown great potential in diagnosing Melanoma using images captured through the camera. Accurate detection of Melanoma using low-end images with machine learning delivers a solution for rapid screening of Melanoma without clinical visits or experts. Although many deep learning methods can be applied to Melanoma detection, there exists a large variance between their performance, given that their parameters, such as learning rates, optimizers, batch size, etc., always differ. In this paper, we carry out a systematic study to validate a more precise deep learning framework for detecting Melanoma and other types of skin lesions. A generic Convolutional Neural Network (CNN) is performed, and transfer learning using a pre-trained framework is proved to help improve the detection accuracy. In addition, data augmentation is applied, which improves the model performance. A series of parameters for the learning rates, batch size, and optimizers etc., are tested within the models. Our study shows tremendous improvement in Melanoma detection with higher accuracy, which can be very useful for medical experts to provide efficient Melanoma detection to patients.
黑色素瘤是一种皮肤癌,通常发展迅速,可以扩散到身体的其他部位,导致大量人群死亡或治疗并发症。早期发现黑色素瘤是增加患者生存机会的关键。虽然可以通过临床检查和皮肤科检查进行准确的黑色素瘤诊断,但这些程序通常耗时,昂贵,并且由于患者的情绪方面或其他障碍而严重延迟。最近,机器学习方法,特别是深度学习,在使用相机捕获的图像诊断黑色素瘤方面显示出巨大的潜力。使用低端图像和机器学习进行黑色素瘤的准确检测为快速筛查黑色素瘤提供了解决方案,无需临床访问或专家。尽管许多深度学习方法可以应用于黑色素瘤检测,但由于它们的参数(如学习率、优化器、批处理大小等)总是不同的,因此它们的性能之间存在很大差异。在本文中,我们进行了一项系统研究,以验证用于检测黑色素瘤和其他类型皮肤病变的更精确的深度学习框架。使用通用卷积神经网络(CNN),并证明使用预训练框架的迁移学习有助于提高检测精度。此外,还采用了数据增强技术,提高了模型的性能。在模型中测试了学习率、批处理大小和优化器等一系列参数。我们的研究显示了黑色素瘤检测的巨大进步和更高的准确性,这对于医学专家为患者提供有效的黑色素瘤检测非常有用。
{"title":"An Empirical Study of Deep Learning Frameworks for Melanoma Cancer Detection using Transfer Learning and Data Augmentation","authors":"Divya Gangwani, Qianxin Liang, Shuwen Wang, Xingquan Zhu","doi":"10.1109/ICKG52313.2021.00015","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00015","url":null,"abstract":"Melanoma is a type of skin cancer that usually develops rapidly and can spread to other parts of the body, causing death or complication of treatment to a large population. Early detection of Melanoma is the key to increase the chances of patients' survival. While accurate Melanoma diagnosis can be performed through clinical examination and dermatology tests, such procedures are usually time-consuming, costly, and severely delayed due to patients' emotional aspects or other obstacles. Recently, machine learning approaches, deep learning, in particular, have shown great potential in diagnosing Melanoma using images captured through the camera. Accurate detection of Melanoma using low-end images with machine learning delivers a solution for rapid screening of Melanoma without clinical visits or experts. Although many deep learning methods can be applied to Melanoma detection, there exists a large variance between their performance, given that their parameters, such as learning rates, optimizers, batch size, etc., always differ. In this paper, we carry out a systematic study to validate a more precise deep learning framework for detecting Melanoma and other types of skin lesions. A generic Convolutional Neural Network (CNN) is performed, and transfer learning using a pre-trained framework is proved to help improve the detection accuracy. In addition, data augmentation is applied, which improves the model performance. A series of parameters for the learning rates, batch size, and optimizers etc., are tested within the models. Our study shows tremendous improvement in Melanoma detection with higher accuracy, which can be very useful for medical experts to provide efficient Melanoma detection to patients.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130835934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
DiffXtract: Joint Discriminative Product Attribute-Value Extraction diffextraction:联合判别乘积属性值提取
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00044
Varun R. Embar, Andrey Kan, Bunyamin Sisman, C. Faloutsos, L. Getoor
Identifying discriminative attributes between prod-uct variations, e.g., the same wristwatch models but in different finishes, is crucial for improving e-commerce search engines and recommender systems. Despite the importance of these discrimi-native attributes, values for such attributes are often not available explicitly and instead are mentioned only in unstructured fields such as product title or product description. In this work, we introduce the novel task of discriminative attribute extraction which involves identifying the attributes that distinguish product variations, such as finish, and also, at the same time, extracting the values for these attributes from unstructured text. This task differs from the standard attribute value extraction task that has been well-studied in literature, as in our task we also need to identify the attribute, in addition to finding the value. We propose DiffXtract, a novel end-to-end, deep learning based approach that jointly identifies both the discriminative attribute and extracts its values from the product variations. The proposed approach is trained using a multitask objective and explicitly models the semantic representation of the discriminative attribute and uses it to extract the attribute values. We show that existing product attribute extraction approaches have several drawbacks, both theoretically and empirically. We also introduce a novel dataset based on a corpus of data previously crawled from a large number of e-commerce websites. In our empirical evaluation, we show that DiffXtract outperforms state-of-the-art deep learning-based and dictionary-based attribute extraction approaches by up to 8% F1 score when identifying attributes, and up to 10% F1 score when extracting attribute values.
识别不同产品之间的区别属性(例如,相同的手表型号但不同的饰面)对于改进电子商务搜索引擎和推荐系统至关重要。尽管这些区别性属性很重要,但这些属性的值通常不显式地可用,而是只在产品标题或产品描述等非结构化字段中提到。在这项工作中,我们引入了鉴别属性提取的新任务,该任务涉及识别区分产品变化的属性,如finish,同时从非结构化文本中提取这些属性的值。这个任务不同于文献中已经得到充分研究的标准属性值提取任务,因为在我们的任务中,除了找到值之外,我们还需要识别属性。我们提出了DiffXtract,这是一种新颖的端到端,基于深度学习的方法,可以联合识别判别属性并从产品变化中提取其值。该方法采用多任务目标进行训练,对判别属性的语义表示进行显式建模,并利用其提取属性值。我们从理论上和经验上证明了现有的产品属性提取方法存在一些缺陷。我们还介绍了一个基于先前从大量电子商务网站抓取的数据语料库的新数据集。在我们的经验评估中,我们表明DiffXtract在识别属性时比最先进的基于深度学习和基于字典的属性提取方法高出8%的F1分数,在提取属性值时高出10%的F1分数。
{"title":"DiffXtract: Joint Discriminative Product Attribute-Value Extraction","authors":"Varun R. Embar, Andrey Kan, Bunyamin Sisman, C. Faloutsos, L. Getoor","doi":"10.1109/ICKG52313.2021.00044","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00044","url":null,"abstract":"Identifying discriminative attributes between prod-uct variations, e.g., the same wristwatch models but in different finishes, is crucial for improving e-commerce search engines and recommender systems. Despite the importance of these discrimi-native attributes, values for such attributes are often not available explicitly and instead are mentioned only in unstructured fields such as product title or product description. In this work, we introduce the novel task of discriminative attribute extraction which involves identifying the attributes that distinguish product variations, such as finish, and also, at the same time, extracting the values for these attributes from unstructured text. This task differs from the standard attribute value extraction task that has been well-studied in literature, as in our task we also need to identify the attribute, in addition to finding the value. We propose DiffXtract, a novel end-to-end, deep learning based approach that jointly identifies both the discriminative attribute and extracts its values from the product variations. The proposed approach is trained using a multitask objective and explicitly models the semantic representation of the discriminative attribute and uses it to extract the attribute values. We show that existing product attribute extraction approaches have several drawbacks, both theoretically and empirically. We also introduce a novel dataset based on a corpus of data previously crawled from a large number of e-commerce websites. In our empirical evaluation, we show that DiffXtract outperforms state-of-the-art deep learning-based and dictionary-based attribute extraction approaches by up to 8% F1 score when identifying attributes, and up to 10% F1 score when extracting attribute values.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"3 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130043951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-level Spatio-temporal Matching Network for Multi-turn Response Selection in Retrieval-based Dialogue Systems 基于检索的对话系统多回合响应选择的多层次时空匹配网络
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00047
Mei Ma, Jianji Wang, Xuguang Lan, N. Zheng
The important task of multi-turn response selection in conversation systems must consider sufficient semantic information and spatio-temporal information when building retrieval-based chatbots. However, existing studies do not pay enough attention to both factors. In this study, a scheme of multi-turn response selection that combines a primary temporal matching module, an advanced temporal matching module, and a spatial matching module is proposed to extract matching information from context and response. The temporal matching modules progressively construct representations of the context and candidate responses at different granularities. Similarity matrices of the context and candidate responses are calculated and stacked using the spatial matching module. Convolutional neural network is then utilized to extract the spatial matching information. Finally, matching vectors of the three modules are fused to calculate the final matching score. Experimental results on two public datasets verify that our model can outperform state-of-the-art methods.
在构建基于检索的聊天机器人时,会话系统中多回合响应选择的重要任务必须考虑足够的语义信息和时空信息。然而,现有的研究对这两个因素的重视程度不够。本研究提出了一种结合初级时间匹配模块、高级时间匹配模块和空间匹配模块的多回合响应选择方案,从上下文和响应中提取匹配信息。时间匹配模块逐步构建不同粒度的上下文和候选响应的表示。使用空间匹配模块计算上下文和候选响应的相似矩阵并进行堆叠。然后利用卷积神经网络提取空间匹配信息。最后,对三个模块的匹配向量进行融合,计算出最终的匹配分数。在两个公共数据集上的实验结果验证了我们的模型可以优于最先进的方法。
{"title":"Multi-level Spatio-temporal Matching Network for Multi-turn Response Selection in Retrieval-based Dialogue Systems","authors":"Mei Ma, Jianji Wang, Xuguang Lan, N. Zheng","doi":"10.1109/ICKG52313.2021.00047","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00047","url":null,"abstract":"The important task of multi-turn response selection in conversation systems must consider sufficient semantic information and spatio-temporal information when building retrieval-based chatbots. However, existing studies do not pay enough attention to both factors. In this study, a scheme of multi-turn response selection that combines a primary temporal matching module, an advanced temporal matching module, and a spatial matching module is proposed to extract matching information from context and response. The temporal matching modules progressively construct representations of the context and candidate responses at different granularities. Similarity matrices of the context and candidate responses are calculated and stacked using the spatial matching module. Convolutional neural network is then utilized to extract the spatial matching information. Finally, matching vectors of the three modules are fused to calculate the final matching score. Experimental results on two public datasets verify that our model can outperform state-of-the-art methods.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125595472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Intuitionistic Fuzzy Requirements Aggregation for Graph Pattern Matching with Group Decision Makers 面向群体决策者的图模式匹配直觉模糊需求聚合
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00023
Haixia Zhao, Guliu Liu, Lei Li, Jiao Li
Graph Pattern Matching (GPM) plays an important role in the field of multi-attribute decision making. By designing a pattern graph involving multiple attribute constraints of the Decision Maker (DM), the sub graphs can be matched from the data graph. However, the existing work rarely considers the requirements from group DMs. In this case, the requirements on each attribute have multiple values from different DMs. How to aggregate these requirements and perform efficient sub graph matching is a challenging task. In this paper, first, a sub graph query problem that needs to consider the multiple requirements from group DMs is proposed. Then, to solve this problem, a Multi-Requirement-based Sub graph Query model (MR-SQ) is proposed, which is mainly composed of two stages: group require-ments aggregation and GPM. For the first stage, an Intuitionistic Fuzzy Requirements Aggregation (IFRA) method is proposed for requirements aggregation. Then, to solve the efficiency problem of large-scale GPM, a parallel strategy is designed for the GPM stage. Finally, the practicability and effectiveness of the proposed model have been verified through an illustrative example and time- performance comparison experiments.
图模式匹配(GPM)在多属性决策领域发挥着重要作用。通过设计包含多个属性约束的模式图,可以从数据图中匹配子图。然而,现有的工作很少考虑来自组dm的需求。在这种情况下,每个属性的需求具有来自不同dm的多个值。如何聚合这些需求并执行有效的子图匹配是一项具有挑战性的任务。本文首先提出了一种需要考虑群dm多个需求的子图查询问题。然后,针对这一问题,提出了一种基于多需求的子图查询模型(MR-SQ),该模型主要由组需求聚合和GPM两个阶段组成。第一阶段,提出了一种直观模糊需求聚合(IFRA)方法。然后,为解决大规模GPM的效率问题,设计了GPM阶段的并行策略。最后,通过一个算例和时间性能对比实验,验证了所提模型的实用性和有效性。
{"title":"Intuitionistic Fuzzy Requirements Aggregation for Graph Pattern Matching with Group Decision Makers","authors":"Haixia Zhao, Guliu Liu, Lei Li, Jiao Li","doi":"10.1109/ICKG52313.2021.00023","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00023","url":null,"abstract":"Graph Pattern Matching (GPM) plays an important role in the field of multi-attribute decision making. By designing a pattern graph involving multiple attribute constraints of the Decision Maker (DM), the sub graphs can be matched from the data graph. However, the existing work rarely considers the requirements from group DMs. In this case, the requirements on each attribute have multiple values from different DMs. How to aggregate these requirements and perform efficient sub graph matching is a challenging task. In this paper, first, a sub graph query problem that needs to consider the multiple requirements from group DMs is proposed. Then, to solve this problem, a Multi-Requirement-based Sub graph Query model (MR-SQ) is proposed, which is mainly composed of two stages: group require-ments aggregation and GPM. For the first stage, an Intuitionistic Fuzzy Requirements Aggregation (IFRA) method is proposed for requirements aggregation. Then, to solve the efficiency problem of large-scale GPM, a parallel strategy is designed for the GPM stage. Finally, the practicability and effectiveness of the proposed model have been verified through an illustrative example and time- performance comparison experiments.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122963979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 IEEE International Conference on Big Knowledge (ICBK)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1