Vijay Ravi, Jinhan Wang, Jonathan Flint, Abeer Alwan
The proposed method focuses on speaker disentanglement in the context of depression detection from speech signals. Previous approaches require patient/speaker labels, encounter instability due to loss maximization, and introduce unnecessary parameters for adversarial domain prediction. In contrast, the proposed unsupervised approach reduces cosine similarity between latent spaces of depression and pre-trained speaker classification models. This method outperforms baseline models, matches or exceeds adversarial methods in performance, and does so without relying on speaker labels or introducing additional model parameters, leading to a reduction in model complexity. The higher the speaker de-identification score (DeID), the better the depression detection system is in masking a patient's identity thereby enhancing the privacy attributes of depression detection systems. On the DAIC-WOZ dataset with ComparE16 features and an LSTM-only model, our method achieves an F1-Score of 0.776 and a DeID score of 92.87%, outperforming its adversarial counterpart which has an F1Score of 0.762 and 68.37% DeID, respectively. Furthermore, we demonstrate that speaker-disentanglement methods are complementary to text-based approaches, and a score-level fusion with a Word2vec-based depression detection model further enhances the overall performance to an F1-Score of 0.830.
所提出的方法侧重于在从语音信号中检测抑郁的背景下进行扬声器分离。以往的方法需要患者/说话人标签,会因损失最大化而导致不稳定性,并为对抗域预测引入不必要的参数。相比之下,所提出的无监督方法降低了抑郁潜空间与预训练说话人分类模型之间的余弦相似度。这种方法的性能优于基线模型,在性能上与对抗方法不相上下,甚至更胜一筹,而且无需依赖说话者标签或引入额外的模型参数,从而降低了模型的复杂性。说话者去身份化得分(DeID)越高,抑郁检测系统在掩盖患者身份方面的表现就越好,从而提高了抑郁检测系统的隐私属性。在使用 ComparE16 特征和纯 LSTM 模型的 DAIC-WOZ 数据集上,我们的方法获得了 0.776 的 F1 分数和 92.87% 的 DeID 分数,优于其 F1 分数为 0.762 和 DeID 分数为 68.37% 的对抗方法。此外,我们还证明了扬声器分离方法与基于文本的方法是互补的,而与基于 Word2vec 的抑郁检测模型进行分数级融合则进一步提高了整体性能,使 F1 分数达到 0.830。
{"title":"A Privacy-Preserving Unsupervised Speaker Disentanglement Method for Depression Detection from Speech.","authors":"Vijay Ravi, Jinhan Wang, Jonathan Flint, Abeer Alwan","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>The proposed method focuses on speaker disentanglement in the context of depression detection from speech signals. Previous approaches require patient/speaker labels, encounter instability due to loss maximization, and introduce unnecessary parameters for adversarial domain prediction. In contrast, the proposed unsupervised approach reduces cosine similarity between latent spaces of depression and pre-trained speaker classification models. This method outperforms baseline models, matches or exceeds adversarial methods in performance, and does so without relying on speaker labels or introducing additional model parameters, leading to a reduction in model complexity. The higher the speaker de-identification score (<i>DeID</i>), the better the depression detection system is in masking a patient's identity thereby enhancing the privacy attributes of depression detection systems. On the DAIC-WOZ dataset with ComparE16 features and an LSTM-only model, our method achieves an F1-Score of 0.776 and a <i>DeID</i> score of 92.87%, outperforming its adversarial counterpart which has an F1Score of 0.762 and 68.37% <i>DeID</i>, respectively. Furthermore, we demonstrate that speaker-disentanglement methods are complementary to text-based approaches, and a score-level fusion with a Word2vec-based depression detection model further enhances the overall performance to an F1-Score of 0.830.</p>","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3649 ","pages":"57-63"},"PeriodicalIF":0.0,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11034881/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140875057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Maneesh Bilalpur, Mert Inan, Dorsa Zeinali, Jeffrey F Cohn, Malihe Alikhani
Addressing the critical shortage of mental health resources for effective screening, diagnosis, and treatment remains a significant challenge. This scarcity underscores the need for innovative solutions, particularly in enhancing the accessibility and efficacy of therapeutic support. Embodied agents with advanced interactive capabilities emerge as a promising and cost-effective supplement to traditional caregiving methods. Crucial to these agents' effectiveness is their ability to simulate non-verbal behaviors, like backchannels, that are pivotal in establishing rapport and understanding in therapeutic contexts but remain under-explored. To improve the rapport-building capabilities of embodied agents we annotated backchannel smiles in videos of intimate face-to-face conversations over topics such as mental health, illness, and relationships. We hypothesized that both speaker and listener behaviors affect the duration and intensity of backchannel smiles. Using cues from speech prosody and language along with the demographics of the speaker and listener, we found them to contain significant predictors of the intensity of backchannel smiles. Based on our findings, we introduce backchannel smile production in embodied agents as a generation problem. Our attention-based generative model suggests that listener information offers performance improvements over the baseline speaker-centric generation approach. Conditioned generation using the significant predictors of smile intensity provides statistically significant improvements in empirical measures of generation quality. Our user study by transferring generated smiles to an embodied agent suggests that agent with backchannel smiles is perceived to be more human-like and is an attractive alternative for non-personal conversations over agent without backchannel smiles.
{"title":"Learning to Generate Context-Sensitive Backchannel Smiles for Embodied AI Agents with Applications in Mental Health Dialogues.","authors":"Maneesh Bilalpur, Mert Inan, Dorsa Zeinali, Jeffrey F Cohn, Malihe Alikhani","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Addressing the critical shortage of mental health resources for effective screening, diagnosis, and treatment remains a significant challenge. This scarcity underscores the need for innovative solutions, particularly in enhancing the accessibility and efficacy of therapeutic support. Embodied agents with advanced interactive capabilities emerge as a promising and cost-effective supplement to traditional caregiving methods. Crucial to these agents' effectiveness is their ability to simulate non-verbal behaviors, like backchannels, that are pivotal in establishing rapport and understanding in therapeutic contexts but remain under-explored. To improve the rapport-building capabilities of embodied agents we annotated backchannel smiles in videos of intimate face-to-face conversations over topics such as mental health, illness, and relationships. We hypothesized that both speaker and listener behaviors affect the duration and intensity of backchannel smiles. Using cues from speech prosody and language along with the demographics of the speaker and listener, we found them to contain significant predictors of the intensity of backchannel smiles. Based on our findings, we introduce backchannel smile production in embodied agents as a generation problem. Our attention-based generative model suggests that listener information offers performance improvements over the baseline speaker-centric generation approach. Conditioned generation using the significant predictors of smile intensity provides statistically significant improvements in empirical measures of generation quality. Our user study by transferring generated smiles to an embodied agent suggests that agent with backchannel smiles is perceived to be more human-like and is an attractive alternative for non-personal conversations over agent without backchannel smiles.</p>","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3649 ","pages":"12-22"},"PeriodicalIF":0.0,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11608428/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142775368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-09-12DOI: 10.55056/ceur-ws.org/vol-3482/paper041
Inna A. Kravtsova, Alina O. Mankuta, Vita A. Hamaniuk, Olga S. Bilozir, Andrei V. Voznyak
The paper explores the challenges and opportunities of developing professional competence of primary school teachers in teaching foreign languages according to the New Ukrainian School concept. The paper analyzes and describes various Internet resources that can facilitate and enhance foreign language learning outcomes in primary school. The paper argues that Internet resources can help modernize foreign language education in primary school and align it with the New Ukrainian School concept. The paper also discusses the importance of training primary school teachers in the methods of organizing distance learning, which is a priority for higher education institutions in the context of continuous education.
{"title":"Internet resources for foreign language education in primary school: challenges and opportunities","authors":"Inna A. Kravtsova, Alina O. Mankuta, Vita A. Hamaniuk, Olga S. Bilozir, Andrei V. Voznyak","doi":"10.55056/ceur-ws.org/vol-3482/paper041","DOIUrl":"https://doi.org/10.55056/ceur-ws.org/vol-3482/paper041","url":null,"abstract":"The paper explores the challenges and opportunities of developing professional competence of primary school teachers in teaching foreign languages according to the New Ukrainian School concept. The paper analyzes and describes various Internet resources that can facilitate and enhance foreign language learning outcomes in primary school. The paper argues that Internet resources can help modernize foreign language education in primary school and align it with the New Ukrainian School concept. The paper also discusses the importance of training primary school teachers in the methods of organizing distance learning, which is a priority for higher education institutions in the context of continuous education.","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135935615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-09-12DOI: 10.55056/ceur-ws.org/vol-3482/paper116
Olha V. Chorna, Vita A. Hamaniuk, Oksana Y. Markheva, Andrei V. Voznyak
The integration of information and communication technologies (ICT) in education has increased the possibilities and expanded the boundaries of the learning process. It is also a prerequisite for implementing distance learning. Various online resources, such as e-mail, blogs, forums, online applications, and video hosting sites, can be used to create open learning and education environments. This study focuses on the use of informational educational technologies for learning foreign languages, especially German. The article presents the results of a theoretical analysis of the content of YouTube video materials in terms of their personal and didactic relevance for teaching German as a first or second foreign language in higher education, specifically at a pedagogical university. Based on the practical experience of using several popular thematic YouTube channels with a large and stable audience, a brief didactic analysis of their products is provided and suggestions are made on how to transform video content into methodological material for the practical course of German language for future teachers. The article also explores the potential of using alternative YouTube resources for distance learning with regard to the development of mediation skills as defined by the authors of the CEFR Companion Volume with New Descriptors. Four types of resources that can serve as teaching materials are identified and analyzed; some examples of their preparation and use for the training of future foreign language teachers are given. The article also discusses the open resources ONCOO and TWINE, which can be used to foster the autonomy of future foreign language teachers, and describes their features. The proposed recommendations can help to achieve the following objectives: enriching vocabulary; semanticizing phraseological units, fixed expressions, clichés; developing pronunciation skills; enhancing linguistic and ICT competencies; improving listening and speaking skills; increasing motivation to learn, etc.
{"title":"YouTube as an open resource for foreign language learning: a case study of German","authors":"Olha V. Chorna, Vita A. Hamaniuk, Oksana Y. Markheva, Andrei V. Voznyak","doi":"10.55056/ceur-ws.org/vol-3482/paper116","DOIUrl":"https://doi.org/10.55056/ceur-ws.org/vol-3482/paper116","url":null,"abstract":"The integration of information and communication technologies (ICT) in education has increased the possibilities and expanded the boundaries of the learning process. It is also a prerequisite for implementing distance learning. Various online resources, such as e-mail, blogs, forums, online applications, and video hosting sites, can be used to create open learning and education environments. This study focuses on the use of informational educational technologies for learning foreign languages, especially German. The article presents the results of a theoretical analysis of the content of YouTube video materials in terms of their personal and didactic relevance for teaching German as a first or second foreign language in higher education, specifically at a pedagogical university. Based on the practical experience of using several popular thematic YouTube channels with a large and stable audience, a brief didactic analysis of their products is provided and suggestions are made on how to transform video content into methodological material for the practical course of German language for future teachers. The article also explores the potential of using alternative YouTube resources for distance learning with regard to the development of mediation skills as defined by the authors of the CEFR Companion Volume with New Descriptors. Four types of resources that can serve as teaching materials are identified and analyzed; some examples of their preparation and use for the training of future foreign language teachers are given. The article also discusses the open resources ONCOO and TWINE, which can be used to foster the autonomy of future foreign language teachers, and describes their features. The proposed recommendations can help to achieve the following objectives: enriching vocabulary; semanticizing phraseological units, fixed expressions, clichés; developing pronunciation skills; enhancing linguistic and ICT competencies; improving listening and speaking skills; increasing motivation to learn, etc.","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135935773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Anthony Huffman, Edison Ong, Tim Brunson, Nasim Sanati, Jie Zheng, Anna Maria Masci, Guanming Wu, Yongqun He
ImmPort, the world's largest repository of immunology data, includes many vaccine immune response datasets. ImmPort maps the metadata of these studies to ontology and database schema. As of February 28, 2023, our ImmPort data analysis identified 6.258 immune exposures using 47 vaccines in 4,607 human subjects, and 324 cohort studies from the ImmPort. We hypothesized that an integrative ontological representation of the data from these studies would enhance our understanding and analysis of these ImmPort vaccine studies, and with ontological classification and tools such as VIGET, we could further study the effects of different conditions such as vaccine types and host biological sex on the vaccine response gene expression profiles. Our Vaccine Ontology (VO) analysis classified these 37 vaccines into bacterial, viral, and protozoan vaccine types with different vaccine properties. The ImmPort metadata types were modeled with the Vaccine Investigation Ontology (VIO). Our new ontology-based pipeline extracted vaccine response data from the ImmPort database, annotated them based on ontology, obtained corresponding gene expression data from the GEO, and performed consistent omics data analysis. Our use case found gene profiles shared and differed from live and killed inactivated influenza vaccines. Furthermore, our Omics data analysis using the VIGET tool found that female and male human subjects have differential host responses for influenza vaccines. For example, our study showed much stronger early female responses to influenza vaccination than males, and males was able to show active immune responses at a later stage. Interestingly, the female (but not male) human subject group also showed significantly enriched neutrophil degranulation at Day 3 after influenza vaccination; however, males (but not females) displayed significantly enriched neutrophil degranulation at Day 14 after influenza vaccination. These mechanisms have been used to find differences between the gene lists and pathways of host responses to different vaccines conditional to different factors including vaccine types and host biological sex. Moreover, this framework can be expanded to other vaccines and vaccine categories easily.
{"title":"Ontology-based representation and analysis of conditional vaccine immune responses using Omics data.","authors":"Anthony Huffman, Edison Ong, Tim Brunson, Nasim Sanati, Jie Zheng, Anna Maria Masci, Guanming Wu, Yongqun He","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>ImmPort, the world's largest repository of immunology data, includes many vaccine immune response datasets. ImmPort maps the metadata of these studies to ontology and database schema. As of February 28, 2023, our ImmPort data analysis identified 6.258 immune exposures using 47 vaccines in 4,607 human subjects, and 324 cohort studies from the ImmPort. We hypothesized that an integrative ontological representation of the data from these studies would enhance our understanding and analysis of these ImmPort vaccine studies, and with ontological classification and tools such as VIGET, we could further study the effects of different conditions such as vaccine types and host biological sex on the vaccine response gene expression profiles. Our Vaccine Ontology (VO) analysis classified these 37 vaccines into bacterial, viral, and protozoan vaccine types with different vaccine properties. The ImmPort metadata types were modeled with the Vaccine Investigation Ontology (VIO). Our new ontology-based pipeline extracted vaccine response data from the ImmPort database, annotated them based on ontology, obtained corresponding gene expression data from the GEO, and performed consistent omics data analysis. Our use case found gene profiles shared and differed from live and killed inactivated influenza vaccines. Furthermore, our Omics data analysis using the VIGET tool found that female and male human subjects have differential host responses for influenza vaccines. For example, our study showed much stronger early female responses to influenza vaccination than males, and males was able to show active immune responses at a later stage. Interestingly, the female (but not male) human subject group also showed significantly enriched neutrophil degranulation at Day 3 after influenza vaccination; however, males (but not females) displayed significantly enriched neutrophil degranulation at Day 14 after influenza vaccination. These mechanisms have been used to find differences between the gene lists and pathways of host responses to different vaccines conditional to different factors including vaccine types and host biological sex. Moreover, this framework can be expanded to other vaccines and vaccine categories easily.</p>","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3603 ","pages":"1-12"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11584146/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142711244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Evangelos Sariyanidi, Casey J Zampella, Ellis DeJardin, John D Herrington, Robert T Schultz, Birkan Tunc
Advances in computational behavior analysis via artificial intelligence (AI) promise to improve mental healthcare services by providing clinicians with tools to assist diagnosis or measurement of treatment outcomes. This potential has spurred an increasing number of studies in which automated pipelines predict diagnoses of mental health conditions. However, a fundamental question remains unanswered: How do the predictions of the AI algorithms correspond and compare with the predictions of humans? This is a critical question if AI technology is to be used as an assistive tool, because the utility of an AI algorithm would be negligible if it provides little information beyond what clinicians can readily infer. In this paper, we compare the performance of 19 human raters (8 autism experts and 11 non-experts) and that of an AI algorithm in terms of predicting autism diagnosis from short (3-minute) videos of N = 42 participants in a naturalistic conversation. Results show that the AI algorithm achieves an average accuracy of 80.5%, which is comparable to that of clinicians with expertise in autism (83.1%) and clinical research staff without specialized expertise (78.3%). Critically, diagnoses that were inaccurately predicted by most humans (experts and non-experts, alike) were typically correctly predicted by AI. Our results highlight the potential of AI as an assistive tool that can augment clinician diagnostic decision-making.
{"title":"Comparison of Human Experts and AI in Predicting Autism from Facial Behavior.","authors":"Evangelos Sariyanidi, Casey J Zampella, Ellis DeJardin, John D Herrington, Robert T Schultz, Birkan Tunc","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Advances in computational behavior analysis via artificial intelligence (AI) promise to improve mental healthcare services by providing clinicians with tools to assist diagnosis or measurement of treatment outcomes. This potential has spurred an increasing number of studies in which automated pipelines predict diagnoses of mental health conditions. However, a fundamental question remains unanswered: How do the predictions of the AI algorithms correspond and compare with the predictions of humans? This is a critical question if AI technology is to be used as an assistive tool, because the utility of an AI algorithm would be negligible if it provides little information beyond what clinicians can readily infer. In this paper, we compare the performance of 19 human raters (8 autism experts and 11 non-experts) and that of an AI algorithm in terms of predicting autism diagnosis from short (3-minute) videos of <i>N</i> = 42 participants in a naturalistic conversation. Results show that the AI algorithm achieves an average accuracy of 80.5%, which is comparable to that of clinicians with expertise in autism (83.1%) and clinical research staff without specialized expertise (78.3%). Critically, diagnoses that were inaccurately predicted by most humans (experts and non-experts, alike) were typically correctly predicted by AI. Our results highlight the potential of AI as an assistive tool that can augment clinician diagnostic decision-making.</p>","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3359 ITAH","pages":"48-57"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10687770/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138464759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Michael Rabenberg, Anuwat Pengput, Werner Ceusters
Adequately representing kinship relations is crucial for a variety of medical and biomedical applications. Several kinship ontologies have been proposed but none of them have been designed thus far in line with the Basic Formal Ontology. In this paper, we propose a novel kinship ontology that exhibits the following characteristics: (1) it is fully axiomatized in First Order Logic following the rules governing predicate formation as proposed in BFO2020-FOL, (2) it is modularized in 6 separate files written in the Common Logic Interface Format (CLIF) each one of which can be imported based on specific needs, (3) it provides bridging axioms to and from SNOMED CT, and (4) it contains an extra module with axioms which would not be literally true when phrased naively but are crafted in such a way that they highlight the unusual kinship relations they represent and can be used to generate alerts on possible data entry mistakes. We describe design considerations and challenges encountered.
{"title":"An Extendible Realism-Based Ontology for Kinship.","authors":"Michael Rabenberg, Anuwat Pengput, Werner Ceusters","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Adequately representing kinship relations is crucial for a variety of medical and biomedical applications. Several kinship ontologies have been proposed but none of them have been designed thus far in line with the Basic Formal Ontology. In this paper, we propose a novel kinship ontology that exhibits the following characteristics: (1) it is fully axiomatized in First Order Logic following the rules governing predicate formation as proposed in BFO2020-FOL, (2) it is modularized in 6 separate files written in the Common Logic Interface Format (CLIF) each one of which can be imported based on specific needs, (3) it provides bridging axioms to and from SNOMED CT, and (4) it contains an extra module with axioms which would not be literally true when phrased naively but are crafted in such a way that they highlight the unusual kinship relations they represent and can be used to generate alerts on possible data entry mistakes. We describe design considerations and challenges encountered.</p>","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3603 ","pages":"25-35"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11131162/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141163042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ali Behrouz, Mathias Lécuyer, Cynthia Rudin, Margo Seltzer
Sparse decision trees are one of the most common forms of interpretable models. While recent advances have produced algorithms that fully optimize sparse decision trees for prediction, that work does not address policy design, because the algorithms cannot handle weighted data samples. Specifically, they rely on the discreteness of the loss function, which means that real-valued weights cannot be directly used. For example, none of the existing techniques produce policies that incorporate inverse propensity weighting on individual data points. We present three algorithms for efficient sparse weighted decision tree optimization. The first approach directly optimizes the weighted loss function; however, it tends to be computationally inefficient for large datasets. Our second approach, which scales more efficiently, transforms weights to integer values and uses data duplication to transform the weighted decision tree optimization problem into an unweighted (but larger) counterpart. Our third algorithm, which scales to much larger datasets, uses a randomized procedure that samples each data point with a probability proportional to its weight. We present theoretical bounds on the error of the two fast methods and show experimentally that these methods can be two orders of magnitude faster than the direct optimization of the weighted loss, without losing significant accuracy.
{"title":"Fast Optimization of Weighted Sparse Decision Trees for use in Optimal Treatment Regimes and Optimal Policy Design.","authors":"Ali Behrouz, Mathias Lécuyer, Cynthia Rudin, Margo Seltzer","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Sparse decision trees are one of the most common forms of interpretable models. While recent advances have produced algorithms that fully optimize sparse decision trees for <i>prediction</i>, that work does not address <i>policy design</i>, because the algorithms cannot handle weighted data samples. Specifically, they rely on the discreteness of the loss function, which means that real-valued weights cannot be directly used. For example, none of the existing techniques produce policies that incorporate inverse propensity weighting on individual data points. We present three algorithms for efficient sparse weighted decision tree optimization. The first approach directly optimizes the weighted loss function; however, it tends to be computationally inefficient for large datasets. Our second approach, which scales more efficiently, transforms weights to integer values and uses data duplication to transform the weighted decision tree optimization problem into an unweighted (but larger) counterpart. Our third algorithm, which scales to much larger datasets, uses a randomized procedure that samples each data point with a probability proportional to its weight. We present theoretical bounds on the error of the two fast methods and show experimentally that these methods can be two orders of magnitude faster than the direct optimization of the weighted loss, without losing significant accuracy.</p>","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3318 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10039433/pdf/nihms-1883491.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9197753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-01DOI: 10.48550/arXiv.2210.06825
Ali Behrouz, Mathias Lécuyer, C. Rudin, M. Seltzer
Sparse decision trees are one of the most common forms of interpretable models. While recent advances have produced algorithms that fully optimize sparse decision trees for prediction, that work does not address policy design, because the algorithms cannot handle weighted data samples. Specifically, they rely on the discreteness of the loss function, which means that real-valued weights cannot be directly used. For example, none of the existing techniques produce policies that incorporate inverse propensity weighting on individual data points. We present three algorithms for efficient sparse weighted decision tree optimization. The first approach directly optimizes the weighted loss function; however, it tends to be computationally inefficient for large datasets. Our second approach, which scales more efficiently, transforms weights to integer values and uses data duplication to transform the weighted decision tree optimization problem into an unweighted (but larger) counterpart. Our third algorithm, which scales to much larger datasets, uses a randomized procedure that samples each data point with a probability proportional to its weight. We present theoretical bounds on the error of the two fast methods and show experimentally that these methods can be two orders of magnitude faster than the direct optimization of the weighted loss, without losing significant accuracy.
{"title":"Fast optimization of weighted sparse decision trees for use in optimal treatment regimes and optimal policy design","authors":"Ali Behrouz, Mathias Lécuyer, C. Rudin, M. Seltzer","doi":"10.48550/arXiv.2210.06825","DOIUrl":"https://doi.org/10.48550/arXiv.2210.06825","url":null,"abstract":"Sparse decision trees are one of the most common forms of interpretable models. While recent advances have produced algorithms that fully optimize sparse decision trees for prediction, that work does not address policy design, because the algorithms cannot handle weighted data samples. Specifically, they rely on the discreteness of the loss function, which means that real-valued weights cannot be directly used. For example, none of the existing techniques produce policies that incorporate inverse propensity weighting on individual data points. We present three algorithms for efficient sparse weighted decision tree optimization. The first approach directly optimizes the weighted loss function; however, it tends to be computationally inefficient for large datasets. Our second approach, which scales more efficiently, transforms weights to integer values and uses data duplication to transform the weighted decision tree optimization problem into an unweighted (but larger) counterpart. Our third algorithm, which scales to much larger datasets, uses a randomized procedure that samples each data point with a probability proportional to its weight. We present theoretical bounds on the error of the two fast methods and show experimentally that these methods can be two orders of magnitude faster than the direct optimization of the weighted loss, without losing significant accuracy.","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3318 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45358116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Asiyah Yu Lin, Yuki Yamagata, William D Duncan, Leigh C Carmody, Tatsuya Kushida, Hiroshi Masuya, John Beverley, Biswanath Dutta, Michael DeBellis, Zoë May Pendlington, Paola Roncaglia, Yongqun He
Ontologies have emerged to become critical to support data and knowledge representation, standardization, integration, and analysis. The SARS-CoV-2 pandemic led to the rapid proliferation of COVID-19 data, as well as the development of many COVID-19 ontologies. In the interest of supporting data interoperability, we initiated a community-based effort to harmonize COVID-19 ontologies. Our effort involves the collaborative discussion among developers of seven COVID-19 related ontologies, and the merging of four ontologies. This effort demonstrates the feasibility of harmonizing these ontologies in an interoperable framework to support integrative representation and analysis of COVID-19 related data and knowledge.
{"title":"A community effort for COVID-19 Ontology Harmonization.","authors":"Asiyah Yu Lin, Yuki Yamagata, William D Duncan, Leigh C Carmody, Tatsuya Kushida, Hiroshi Masuya, John Beverley, Biswanath Dutta, Michael DeBellis, Zoë May Pendlington, Paola Roncaglia, Yongqun He","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Ontologies have emerged to become critical to support data and knowledge representation, standardization, integration, and analysis. The SARS-CoV-2 pandemic led to the rapid proliferation of COVID-19 data, as well as the development of many COVID-19 ontologies. In the interest of supporting data interoperability, we initiated a community-based effort to harmonize COVID-19 ontologies. Our effort involves the collaborative discussion among developers of seven COVID-19 related ontologies, and the merging of four ontologies. This effort demonstrates the feasibility of harmonizing these ontologies in an interoperable framework to support integrative representation and analysis of COVID-19 related data and knowledge.</p>","PeriodicalId":72554,"journal":{"name":"CEUR workshop proceedings","volume":"3073 ","pages":"122-127"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10262777/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10024339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}