{"title":"通过可解释的人工智能增强仇恨言论检测","authors":"D. Mittal, Harmeet Singh","doi":"10.1109/ICSMDI57622.2023.00028","DOIUrl":null,"url":null,"abstract":"The potential of XAI in detecting hate speech using deep learning models is versatile and multifaceted. To better understand the decision-making process of complex AI models, this study applied XAI to the dataset and investigated the interpretability and explanation of their decisions. The data was preprocessed by cleaning, tokenizing, lemmatizing, and removing inconsistencies in tweets. Simplification of categorical variables was also performed during training. Exploratory data analysis was conducted to identify patterns and insights in the dataset. The study used a set of existing models, including LIME, SHAP, XGBoost, and KTrain, to analyze the accuracy. The KTrain model achieved the highest accuracy and lowest loss among the variants developed to increase explainability.","PeriodicalId":373017,"journal":{"name":"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancing Hate Speech Detection through Explainable AI\",\"authors\":\"D. Mittal, Harmeet Singh\",\"doi\":\"10.1109/ICSMDI57622.2023.00028\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The potential of XAI in detecting hate speech using deep learning models is versatile and multifaceted. To better understand the decision-making process of complex AI models, this study applied XAI to the dataset and investigated the interpretability and explanation of their decisions. The data was preprocessed by cleaning, tokenizing, lemmatizing, and removing inconsistencies in tweets. Simplification of categorical variables was also performed during training. Exploratory data analysis was conducted to identify patterns and insights in the dataset. The study used a set of existing models, including LIME, SHAP, XGBoost, and KTrain, to analyze the accuracy. The KTrain model achieved the highest accuracy and lowest loss among the variants developed to increase explainability.\",\"PeriodicalId\":373017,\"journal\":{\"name\":\"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)\",\"volume\":\"39 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICSMDI57622.2023.00028\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSMDI57622.2023.00028","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Enhancing Hate Speech Detection through Explainable AI
The potential of XAI in detecting hate speech using deep learning models is versatile and multifaceted. To better understand the decision-making process of complex AI models, this study applied XAI to the dataset and investigated the interpretability and explanation of their decisions. The data was preprocessed by cleaning, tokenizing, lemmatizing, and removing inconsistencies in tweets. Simplification of categorical variables was also performed during training. Exploratory data analysis was conducted to identify patterns and insights in the dataset. The study used a set of existing models, including LIME, SHAP, XGBoost, and KTrain, to analyze the accuracy. The KTrain model achieved the highest accuracy and lowest loss among the variants developed to increase explainability.