{"title":"面向阿拉伯语情感分析的预训练词嵌入比较研究","authors":"Mohamed Zouidine, Mohammed Khalil","doi":"10.1109/COMPSAC54236.2022.00196","DOIUrl":null,"url":null,"abstract":"In this paper, we conduct a series of experiments to systematically study both context-independent and context-dependent word embeddings for the purpose of Arabic sentiment analysis. We use pretrained word embeddings as fixed features extractors to provide input features for a CNN model. Experimental results with two different Arabic sentiment analysis datasets indicate that the pre-trained contextualized AraBERT model is the most suitable for such tasks. AraBERT reaches an accuracy score of 91.4% and 95.49% on the large Arabic book reviews dataset (LABR) and the hotel Arabic-reviews dataset (HARD), respectively.","PeriodicalId":330838,"journal":{"name":"2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Comparative Study of Pre-trained Word Embeddings for Arabic Sentiment Analysis\",\"authors\":\"Mohamed Zouidine, Mohammed Khalil\",\"doi\":\"10.1109/COMPSAC54236.2022.00196\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we conduct a series of experiments to systematically study both context-independent and context-dependent word embeddings for the purpose of Arabic sentiment analysis. We use pretrained word embeddings as fixed features extractors to provide input features for a CNN model. Experimental results with two different Arabic sentiment analysis datasets indicate that the pre-trained contextualized AraBERT model is the most suitable for such tasks. AraBERT reaches an accuracy score of 91.4% and 95.49% on the large Arabic book reviews dataset (LABR) and the hotel Arabic-reviews dataset (HARD), respectively.\",\"PeriodicalId\":330838,\"journal\":{\"name\":\"2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)\",\"volume\":\"61 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/COMPSAC54236.2022.00196\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/COMPSAC54236.2022.00196","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Comparative Study of Pre-trained Word Embeddings for Arabic Sentiment Analysis
In this paper, we conduct a series of experiments to systematically study both context-independent and context-dependent word embeddings for the purpose of Arabic sentiment analysis. We use pretrained word embeddings as fixed features extractors to provide input features for a CNN model. Experimental results with two different Arabic sentiment analysis datasets indicate that the pre-trained contextualized AraBERT model is the most suitable for such tasks. AraBERT reaches an accuracy score of 91.4% and 95.49% on the large Arabic book reviews dataset (LABR) and the hotel Arabic-reviews dataset (HARD), respectively.