{"title":"Multi-label Prediction for Visual Sentiment Analysis using Eight Different Emotions based on Psychology","authors":"Tetsuya Asakawa, Masaki Aono","doi":"10.1145/3484274.3484296","DOIUrl":null,"url":null,"abstract":"In visual sentiment analysis, sentiment estimation from images is a challenging research problem. Previous studies focused on a few specific sentiments and their intensities and have not captured abundant psychological human feelings. In addition, multi-label sentiment estimation from images has not been sufficiently investigated. The purpose of this research is to build a visual sentiment dataset, accurately estimate the sentiments as a multi-label multi-class problem from images that simultaneously evoke multiple emotions. We built a visual sentiment dataset based on Plutchik's wheel of emotions. We describe this ‘Senti8PW’ dataset, then perform multi-label sentiment analysis using the dataset, where we propose a combined deep neural network model that enables inputs from both hand-crafted features and CNN features. We also introduce a threshold-based multi-label prediction algorithm, in which we assume that each emotion has a probability distribution. In other words, after training our deep neural network, we predict evoked emotions for an image if the intensity of the emotion is larger than the threshold of the corresponding emotion. Extensive experiments were conducted on our dataset. Our model achieves superior results compared to the state-of-the-art algorithms in terms of subsets.","PeriodicalId":143540,"journal":{"name":"Proceedings of the 4th International Conference on Control and Computer Vision","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 4th International Conference on Control and Computer Vision","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3484274.3484296","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In visual sentiment analysis, sentiment estimation from images is a challenging research problem. Previous studies focused on a few specific sentiments and their intensities and have not captured abundant psychological human feelings. In addition, multi-label sentiment estimation from images has not been sufficiently investigated. The purpose of this research is to build a visual sentiment dataset, accurately estimate the sentiments as a multi-label multi-class problem from images that simultaneously evoke multiple emotions. We built a visual sentiment dataset based on Plutchik's wheel of emotions. We describe this ‘Senti8PW’ dataset, then perform multi-label sentiment analysis using the dataset, where we propose a combined deep neural network model that enables inputs from both hand-crafted features and CNN features. We also introduce a threshold-based multi-label prediction algorithm, in which we assume that each emotion has a probability distribution. In other words, after training our deep neural network, we predict evoked emotions for an image if the intensity of the emotion is larger than the threshold of the corresponding emotion. Extensive experiments were conducted on our dataset. Our model achieves superior results compared to the state-of-the-art algorithms in terms of subsets.