{"title":"基于文本内容的多级讲座视频分类","authors":"Veysel Sercan Ağzıyağlı, H. Oğul","doi":"10.1109/AICT50176.2020.9368692","DOIUrl":null,"url":null,"abstract":"Recent interest in e-learning and distance education services has significantly increased the amount of lecture video data in public and institutional repositories. In their current forms, users can browse in these collections using meta-data-based search queries such as course name, description, instructor and syllabus. However, lecture video entries have rich contents, including image, text and speech, which can not be easily represented by meta-data annotations. Therefore, there is an emerging need to develop tools that will automatically annotate lecture videos to facilitate more targeted search. A simple way to realize this is to classify lectures into known categories. With this objective, this paper presents a method for classifying videos based on extracted text content in several semantic levels. The method is based on Bidirectional Long-Short Term Memory (Bi-LSTM) applied on word embedding vectors of text content extracted by Optical Character Recognition (OCR). This approach can outperform conventional machine learning models and provide a useful solution for automatic lecture video annotation to support online education.","PeriodicalId":136491,"journal":{"name":"2020 IEEE 14th International Conference on Application of Information and Communication Technologies (AICT)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-level lecture video classification using text content\",\"authors\":\"Veysel Sercan Ağzıyağlı, H. Oğul\",\"doi\":\"10.1109/AICT50176.2020.9368692\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent interest in e-learning and distance education services has significantly increased the amount of lecture video data in public and institutional repositories. In their current forms, users can browse in these collections using meta-data-based search queries such as course name, description, instructor and syllabus. However, lecture video entries have rich contents, including image, text and speech, which can not be easily represented by meta-data annotations. Therefore, there is an emerging need to develop tools that will automatically annotate lecture videos to facilitate more targeted search. A simple way to realize this is to classify lectures into known categories. With this objective, this paper presents a method for classifying videos based on extracted text content in several semantic levels. The method is based on Bidirectional Long-Short Term Memory (Bi-LSTM) applied on word embedding vectors of text content extracted by Optical Character Recognition (OCR). This approach can outperform conventional machine learning models and provide a useful solution for automatic lecture video annotation to support online education.\",\"PeriodicalId\":136491,\"journal\":{\"name\":\"2020 IEEE 14th International Conference on Application of Information and Communication Technologies (AICT)\",\"volume\":\"9 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-10-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE 14th International Conference on Application of Information and Communication Technologies (AICT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/AICT50176.2020.9368692\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE 14th International Conference on Application of Information and Communication Technologies (AICT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AICT50176.2020.9368692","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-level lecture video classification using text content
Recent interest in e-learning and distance education services has significantly increased the amount of lecture video data in public and institutional repositories. In their current forms, users can browse in these collections using meta-data-based search queries such as course name, description, instructor and syllabus. However, lecture video entries have rich contents, including image, text and speech, which can not be easily represented by meta-data annotations. Therefore, there is an emerging need to develop tools that will automatically annotate lecture videos to facilitate more targeted search. A simple way to realize this is to classify lectures into known categories. With this objective, this paper presents a method for classifying videos based on extracted text content in several semantic levels. The method is based on Bidirectional Long-Short Term Memory (Bi-LSTM) applied on word embedding vectors of text content extracted by Optical Character Recognition (OCR). This approach can outperform conventional machine learning models and provide a useful solution for automatic lecture video annotation to support online education.