{"title":"基于遥感图像的视觉问答","authors":"Sylvain Lobry, J. Murray, Diego Marcos, D. Tuia","doi":"10.1109/IGARSS.2019.8898891","DOIUrl":null,"url":null,"abstract":"Remote sensing images carry wide amounts of information beyond land cover or land use. Images contain visual and structural information that can be queried to obtain high level information about specific image content or relational dependencies between the objects sensed. This paper explores the possibility to use questions formulated in natural language as a generic and accessible way to extract this type of information from remote sensing images, i.e. visual question answering. We introduce an automatic way to create a dataset using OpenStreetMap1 data and present some preliminary results. Our proposed approach is based on deep learning, and is trained using our new dataset.","PeriodicalId":13262,"journal":{"name":"IGARSS 2019 - 2019 IEEE International Geoscience and Remote Sensing Symposium","volume":"17 1","pages":"4951-4954"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"13","resultStr":"{\"title\":\"Visual Question Answering From Remote Sensing Images\",\"authors\":\"Sylvain Lobry, J. Murray, Diego Marcos, D. Tuia\",\"doi\":\"10.1109/IGARSS.2019.8898891\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Remote sensing images carry wide amounts of information beyond land cover or land use. Images contain visual and structural information that can be queried to obtain high level information about specific image content or relational dependencies between the objects sensed. This paper explores the possibility to use questions formulated in natural language as a generic and accessible way to extract this type of information from remote sensing images, i.e. visual question answering. We introduce an automatic way to create a dataset using OpenStreetMap1 data and present some preliminary results. Our proposed approach is based on deep learning, and is trained using our new dataset.\",\"PeriodicalId\":13262,\"journal\":{\"name\":\"IGARSS 2019 - 2019 IEEE International Geoscience and Remote Sensing Symposium\",\"volume\":\"17 1\",\"pages\":\"4951-4954\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-11-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"13\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IGARSS 2019 - 2019 IEEE International Geoscience and Remote Sensing Symposium\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IGARSS.2019.8898891\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IGARSS 2019 - 2019 IEEE International Geoscience and Remote Sensing Symposium","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IGARSS.2019.8898891","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Visual Question Answering From Remote Sensing Images
Remote sensing images carry wide amounts of information beyond land cover or land use. Images contain visual and structural information that can be queried to obtain high level information about specific image content or relational dependencies between the objects sensed. This paper explores the possibility to use questions formulated in natural language as a generic and accessible way to extract this type of information from remote sensing images, i.e. visual question answering. We introduce an automatic way to create a dataset using OpenStreetMap1 data and present some preliminary results. Our proposed approach is based on deep learning, and is trained using our new dataset.