首页 > 最新文献

Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery最新文献

英文 中文
Cross-Modal Learning of Housing Quality in Amsterdam 阿姆斯特丹住房质量的跨模式学习
A. Levering, Diego Marcos, Ilan Havinga, D. Tuia
In our research we test data and models for the recognition of housing quality in the city of Amsterdam from ground-level and aerial imagery. For ground-level images we compare Google StreetView (GSV) to Flickr images. Our results show that GSV predicts the most accurate building quality scores, approximately 30% better than using only aerial images. However, we find that through careful filtering and by using the right pre-trained model, Flickr image features combined with aerial image features are able to halve the performance gap to GSV features from 30% to 15%. Our results indicate that there are viable alternatives to GSV for liveability factor prediction, which is encouraging as GSV images are more difficult to acquire and not always available.
在我们的研究中,我们从地面和航空图像中测试了识别阿姆斯特丹市住房质量的数据和模型。对于地面图像,我们将Google StreetView (GSV)与Flickr图像进行比较。我们的研究结果表明,GSV预测最准确的建筑质量分数,大约比仅使用航空图像好30%。然而,我们发现,通过仔细过滤和使用正确的预训练模型,Flickr图像特征结合航空图像特征能够将与GSV特征的性能差距从30%减少到15%。我们的研究结果表明,有可行的替代GSV来预测宜居系数,这是令人鼓舞的,因为GSV图像更难获取,而且并不总是可用的。
{"title":"Cross-Modal Learning of Housing Quality in Amsterdam","authors":"A. Levering, Diego Marcos, Ilan Havinga, D. Tuia","doi":"10.1145/3486635.3491067","DOIUrl":"https://doi.org/10.1145/3486635.3491067","url":null,"abstract":"In our research we test data and models for the recognition of housing quality in the city of Amsterdam from ground-level and aerial imagery. For ground-level images we compare Google StreetView (GSV) to Flickr images. Our results show that GSV predicts the most accurate building quality scores, approximately 30% better than using only aerial images. However, we find that through careful filtering and by using the right pre-trained model, Flickr image features combined with aerial image features are able to halve the performance gap to GSV features from 30% to 15%. Our results indicate that there are viable alternatives to GSV for liveability factor prediction, which is encouraging as GSV images are more difficult to acquire and not always available.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128519591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Synthetic Map Generation to Provide Unlimited Training Data for Historical Map Text Detection 合成地图生成为历史地图文本检测提供无限的训练数据
Zekun Li, Runyu Guan, Qianmu Yu, Yao-Yi Chiang, Craig A. Knoblock
Many historical map sheets are publicly available for studies that require long-term historical geographic data. The cartographic design of these maps includes a combination of map symbols and text labels. Automatically reading text labels from map images could greatly speed up the map interpretation and helps generate rich metadata describing the map content. Many text detection algorithms have been proposed to locate text regions in map images automatically, but most of the algorithms are trained on out-of-domain datasets (e.g., scenic images). Training data determines the quality of machine learning models, and manually annotating text regions in map images is labor-extensive and time-consuming. On the other hand, existing geographic data sources, such as Open-StreetMap (OSM), contain machine-readable map layers, which allow us to separate out the text layer and obtain text label annotations easily. However, the cartographic styles between OSM map tiles and historical maps are significantly different. This paper proposes a method to automatically generate an unlimited amount of annotated historical map images for training text detection models. We use a style transfer model to convert contemporary map images into historical style and place text labels upon them. We show that the state-of-the-art text detection models (e.g., PSENet) can benefit from the synthetic historical maps and achieve significant improvement for historical map text detection.
对于需要长期历史地理数据的研究,许多历史地图都是公开的。这些地图的制图设计包括地图符号和文本标签的组合。从地图图像中自动读取文本标签可以大大加快地图解释的速度,并有助于生成描述地图内容的丰富元数据。许多文本检测算法已经被提出来自动定位地图图像中的文本区域,但大多数算法都是在域外数据集(如风景图像)上训练的。训练数据决定了机器学习模型的质量,手动标注地图图像中的文本区域是费时费力的。另一方面,现有的地理数据源,如Open-StreetMap (OSM),包含机器可读的地图层,这使得我们可以很容易地分离出文本层并获得文本标签注释。然而,OSM地图瓷砖和历史地图之间的制图风格有很大不同。本文提出了一种自动生成无限量带注释的历史地图图像用于训练文本检测模型的方法。我们使用风格转换模型将当代地图图像转换为历史风格,并在其上放置文本标签。我们证明了最先进的文本检测模型(例如PSENet)可以从合成历史地图中受益,并在历史地图文本检测方面取得了显着改进。
{"title":"Synthetic Map Generation to Provide Unlimited Training Data for Historical Map Text Detection","authors":"Zekun Li, Runyu Guan, Qianmu Yu, Yao-Yi Chiang, Craig A. Knoblock","doi":"10.1145/3486635.3491070","DOIUrl":"https://doi.org/10.1145/3486635.3491070","url":null,"abstract":"Many historical map sheets are publicly available for studies that require long-term historical geographic data. The cartographic design of these maps includes a combination of map symbols and text labels. Automatically reading text labels from map images could greatly speed up the map interpretation and helps generate rich metadata describing the map content. Many text detection algorithms have been proposed to locate text regions in map images automatically, but most of the algorithms are trained on out-of-domain datasets (e.g., scenic images). Training data determines the quality of machine learning models, and manually annotating text regions in map images is labor-extensive and time-consuming. On the other hand, existing geographic data sources, such as Open-StreetMap (OSM), contain machine-readable map layers, which allow us to separate out the text layer and obtain text label annotations easily. However, the cartographic styles between OSM map tiles and historical maps are significantly different. This paper proposes a method to automatically generate an unlimited amount of annotated historical map images for training text detection models. We use a style transfer model to convert contemporary map images into historical style and place text labels upon them. We show that the state-of-the-art text detection models (e.g., PSENet) can benefit from the synthetic historical maps and achieve significant improvement for historical map text detection.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128062318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Few-shot Learning for Post-disaster Structure Damage Assessment 灾后结构损伤评估的短时学习
Jordan Bowman, Lexie Yang
Automating post-disaster damage assessment with remote sensing data is critical for faster surveys of structures impacted by natural disasters. One significant obstacle to training state-of-the-art deep neural networks to support this automation is that large quantities of labelled data are often required. However, obtaining those labels is particularly unrealistic to support post-disaster damage assessment in a timely manner. Few-shot learning methods could help to mitigate this by reducing the amount of labelled data required to successfully train a model while achieving satisfactory results. To this end, we explore a feature reweighting method to the YOLOv3 object detection architecture to achieve few-shot learning of damage assessment models on the xBD dataset. Our results show that the feature reweighting approach yield improved mAP over the baseline with significantly fewer labelled samples. In addition, we use t-SNE to analyze the class-specific reweighting vectors generated by the reweighting module in order to evaluate their inter-class and intra-class similarity. We find that the vectors form clusters based on class, and that these clusters overlap with visually similar classes. Those results show the potential to employ this few-shot learning strategy for rapid damage assessment with post-event remote sensing images.
利用遥感数据进行灾后损害评估的自动化对于更快地调查受自然灾害影响的建筑物至关重要。训练最先进的深度神经网络来支持这种自动化的一个重大障碍是,通常需要大量的标记数据。然而,获得这些标签尤其不现实,无法及时支持灾后损害评估。通过减少成功训练模型所需的标记数据的数量,同时获得令人满意的结果,Few-shot学习方法可以帮助缓解这一问题。为此,我们探索了YOLOv3目标检测体系结构的特征重加权方法,以实现xBD数据集上损伤评估模型的少镜头学习。我们的结果表明,特征重加权方法在标记样本显著减少的情况下,在基线上产生了改进的mAP。此外,我们使用t-SNE来分析由重权模块生成的特定于类的重权向量,以评估它们的类间和类内相似性。我们发现向量形成基于类的簇,这些簇与视觉上相似的类重叠。这些结果显示了将这种少量学习策略应用于事件后遥感图像的快速损害评估的潜力。
{"title":"Few-shot Learning for Post-disaster Structure Damage Assessment","authors":"Jordan Bowman, Lexie Yang","doi":"10.1145/3486635.3491071","DOIUrl":"https://doi.org/10.1145/3486635.3491071","url":null,"abstract":"Automating post-disaster damage assessment with remote sensing data is critical for faster surveys of structures impacted by natural disasters. One significant obstacle to training state-of-the-art deep neural networks to support this automation is that large quantities of labelled data are often required. However, obtaining those labels is particularly unrealistic to support post-disaster damage assessment in a timely manner. Few-shot learning methods could help to mitigate this by reducing the amount of labelled data required to successfully train a model while achieving satisfactory results. To this end, we explore a feature reweighting method to the YOLOv3 object detection architecture to achieve few-shot learning of damage assessment models on the xBD dataset. Our results show that the feature reweighting approach yield improved mAP over the baseline with significantly fewer labelled samples. In addition, we use t-SNE to analyze the class-specific reweighting vectors generated by the reweighting module in order to evaluate their inter-class and intra-class similarity. We find that the vectors form clusters based on class, and that these clusters overlap with visually similar classes. Those results show the potential to employ this few-shot learning strategy for rapid damage assessment with post-event remote sensing images.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117177661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
VTSV
Jinmeng Rao, Song Gao, Xiaojin Zhu
Trajectory data is among the most sensitive data and the society increasingly raises privacy concerns. In this demo paper, we present a privacy-preserving Vehicle Trajectory Simulation and Visualization (VTSV) web platform (demo video: https://youtu.be/NY5L4bu2kTU), which automatically generates navigation routes between given pairs of origins and destinations and employs a deep reinforcement learning model to simulate vehicle trajectories with customized driving behaviors such as normal driving, overspeed, aggressive acceleration, and aggressive turning. The simulated vehicle trajectory data contain high-sample-rate of attributes including GPS location, speed, acceleration, and steering angle, and such data are visualized in VTSV using streetscape.gl, an autonomous driving data visualization framework. Location privacy protection methods such as origin-destination geomasking and trajectory k-anonymity are integrated into the platform to support privacy-preserving trajectory data generation and publication. We design two application scenarios to demonstrate how VTSV performs location privacy protection and customize driving behavior, respectively. The demonstration shows that VTSV is able to mitigate data privacy, sparsity, and imbalance sampling issues, which offers new insights into driving trajectory simulation and GeoAI-powered privacy-preserving data publication.
{"title":"VTSV","authors":"Jinmeng Rao, Song Gao, Xiaojin Zhu","doi":"10.1145/3486635.3491073","DOIUrl":"https://doi.org/10.1145/3486635.3491073","url":null,"abstract":"Trajectory data is among the most sensitive data and the society increasingly raises privacy concerns. In this demo paper, we present a privacy-preserving Vehicle Trajectory Simulation and Visualization (VTSV) web platform (demo video: https://youtu.be/NY5L4bu2kTU), which automatically generates navigation routes between given pairs of origins and destinations and employs a deep reinforcement learning model to simulate vehicle trajectories with customized driving behaviors such as normal driving, overspeed, aggressive acceleration, and aggressive turning. The simulated vehicle trajectory data contain high-sample-rate of attributes including GPS location, speed, acceleration, and steering angle, and such data are visualized in VTSV using streetscape.gl, an autonomous driving data visualization framework. Location privacy protection methods such as origin-destination geomasking and trajectory k-anonymity are integrated into the platform to support privacy-preserving trajectory data generation and publication. We design two application scenarios to demonstrate how VTSV performs location privacy protection and customize driving behavior, respectively. The demonstration shows that VTSV is able to mitigate data privacy, sparsity, and imbalance sampling issues, which offers new insights into driving trajectory simulation and GeoAI-powered privacy-preserving data publication.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125577245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Mapping Road Safety Barriers Across Street View Image Sequences: A Hybrid Object Detection and Recurrent Model 在街景图像序列中绘制道路安全屏障:一种混合目标检测和循环模型
Md. Mostafijur Rahman, Arpan Man Sainju, Dan Yan, Zhe Jiang
Road safety barriers (e.g., concrete barriers, metal crash barriers, rumble strips) play an important role in preventing or mitigating vehicle crashes. Accurate maps of road safety barriers are critical components of safety infrastructure management systems at federal or state transportation agencies. In current practice, mapping road safety barriers is largely done manually (e.g., driving on the road or visual interpretation of street view imagery), which is slow, tedious, and expensive. We propose a deep learning approach to automatically map road safety barriers from street view imagery. Our approach considers road barriers as long objects spanning across consecutive street view images in a sequence and use a hybrid object-detection and recurrent-network model. Preliminary results on real-world street view imagery show that the proposed model outperforms several baseline methods.
道路安全屏障(如混凝土屏障、金属防撞屏障、防撞带)在预防或减轻车辆碰撞方面发挥着重要作用。道路安全屏障的精确地图是联邦或州交通机构安全基础设施管理系统的重要组成部分。在目前的实践中,道路安全屏障的测绘主要是手动完成的(例如,在道路上驾驶或街景图像的视觉解释),这是缓慢、繁琐和昂贵的。我们提出了一种深度学习方法,从街景图像中自动绘制道路安全屏障。我们的方法将道路障碍物视为跨越连续街景图像的长物体,并使用混合物体检测和循环网络模型。在真实街景图像上的初步结果表明,该模型优于几种基线方法。
{"title":"Mapping Road Safety Barriers Across Street View Image Sequences: A Hybrid Object Detection and Recurrent Model","authors":"Md. Mostafijur Rahman, Arpan Man Sainju, Dan Yan, Zhe Jiang","doi":"10.1145/3486635.3491074","DOIUrl":"https://doi.org/10.1145/3486635.3491074","url":null,"abstract":"Road safety barriers (e.g., concrete barriers, metal crash barriers, rumble strips) play an important role in preventing or mitigating vehicle crashes. Accurate maps of road safety barriers are critical components of safety infrastructure management systems at federal or state transportation agencies. In current practice, mapping road safety barriers is largely done manually (e.g., driving on the road or visual interpretation of street view imagery), which is slow, tedious, and expensive. We propose a deep learning approach to automatically map road safety barriers from street view imagery. Our approach considers road barriers as long objects spanning across consecutive street view images in a sequence and use a hybrid object-detection and recurrent-network model. Preliminary results on real-world street view imagery show that the proposed model outperforms several baseline methods.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134403327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Conflation of Geospatial POI Data and Ground-level Imagery via Link Prediction on Joint Semantic Graph 基于联合语义图链接预测的地理空间POI数据与地面图像合并
Rutuja Gurav, Debraj De, Gautam S. Thakur, Junchuan Fan
With the proliferation of smartphone cameras and social networks, we have rich, multi-modal data about points of interest (POIs) - like cultural landmarks, institutions, businesses, etc. - within a given areas of interest (AOI) (e.g., a county, city or a neighborhood) available to us. Data conflation across multiple modalities of data sources is one of the key challenges in maintaining a geographical information system (GIS) which accumulate data about POIs. Given POI data from nine different sources, and ground-level geo-tagged and scene-captioned images from two different image hosting platforms, in this work we explore the application of graph neural networks (GNNs) to perform data conflation, while leveraging a natural graph structure evident in geospatial data. The preliminary results demonstrate the capacity of a GNN operation to learn distributions of entity (POIs and images) features, coupled with topological structure of entity's local neighborhood in a semantic nearest neighbor graph, in order to predict links between a pair of entities.
随着智能手机摄像头和社交网络的普及,我们拥有丰富的、多模式的兴趣点(poi)数据,比如文化地标、机构、企业等,这些数据都在特定的兴趣区域(AOI)内(例如,一个县、一个城市或一个社区)提供给我们。跨多种数据源模式的数据合并是维护地理信息系统(GIS)的关键挑战之一。鉴于来自九个不同来源的POI数据,以及来自两个不同图像托管平台的地面地理标记和场景字幕图像,在这项工作中,我们探索了图神经网络(gnn)在利用地理空间数据中明显的自然图结构的同时执行数据合并的应用。初步结果表明,GNN操作能够学习实体(poi和图像)特征的分布,并结合语义最近邻图中实体局部邻域的拓扑结构,从而预测一对实体之间的链接。
{"title":"Conflation of Geospatial POI Data and Ground-level Imagery via Link Prediction on Joint Semantic Graph","authors":"Rutuja Gurav, Debraj De, Gautam S. Thakur, Junchuan Fan","doi":"10.1145/3486635.3491068","DOIUrl":"https://doi.org/10.1145/3486635.3491068","url":null,"abstract":"With the proliferation of smartphone cameras and social networks, we have rich, multi-modal data about points of interest (POIs) - like cultural landmarks, institutions, businesses, etc. - within a given areas of interest (AOI) (e.g., a county, city or a neighborhood) available to us. Data conflation across multiple modalities of data sources is one of the key challenges in maintaining a geographical information system (GIS) which accumulate data about POIs. Given POI data from nine different sources, and ground-level geo-tagged and scene-captioned images from two different image hosting platforms, in this work we explore the application of graph neural networks (GNNs) to perform data conflation, while leveraging a natural graph structure evident in geospatial data. The preliminary results demonstrate the capacity of a GNN operation to learn distributions of entity (POIs and images) features, coupled with topological structure of entity's local neighborhood in a semantic nearest neighbor graph, in order to predict links between a pair of entities.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129415555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Semantic Segmentation in Aerial Images Using Class-Aware Unsupervised Domain Adaptation 基于类别感知的无监督域自适应航空图像语义分割
Ying Chen, Xu Ouyang, Kaiyue Zhu, G. Agam
Semantic segmentation using deep neural networks is an important component of aerial image understanding. However, models trained using data from one domain may not generalize well to another domain due to a domain shift between data distributions in the two domains. Such a domain gap is common in aerial images due to large visual appearance changes, and so substantial accuracy loss may occur when using a trained model for inference on new data. In this paper, we propose a novel unsupervised domain adaptation framework to address domain shift in the context of semantic segmentation of aerial images. To this end, we address the problem of domain shift by learning class-aware distribution differences between the source and target domains. Further, we employ entropy minimization on the target domain to produce high-confidence predictions. We demonstrate the effectiveness of the proposed approach using a challenge segmentation dataset by ISPRS, and show improvement over state-of-the-art methods.
基于深度神经网络的语义分割是航空图像理解的重要组成部分。然而,使用来自一个领域的数据训练的模型可能不能很好地推广到另一个领域,因为两个领域的数据分布之间存在领域转移。这种领域差距在航空图像中很常见,因为视觉外观变化很大,所以当使用训练模型对新数据进行推理时,可能会出现大量的精度损失。在本文中,我们提出了一种新的无监督域自适应框架来解决航空图像语义分割中的域转移问题。为此,我们通过学习源领域和目标领域之间的类感知分布差异来解决领域转移问题。此外,我们在目标域上使用熵最小化来产生高置信度的预测。我们使用ISPRS的挑战分割数据集证明了所提出方法的有效性,并展示了对最先进方法的改进。
{"title":"Semantic Segmentation in Aerial Images Using Class-Aware Unsupervised Domain Adaptation","authors":"Ying Chen, Xu Ouyang, Kaiyue Zhu, G. Agam","doi":"10.1145/3486635.3491069","DOIUrl":"https://doi.org/10.1145/3486635.3491069","url":null,"abstract":"Semantic segmentation using deep neural networks is an important component of aerial image understanding. However, models trained using data from one domain may not generalize well to another domain due to a domain shift between data distributions in the two domains. Such a domain gap is common in aerial images due to large visual appearance changes, and so substantial accuracy loss may occur when using a trained model for inference on new data. In this paper, we propose a novel unsupervised domain adaptation framework to address domain shift in the context of semantic segmentation of aerial images. To this end, we address the problem of domain shift by learning class-aware distribution differences between the source and target domains. Further, we employ entropy minimization on the target domain to produce high-confidence predictions. We demonstrate the effectiveness of the proposed approach using a challenge segmentation dataset by ISPRS, and show improvement over state-of-the-art methods.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131841638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Location Classification Based on Tweets 基于Tweets的位置分类
Elad Kravi, Y. Kanza, B. Kimelfeld, Roi Reichart
Location classification is used for associating type to locations, to enrich maps and support a plethora of geospatial applications that rely on location types. Classification can be performed by humans, but using machine learning is more efficient and faster to react to changes than human-based classification. Machine learning can be used in lieu of human classification or for supporting it. In this paper we study the use of machine learning for Geosocial Location Classification, where the type of a site, e.g., a building, is discovered based on social-media posts, e.g., tweets. Our goal is to correctly associate a set of tweets posted in a small radius around a given location with the corresponding location type, e.g., school, church, restaurant or museum. We explore two approaches to the problem: (a) a pipeline approach, where each post is first classified, and then the location associated with the set of posts is inferred from the individual post labels; and (b) a joint approach where the individual posts are simultaneously processed to yield the desired location type. We tested the two approaches over a data set of geotagged tweets. Our results demonstrate the superiority of the joint approach. Moreover, we show that due to the unique structure of the problem, where weakly-related messages are jointly processed to yield a single final label, linear classifiers outperform deep neural network alternatives.
位置分类用于将类型与位置关联起来,以丰富地图并支持依赖于位置类型的大量地理空间应用程序。分类可以由人类执行,但使用机器学习比基于人类的分类更有效,更快地对变化做出反应。机器学习可以用来代替人类分类或支持人类分类。在本文中,我们研究了机器学习在地理社会定位分类中的应用,其中一个网站的类型,例如,建筑物,是基于社交媒体帖子,例如,推特发现的。我们的目标是正确地将一组在给定位置周围的小半径内发布的tweet与相应的位置类型(例如,学校、教堂、餐馆或博物馆)关联起来。我们探索了两种方法来解决这个问题:(a)管道方法,首先对每个帖子进行分类,然后从单个帖子标签推断出与一组帖子相关的位置;(b)采用联合方法,同时处理个别员额,以产生所需的地点类型。我们在一组带有地理标记的tweet数据集上测试了这两种方法。我们的结果证明了联合方法的优越性。此外,我们表明,由于问题的独特结构,其中弱相关的消息被联合处理以产生单个最终标签,线性分类器优于深度神经网络替代品。
{"title":"Location Classification Based on Tweets","authors":"Elad Kravi, Y. Kanza, B. Kimelfeld, Roi Reichart","doi":"10.1145/3486635.3491075","DOIUrl":"https://doi.org/10.1145/3486635.3491075","url":null,"abstract":"Location classification is used for associating type to locations, to enrich maps and support a plethora of geospatial applications that rely on location types. Classification can be performed by humans, but using machine learning is more efficient and faster to react to changes than human-based classification. Machine learning can be used in lieu of human classification or for supporting it. In this paper we study the use of machine learning for Geosocial Location Classification, where the type of a site, e.g., a building, is discovered based on social-media posts, e.g., tweets. Our goal is to correctly associate a set of tweets posted in a small radius around a given location with the corresponding location type, e.g., school, church, restaurant or museum. We explore two approaches to the problem: (a) a pipeline approach, where each post is first classified, and then the location associated with the set of posts is inferred from the individual post labels; and (b) a joint approach where the individual posts are simultaneously processed to yield the desired location type. We tested the two approaches over a data set of geotagged tweets. Our results demonstrate the superiority of the joint approach. Moreover, we show that due to the unique structure of the problem, where weakly-related messages are jointly processed to yield a single final label, linear classifiers outperform deep neural network alternatives.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130874416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
hex2vec: Context-Aware Embedding H3 Hexagons with OpenStreetMap Tags hex2vec:上下文感知嵌入H3六边形与OpenStreetMap标签
Szymon Wo'zniak, Piotr Szyma'nski
Representation learning of spatial and geographic data is a rapidly developing field which allows for similarity detection between areas and high-quality inference using deep neural networks. Past approaches however concentrated on embedding raster imagery (maps, street or satellite photos), mobility data or road networks. In this paper we propose the first approach to learning vector representations of OpenStreetMap regions with respect to urban functions and land-use in a micro-region grid. We identify a subset of OSM tags related to major characteristics of land-use, building and urban region functions, types of water, green or other natural areas. Through manual verification of tagging quality, we selected 36 cities were for training region representations. Uber's H3 index was used to divide the cities into hexagons, and OSM tags were aggregated for each hexagon. We propose the hex2vec method based on the Skip-gram model with negative sampling. The resulting vector representations showcase semantic structures of the map characteristics, similar to ones found in vector-based language models. We also present insights from region similarity detection in six Polish cities and propose a region typology obtained through agglomerative clustering.
空间和地理数据的表示学习是一个快速发展的领域,它允许区域之间的相似性检测和使用深度神经网络进行高质量的推理。然而,过去的方法集中于嵌入光栅图像(地图、街道或卫星照片)、移动数据或道路网络。在本文中,我们提出了第一种方法来学习OpenStreetMap区域在微区域网格中关于城市功能和土地利用的向量表示。我们确定了与土地利用、建筑和城市区域功能、水、绿色或其他自然区域类型的主要特征相关的OSM标签子集。通过人工验证标注质量,我们选择了36个城市作为训练区域表示。使用Uber的H3指数将城市划分为六边形,并为每个六边形聚合OSM标签。我们提出了基于Skip-gram负采样模型的hex2vec方法。得到的向量表示展示了映射特征的语义结构,类似于基于向量的语言模型。我们还提出了从六个波兰城市的区域相似性检测的见解,并提出了通过聚集聚类获得的区域类型学。
{"title":"hex2vec: Context-Aware Embedding H3 Hexagons with OpenStreetMap Tags","authors":"Szymon Wo'zniak, Piotr Szyma'nski","doi":"10.1145/3486635.3491076","DOIUrl":"https://doi.org/10.1145/3486635.3491076","url":null,"abstract":"Representation learning of spatial and geographic data is a rapidly developing field which allows for similarity detection between areas and high-quality inference using deep neural networks. Past approaches however concentrated on embedding raster imagery (maps, street or satellite photos), mobility data or road networks. In this paper we propose the first approach to learning vector representations of OpenStreetMap regions with respect to urban functions and land-use in a micro-region grid. We identify a subset of OSM tags related to major characteristics of land-use, building and urban region functions, types of water, green or other natural areas. Through manual verification of tagging quality, we selected 36 cities were for training region representations. Uber's H3 index was used to divide the cities into hexagons, and OSM tags were aggregated for each hexagon. We propose the hex2vec method based on the Skip-gram model with negative sampling. The resulting vector representations showcase semantic structures of the map characteristics, similar to ones found in vector-based language models. We also present insights from region similarity detection in six Polish cities and propose a region typology obtained through agglomerative clustering.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128750671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Trinity: A No-Code AI platform for complex spatial datasets Trinity:用于复杂空间数据集的无代码AI平台
C. V. K. Iyer, Feili Hou, Henry Wang, Yonghong Wang, Kay Oh, Swetava Ganguli, Vipul Pandey
We present a no-code Artificial Intelligence (AI) platform called Trinity with the main design goal of enabling both machine learning researchers and non-technical geospatial domain experts to experiment with domain-specific signals and datasets for solving a variety of complex problems on their own. This versatility to solve diverse problems is achieved by transforming complex Spatio-temporal datasets to make them consumable by standard deep learning models, in this case, Convolutional Neural Networks (CNNs), and giving the ability to formulate disparate problems in a standard way, eg. semantic segmentation. With an intuitive user interface, a feature store that hosts derivatives of complex feature engineering, a deep learning kernel, and a scalable data processing mechanism, Trinity provides a powerful platform for domain experts to share the stage with scientists and engineers in solving business-critical problems. It enables quick prototyping, rapid experimentation and reduces the time to production by standardizing model building and deployment. In this paper, we present our motivation behind Trinity and its design along with showcasing sample applications to motivate the idea of lowering the bar to using AI.
我们提出了一个名为Trinity的无代码人工智能(AI)平台,其主要设计目标是使机器学习研究人员和非技术地理空间领域专家能够试验特定领域的信号和数据集,以自行解决各种复杂问题。这种解决不同问题的通用性是通过将复杂的时空数据集转换为标准深度学习模型(在本例中是卷积神经网络(cnn))可使用的数据集来实现的,并赋予以标准方式制定不同问题的能力,例如:语义分割。凭借直观的用户界面、承载复杂特征工程衍生产品的特征存储、深度学习内核和可扩展的数据处理机制,Trinity为领域专家提供了一个强大的平台,可以与科学家和工程师共享解决关键业务问题的平台。它支持快速原型、快速实验,并通过标准化模型构建和部署减少生产时间。在本文中,我们介绍了Trinity及其设计背后的动机,并展示了示例应用程序,以激发降低使用AI的门槛的想法。
{"title":"Trinity: A No-Code AI platform for complex spatial datasets","authors":"C. V. K. Iyer, Feili Hou, Henry Wang, Yonghong Wang, Kay Oh, Swetava Ganguli, Vipul Pandey","doi":"10.1145/3486635.3491072","DOIUrl":"https://doi.org/10.1145/3486635.3491072","url":null,"abstract":"We present a no-code Artificial Intelligence (AI) platform called Trinity with the main design goal of enabling both machine learning researchers and non-technical geospatial domain experts to experiment with domain-specific signals and datasets for solving a variety of complex problems on their own. This versatility to solve diverse problems is achieved by transforming complex Spatio-temporal datasets to make them consumable by standard deep learning models, in this case, Convolutional Neural Networks (CNNs), and giving the ability to formulate disparate problems in a standard way, eg. semantic segmentation. With an intuitive user interface, a feature store that hosts derivatives of complex feature engineering, a deep learning kernel, and a scalable data processing mechanism, Trinity provides a powerful platform for domain experts to share the stage with scientists and engineers in solving business-critical problems. It enables quick prototyping, rapid experimentation and reduces the time to production by standardizing model building and deployment. In this paper, we present our motivation behind Trinity and its design along with showcasing sample applications to motivate the idea of lowering the bar to using AI.","PeriodicalId":448866,"journal":{"name":"Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130987992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
Proceedings of the 4th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1