Pub Date : 2022-06-01Epub Date: 2022-02-12DOI: 10.1145/3502221
Dejun Teng, Yanhui Liang, Hoang Vo, Jun Kong, Fusheng Wang
3D spatial data has been generated at an extreme scale from many emerging applications, such as high definition maps for autonomous driving and 3D Human BioMolecular Atlas. In particular, 3D digital pathology provides a revolutionary approach to map human tissues in 3D, which is highly promising for advancing computer-aided diagnosis and understanding diseases through spatial queries and analysis. However, the exponential increase of data at 3D leads to significant I/O, communication, and computational challenges for 3D spatial queries. The complex structures of 3D objects such as bifurcated vessels make it difficult to effectively support 3D spatial queries with traditional methods. In this article, we present our work on building an efficient and scalable spatial query system, iSPEED, for large-scale 3D data with complex structures. iSPEED adopts effective progressive compression for each 3D object with successive levels of detail. Further, iSPEED exploits structural indexing for complex structured objects in distance-based queries. By querying with data represented in successive levels of details and structural indexes, iSPEED provides an option for users to balance between query efficiency and query accuracy. iSPEED builds in-memory indexes and decompresses data on-demand, which has a minimal memory footprint. iSPEED provides a 3D spatial query engine that can be invoked on-demand to run many instances in parallel implemented with, but not limited to, MapReduce. We evaluate iSPEED with three representative queries: 3D spatial joins, 3D nearest neighbor query, and 3D spatial proximity estimation. The extensive experiments demonstrate that iSPEED significantly improves the performance of existing spatial query systems.
{"title":"Efficient 3D Spatial Queries for Complex Objects.","authors":"Dejun Teng, Yanhui Liang, Hoang Vo, Jun Kong, Fusheng Wang","doi":"10.1145/3502221","DOIUrl":"https://doi.org/10.1145/3502221","url":null,"abstract":"<p><p>3D spatial data has been generated at an extreme scale from many emerging applications, such as high definition maps for autonomous driving and 3D Human BioMolecular Atlas. In particular, 3D digital pathology provides a revolutionary approach to map human tissues in 3D, which is highly promising for advancing computer-aided diagnosis and understanding diseases through spatial queries and analysis. However, the exponential increase of data at 3D leads to significant I/O, communication, and computational challenges for 3D spatial queries. The complex structures of 3D objects such as bifurcated vessels make it difficult to effectively support 3D spatial queries with traditional methods. In this article, we present our work on building an efficient and scalable spatial query system, <i>iSPEED,</i> for large-scale 3D data with complex structures. iSPEED adopts effective progressive compression for each 3D object with successive levels of detail. Further, iSPEED exploits structural indexing for complex structured objects in distance-based queries. By querying with data represented in successive levels of details and structural indexes, iSPEED provides an option for users to balance between query efficiency and query accuracy. iSPEED builds in-memory indexes and decompresses data on-demand, which has a minimal memory footprint. iSPEED provides a 3D spatial query engine that can be invoked on-demand to run many instances in parallel implemented with, but not limited to, MapReduce. We evaluate iSPEED with three representative queries: 3D spatial joins, 3D nearest neighbor query, and 3D spatial proximity estimation. The extensive experiments demonstrate that iSPEED significantly improves the performance of existing spatial query systems.</p>","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9446285/pdf/nihms-1827081.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"33448539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hamada Rizk, H. Yamaguchi, Maged A. Youssef, T. Higashino
Robust and accurate indoor localization has been the goal of several research efforts over the past decade. Toward achieving this goal, WiFi fingerprinting-based indoor localization systems have been proposed. However, fingerprinting involves significant effort—especially when done at high density—and needs to be repeated with any change in the deployment area. While a number of recent systems have been introduced to reduce the calibration effort, these still trade overhead with accuracy. This article presents LiPhi++, an accurate system for enabling fingerprinting-based indoor localization systems without the associated data collection overhead. This is achieved by leveraging the sensing capability of transportable laser range scanners to automatically label WiFi scans, which can subsequently be used to build (and maintain) a fingerprint database. As part of its design, LiPhi++ leverages this database to train a deep long short-term memory network utilizing the signal strength history from the detected access points. LiPhi++ also has provisions for handling practical deployment issues, including the noisy wireless environment, heterogeneous devices, among others. Evaluation of LiPhi++ using Android phones in two realistic testbeds shows that it can match the performance of manual fingerprinting techniques under the same deployment conditions without the overhead associated with the traditional fingerprinting process. In addition, LiPhi++ improves upon the median localization accuracy obtained from crowdsourcing-based and fingerprinting-based systems by 284% and 418%, respectively, when tested with data collected a few months later.
{"title":"Laser Range Scanners for Enabling Zero-overhead WiFi-based Indoor Localization System","authors":"Hamada Rizk, H. Yamaguchi, Maged A. Youssef, T. Higashino","doi":"10.1145/3539659","DOIUrl":"https://doi.org/10.1145/3539659","url":null,"abstract":"Robust and accurate indoor localization has been the goal of several research efforts over the past decade. Toward achieving this goal, WiFi fingerprinting-based indoor localization systems have been proposed. However, fingerprinting involves significant effort—especially when done at high density—and needs to be repeated with any change in the deployment area. While a number of recent systems have been introduced to reduce the calibration effort, these still trade overhead with accuracy. This article presents LiPhi++, an accurate system for enabling fingerprinting-based indoor localization systems without the associated data collection overhead. This is achieved by leveraging the sensing capability of transportable laser range scanners to automatically label WiFi scans, which can subsequently be used to build (and maintain) a fingerprint database. As part of its design, LiPhi++ leverages this database to train a deep long short-term memory network utilizing the signal strength history from the detected access points. LiPhi++ also has provisions for handling practical deployment issues, including the noisy wireless environment, heterogeneous devices, among others. Evaluation of LiPhi++ using Android phones in two realistic testbeds shows that it can match the performance of manual fingerprinting techniques under the same deployment conditions without the overhead associated with the traditional fingerprinting process. In addition, LiPhi++ improves upon the median localization accuracy obtained from crowdsourcing-based and fingerprinting-based systems by 284% and 418%, respectively, when tested with data collected a few months later.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45056540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Philip E. Brown, Krystian Czapiga, Arun Jotshi, Y. Kanza, Velin Kounev, Poornima Suresh
Microwave backhaul links are often used as wireless connections between telecommunication towers, in places where deploying optical fibers is impossible or too expensive. The relatively high frequency of microwaves increases their ability to transfer information at a high rate, but it also makes them susceptible to spatial obstructions and interference. Hence, when deploying wireless links, there are two conflicting considerations. First, the antennas height, selected from the available slots on each tower, should be as low as possible. Second, there should be a line of sight (LoS) between the antennas, and a buffer around the LoS defined by the first Fresnel zone should be clear of obstacles. To compute antenna heights, a planning system for wireless links has to maintain an elevation model, efficiently discover obstacles between towers, and execute Fresnel-zone clearance tests over a 3D model of the deployment area. In this article we present a system and algorithms for computing the height of antennas, by testing LoS and clearance of Fresnel zones. The system handles the following requirements: (1) the need to cover large areas, e.g., all of the USA, (2) big distance between towers, e.g., 100 kilometers, and (3) computing batches of thousands of pairs within a few minutes. We introduce three novel algorithms for efficient computation of antenna heights, we show how to effectively model and manage the large-scale geospatial data needed for the planning, and we present the results of tests over real-world settings.
{"title":"Planning Wireless Backhaul Links by Testing Line of Sight and Fresnel Zone Clearance","authors":"Philip E. Brown, Krystian Czapiga, Arun Jotshi, Y. Kanza, Velin Kounev, Poornima Suresh","doi":"10.1145/3517382","DOIUrl":"https://doi.org/10.1145/3517382","url":null,"abstract":"Microwave backhaul links are often used as wireless connections between telecommunication towers, in places where deploying optical fibers is impossible or too expensive. The relatively high frequency of microwaves increases their ability to transfer information at a high rate, but it also makes them susceptible to spatial obstructions and interference. Hence, when deploying wireless links, there are two conflicting considerations. First, the antennas height, selected from the available slots on each tower, should be as low as possible. Second, there should be a line of sight (LoS) between the antennas, and a buffer around the LoS defined by the first Fresnel zone should be clear of obstacles. To compute antenna heights, a planning system for wireless links has to maintain an elevation model, efficiently discover obstacles between towers, and execute Fresnel-zone clearance tests over a 3D model of the deployment area. In this article we present a system and algorithms for computing the height of antennas, by testing LoS and clearance of Fresnel zones. The system handles the following requirements: (1) the need to cover large areas, e.g., all of the USA, (2) big distance between towers, e.g., 100 kilometers, and (3) computing batches of thousands of pairs within a few minutes. We introduce three novel algorithms for efficient computation of antenna heights, we show how to effectively model and manage the large-scale geospatial data needed for the planning, and we present the results of tests over real-world settings.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2022-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48891574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Traveling salesman problem (TSP) is one of the most studied combinatorial optimization problems; several exact, heuristic or even learning-based strategies have been proposed to solve this challenging issue. Targeting on the research problem of bi-objective non-monotonic Euclidean TSP and based on the concept of the multi-agent-based approach, we propose a two-stage parallel matching approaching for solving TSP. Acting as a divide-and-conquer strategy, the merit lies in the simultaneously clustering and routing in the dividing process. Precisely, we first propose the Two-Stage Parallel Matching algorithm (TSPM) to deal with the bi-objective TSP. We then formulate the Grid-Based Two-Stage Parallel Matching (GRAPE) framework, which can synergize with TSPM, exact method, or other state-of-the-art TSP solvers, for solving large-scale Euclidean TSP. According to this framework, the original problem space is divided into smaller regions and then computed in parallel, which helps to tackle and derive solutions for larger-scale Euclidean TSP within reasonable computational resources. Preliminary evaluation based on TSPLIB testbed shows that our proposed GRAPE framework holds a decent quality of solutions in especially runtime for large-scale Euclidean TSP. Meanwhile, experiments conducted on two real-world datasets demonstrate the efficacy and adaptability of our proposed TSPM in solving the bi-objective non-monotonic TSP.
{"title":"A Grid-Based Two-Stage Parallel Matching Framework for Bi-Objective Euclidean Traveling Salesman Problem","authors":"Fandel Lin, Hsun-Ping Hsieh","doi":"10.1145/3526025","DOIUrl":"https://doi.org/10.1145/3526025","url":null,"abstract":"Traveling salesman problem (TSP) is one of the most studied combinatorial optimization problems; several exact, heuristic or even learning-based strategies have been proposed to solve this challenging issue. Targeting on the research problem of bi-objective non-monotonic Euclidean TSP and based on the concept of the multi-agent-based approach, we propose a two-stage parallel matching approaching for solving TSP. Acting as a divide-and-conquer strategy, the merit lies in the simultaneously clustering and routing in the dividing process. Precisely, we first propose the Two-Stage Parallel Matching algorithm (TSPM) to deal with the bi-objective TSP. We then formulate the Grid-Based Two-Stage Parallel Matching (GRAPE) framework, which can synergize with TSPM, exact method, or other state-of-the-art TSP solvers, for solving large-scale Euclidean TSP. According to this framework, the original problem space is divided into smaller regions and then computed in parallel, which helps to tackle and derive solutions for larger-scale Euclidean TSP within reasonable computational resources. Preliminary evaluation based on TSPLIB testbed shows that our proposed GRAPE framework holds a decent quality of solutions in especially runtime for large-scale Euclidean TSP. Meanwhile, experiments conducted on two real-world datasets demonstrate the efficacy and adaptability of our proposed TSPM in solving the bi-objective non-monotonic TSP.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2022-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46960368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Bayesian networks are a popular class of multivariate probabilistic models as they allow for the translation of prior beliefs about conditional dependencies between variables to be easily encoded into their model structure. Due to their widespread usage, they are often applied to spatial data for inferring properties of the systems under study and also generating predictions for how these systems may behave in the future. We review published research on methodologies for representing spatial data with Bayesian networks and also summarize the application areas for which Bayesian networks are employed in the modeling of spatial data. We find that a wide variety of perspectives are taken, including a GIS-centric focus on efficiently generating geospatial predictions, a statistical focus on rigorously constructing graphical models controlling for spatial correlation, as well as a range of problem-specific heuristics for mitigating the effects of spatial correlation and dependency arising in spatial data analysis. Special attention is also paid to potential future directions for the integration of Bayesian networks with spatial processes.
{"title":"A Review of Bayesian Networks for Spatial Data","authors":"C. Krapu, R. Stewart, A. Rose","doi":"10.1145/3516523","DOIUrl":"https://doi.org/10.1145/3516523","url":null,"abstract":"Bayesian networks are a popular class of multivariate probabilistic models as they allow for the translation of prior beliefs about conditional dependencies between variables to be easily encoded into their model structure. Due to their widespread usage, they are often applied to spatial data for inferring properties of the systems under study and also generating predictions for how these systems may behave in the future. We review published research on methodologies for representing spatial data with Bayesian networks and also summarize the application areas for which Bayesian networks are employed in the modeling of spatial data. We find that a wide variety of perspectives are taken, including a GIS-centric focus on efficiently generating geospatial predictions, a statistical focus on rigorously constructing graphical models controlling for spatial correlation, as well as a range of problem-specific heuristics for mitigating the effects of spatial correlation and dependency arising in spatial data analysis. Special attention is also paid to potential future directions for the integration of Bayesian networks with spatial processes.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2022-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44924090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Urban planning refers to the efforts of designing land-use configurations given a region. However, to obtain effective urban plans, urban experts have to spend much time and effort analyzing sophisticated planning constraints based on domain knowledge and personal experiences. To alleviate the heavy burden of them and produce consistent urban plans, we want to ask that can AI accelerate the urban planning process, so that human planners only adjust generated configurations for specific needs? The recent advance of deep generative models provides a possible answer, which inspires us to automate urban planning from an adversarial learning perspective. However, three major challenges arise: (1) how to define a quantitative land-use configuration? (2) how to automate configuration planning? (3) how to evaluate the quality of a generated configuration? In this article, we systematically address the three challenges. Specifically, (1) We define a land-use configuration as a longitude-latitude-channel tensor. (2) We formulate the automated urban planning problem into a task of deep generative learning. The objective is to generate a configuration tensor given the surrounding contexts of a target region. In particular, we first construct spatial graphs using geographic and human mobility data crawled from websites to learn graph representations. We then combine each target area and its surrounding context representations as a tuple, and categorize all tuples into positive (well-planned areas) and negative samples (poorly-planned areas). Next, we develop an adversarial learning framework, in which a generator takes the surrounding context representations as input to generate a land-use configuration, and a discriminator learns to distinguish between positive and negative samples. (3) We provide quantitative evaluation metrics and conduct extensive experiments to demonstrate the effectiveness of our framework.
{"title":"Automated Urban Planning for Reimagining City Configuration via Adversarial Learning: Quantification, Generation, and Evaluation","authors":"Dongjie Wang, Yanjie Fu, Kunpeng Liu, Fanglan Chen, Pengyang Wang, Chang-Tien Lu","doi":"10.1145/3524302","DOIUrl":"https://doi.org/10.1145/3524302","url":null,"abstract":"Urban planning refers to the efforts of designing land-use configurations given a region. However, to obtain effective urban plans, urban experts have to spend much time and effort analyzing sophisticated planning constraints based on domain knowledge and personal experiences. To alleviate the heavy burden of them and produce consistent urban plans, we want to ask that can AI accelerate the urban planning process, so that human planners only adjust generated configurations for specific needs? The recent advance of deep generative models provides a possible answer, which inspires us to automate urban planning from an adversarial learning perspective. However, three major challenges arise: (1) how to define a quantitative land-use configuration? (2) how to automate configuration planning? (3) how to evaluate the quality of a generated configuration? In this article, we systematically address the three challenges. Specifically, (1) We define a land-use configuration as a longitude-latitude-channel tensor. (2) We formulate the automated urban planning problem into a task of deep generative learning. The objective is to generate a configuration tensor given the surrounding contexts of a target region. In particular, we first construct spatial graphs using geographic and human mobility data crawled from websites to learn graph representations. We then combine each target area and its surrounding context representations as a tuple, and categorize all tuples into positive (well-planned areas) and negative samples (poorly-planned areas). Next, we develop an adversarial learning framework, in which a generator takes the surrounding context representations as input to generate a land-use configuration, and a discriminator learns to distinguish between positive and negative samples. (3) We provide quantitative evaluation metrics and conduct extensive experiments to demonstrate the effectiveness of our framework.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2021-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43034813","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
O. Wolfson, Prabin Giri, S. Jajodia, Goce Trajcevski
We consider surveillance of a geographic region by a collaborative system of drones. The drones assist each other in identifying and managing activities of interest on the ground. We also consider an adversary who can create both genuine and fake activities on the ground. The objective of the adversary is to use fake activities to maximize the response time to genuine activities. We present two collaboration algorithms and analyze their response times, as well as the adversary’s efforts in terms of the number of fake activities required to achieve a certain response time.
{"title":"Geographic-Region Monitoring by Drones in Adversarial Environments","authors":"O. Wolfson, Prabin Giri, S. Jajodia, Goce Trajcevski","doi":"10.1145/3611009","DOIUrl":"https://doi.org/10.1145/3611009","url":null,"abstract":"We consider surveillance of a geographic region by a collaborative system of drones. The drones assist each other in identifying and managing activities of interest on the ground. We also consider an adversary who can create both genuine and fake activities on the ground. The objective of the adversary is to use fake activities to maximize the response time to genuine activities. We present two collaboration algorithms and analyze their response times, as well as the adversary’s efforts in terms of the number of fake activities required to achieve a certain response time.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44787657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In the last decade, more and more spatial data has been acquired on a global scale due to satellite missions, social media, and coordinated governmental activities. This observational data suffers from huge storage footprints and makes global analysis challenging. Therefore, many information products have been designed in which observations are turned into global maps showing features such as land cover or land use, often with only a few discrete values and sparse spatial coverage like only within cities. Traditional coding of such data as a raster image becomes challenging due to the sizes of the datasets and spatially non-local access patterns, for example, when labeling social media streams. This article proposes GloBiMap, a randomized data structure, based on Bloom filters, for modeling low-cardinality sparse raster images of excessive sizes in a configurable amount of memory with pure random access operations avoiding costly intermediate decompression. In addition, the data structure is designed to correct the inevitable errors of the randomized layer in order to have a fully exact representation. We show the feasibility of the approach on several real-world datasets including the Global Urban Footprint in which each pixel denotes whether a particular location contains a building at a resolution of roughly 10m globally as well as on a global Twitter sample of more than 220 million precisely geolocated tweets. In addition, we propose the integration of a denoiser engine based on artificial intelligence in order to reduce the amount of error correction information for extremely compressive GloBiMaps.
{"title":"GloBiMapsAI: An AI-Enhanced Probabilistic Data Structure for Global Raster Datasets","authors":"M. Werner","doi":"10.1145/3453184","DOIUrl":"https://doi.org/10.1145/3453184","url":null,"abstract":"In the last decade, more and more spatial data has been acquired on a global scale due to satellite missions, social media, and coordinated governmental activities. This observational data suffers from huge storage footprints and makes global analysis challenging. Therefore, many information products have been designed in which observations are turned into global maps showing features such as land cover or land use, often with only a few discrete values and sparse spatial coverage like only within cities. Traditional coding of such data as a raster image becomes challenging due to the sizes of the datasets and spatially non-local access patterns, for example, when labeling social media streams. This article proposes GloBiMap, a randomized data structure, based on Bloom filters, for modeling low-cardinality sparse raster images of excessive sizes in a configurable amount of memory with pure random access operations avoiding costly intermediate decompression. In addition, the data structure is designed to correct the inevitable errors of the randomized layer in order to have a fully exact representation. We show the feasibility of the approach on several real-world datasets including the Global Urban Footprint in which each pixel denotes whether a particular location contains a building at a resolution of roughly 10m globally as well as on a global Twitter sample of more than 220 million precisely geolocated tweets. In addition, we propose the integration of a denoiser engine based on artificial intelligence in order to reduce the amount of error correction information for extremely compressive GloBiMaps.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2021-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3453184","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44081864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
D. Teixeira, A. C. Viana, J. Almeida, Mrio S. Alvim
Predicting mobility-related behavior is an important yet challenging task. On the one hand, factors such as one’s routine or preferences for a few favorite locations may help in predicting their mobility. On the other hand, several contextual factors, such as variations in individual preferences, weather, traffic, or even a person’s social contacts, can affect mobility patterns and make its modeling significantly more challenging. A fundamental approach to study mobility-related behavior is to assess how predictable such behavior is, deriving theoretical limits on the accuracy that a prediction model can achieve given a specific dataset. This approach focuses on the inherent nature and fundamental patterns of human behavior captured in that dataset, filtering out factors that depend on the specificities of the prediction method adopted. However, the current state-of-the-art method to estimate predictability in human mobility suffers from two major limitations: low interpretability and hardness to incorporate external factors that are known to help mobility prediction (i.e., contextual information). In this article, we revisit this state-of-the-art method, aiming at tackling these limitations. Specifically, we conduct a thorough analysis of how this widely used method works by looking into two different metrics that are easier to understand and, at the same time, capture reasonably well the effects of the original technique. We evaluate these metrics in the context of two different mobility prediction tasks, notably, next cell and next distinct cell prediction, which have different degrees of difficulty. Additionally, we propose alternative strategies to incorporate different types of contextual information into the existing technique. Our evaluation of these strategies offer quantitative measures of the impact of adding context to the predictability estimate, revealing the challenges associated with doing so in practical scenarios.
{"title":"The Impact of Stationarity, Regularity, and Context on the Predictability of Individual Human Mobility","authors":"D. Teixeira, A. C. Viana, J. Almeida, Mrio S. Alvim","doi":"10.1145/3459625","DOIUrl":"https://doi.org/10.1145/3459625","url":null,"abstract":"Predicting mobility-related behavior is an important yet challenging task. On the one hand, factors such as one’s routine or preferences for a few favorite locations may help in predicting their mobility. On the other hand, several contextual factors, such as variations in individual preferences, weather, traffic, or even a person’s social contacts, can affect mobility patterns and make its modeling significantly more challenging. A fundamental approach to study mobility-related behavior is to assess how predictable such behavior is, deriving theoretical limits on the accuracy that a prediction model can achieve given a specific dataset. This approach focuses on the inherent nature and fundamental patterns of human behavior captured in that dataset, filtering out factors that depend on the specificities of the prediction method adopted. However, the current state-of-the-art method to estimate predictability in human mobility suffers from two major limitations: low interpretability and hardness to incorporate external factors that are known to help mobility prediction (i.e., contextual information). In this article, we revisit this state-of-the-art method, aiming at tackling these limitations. Specifically, we conduct a thorough analysis of how this widely used method works by looking into two different metrics that are easier to understand and, at the same time, capture reasonably well the effects of the original technique. We evaluate these metrics in the context of two different mobility prediction tasks, notably, next cell and next distinct cell prediction, which have different degrees of difficulty. Additionally, we propose alternative strategies to incorporate different types of contextual information into the existing technique. Our evaluation of these strategies offer quantitative measures of the impact of adding context to the predictability estimate, revealing the challenges associated with doing so in practical scenarios.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2021-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1145/3459625","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43962512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Zakaria Mehrab, A. Adiga, M. Marathe, S. Venkatramanan, S. Swarup
High resolution mobility datasets have become increasingly available in the past few years and have enabled detailed models for infectious disease spread including those for COVID-19. However, there are open questions on how such mobility data can be used effectively within epidemic models and for which tasks they are best suited. In this paper, we extract a number of graph-based proximity metrics from high resolution cellphone trace data from X-Mode and use it to study COVID-19 epidemic spread in 50 land grant university counties in the US. We present an approach to estimate the effect of mobility on cases by fitting an ordinary differential equation based model and performing multivariate linear regression to explain the estimated time varying transmissibility. We find that, while mobility plays a significant role, the contribution is heterogeneous across the counties, as exemplified by a subsequent correlation analysis. We also evaluate the metrics’ utility for case surge prediction defined as a supervised classification problem, and show that the learnt model can predict surges with 95% accuracy and an 87% F1-score.
{"title":"Evaluating the Utility of High-Resolution Proximity Metrics in Predicting the Spread of COVID-19","authors":"Zakaria Mehrab, A. Adiga, M. Marathe, S. Venkatramanan, S. Swarup","doi":"10.1145/3531006","DOIUrl":"https://doi.org/10.1145/3531006","url":null,"abstract":"High resolution mobility datasets have become increasingly available in the past few years and have enabled detailed models for infectious disease spread including those for COVID-19. However, there are open questions on how such mobility data can be used effectively within epidemic models and for which tasks they are best suited. In this paper, we extract a number of graph-based proximity metrics from high resolution cellphone trace data from X-Mode and use it to study COVID-19 epidemic spread in 50 land grant university counties in the US. We present an approach to estimate the effect of mobility on cases by fitting an ordinary differential equation based model and performing multivariate linear regression to explain the estimated time varying transmissibility. We find that, while mobility plays a significant role, the contribution is heterogeneous across the counties, as exemplified by a subsequent correlation analysis. We also evaluate the metrics’ utility for case surge prediction defined as a supervised classification problem, and show that the learnt model can predict surges with 95% accuracy and an 87% F1-score.","PeriodicalId":43641,"journal":{"name":"ACM Transactions on Spatial Algorithms and Systems","volume":null,"pages":null},"PeriodicalIF":1.9,"publicationDate":"2021-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47112596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}