Abstract In this work, the geodesic equations and their numerical solution in Cartesian coordinates on an oblate spheroid, presented by Panou and Korakitis (2017), are generalized on a triaxial ellipsoid. A new exact analytical method and a new numerical method of converting Cartesian to ellipsoidal coordinates of a point on a triaxial ellipsoid are presented. An extensive test set for the coordinate conversion is used, in order to evaluate the performance of the two methods. The direct geodesic problem on a triaxial ellipsoid is described as an initial value problem and is solved numerically in Cartesian coordinates. The solution provides the Cartesian coordinates and the angle between the line of constant λ and the geodesic, at any point along the geodesic. Also, the Liouville constant is computed at any point along the geodesic, allowing to check the precision of the method. An extensive data set of geodesics is used, in order to demonstrate the validity of the numerical method for the geodesic problem. We conclude that a complete, stable and precise solution of the problem is accomplished.
{"title":"Geodesic equations and their numerical solution in Cartesian coordinates on a triaxial ellipsoid","authors":"G. Panou, R. Korakitis","doi":"10.1515/jogs-2019-0001","DOIUrl":"https://doi.org/10.1515/jogs-2019-0001","url":null,"abstract":"Abstract In this work, the geodesic equations and their numerical solution in Cartesian coordinates on an oblate spheroid, presented by Panou and Korakitis (2017), are generalized on a triaxial ellipsoid. A new exact analytical method and a new numerical method of converting Cartesian to ellipsoidal coordinates of a point on a triaxial ellipsoid are presented. An extensive test set for the coordinate conversion is used, in order to evaluate the performance of the two methods. The direct geodesic problem on a triaxial ellipsoid is described as an initial value problem and is solved numerically in Cartesian coordinates. The solution provides the Cartesian coordinates and the angle between the line of constant λ and the geodesic, at any point along the geodesic. Also, the Liouville constant is computed at any point along the geodesic, allowing to check the precision of the method. An extensive data set of geodesics is used, in order to demonstrate the validity of the numerical method for the geodesic problem. We conclude that a complete, stable and precise solution of the problem is accomplished.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"7 1","pages":"1 - 12"},"PeriodicalIF":1.3,"publicationDate":"2018-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72741609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract The Tibetan Plateau is the largest region of high elevation in the world. The source of water for a number of important rivers, the Himalayan region is vital to the billions of inhabitants of the Asian continent. Over the last fifty years, the climate in the region has warmed more rapidly than anywhere else at the same latitude. Causes and effects, and the geographical details of these alarming warming trends are as yet not fully known. One way of assessing the effects of climate change in this area is to measure the change in glacier volume in the region, but estimates made on the basis of different techniques have not been conclusive to date, and remain difficult to reconcile. We examine the temporal behavior of the mass flux integrated over four distinct groupings of Tibetan glaciers using satellite gravimetry from the Gravity Recovery and Climate Experiment (GRACE). We use a technique known as spatio-spectral localization using spherical Slepian functions to convert global spherical harmonic expansions of the time-dependent geopotential into monthly estimates of mass changes over the Tibetan Plateau. Subsequent reductions are aimed at interpreting this mass change as due to gains or losses in ice mass. We find that (ice) mass has been decreasing on the Tibetan Plateau between 2002 and 2016 but with significant spatial variability throughout the region. Specifically, in the regions of Himalaya, Pamir, Qilian, and Tien Shan, glaciers have been losing ice mass at a rate of −11±3, −1±2, +8±2, and −6±1 Gt/yr, respectively, over the last decade.
{"title":"The changing mass of glaciers on the Tibetan Plateau, 2002–2016, using time-variable gravity from the GRACE satellite mission","authors":"Alyson K. Beveridge, C. Harig, F. Simons","doi":"10.1515/jogs-2018-0010","DOIUrl":"https://doi.org/10.1515/jogs-2018-0010","url":null,"abstract":"Abstract The Tibetan Plateau is the largest region of high elevation in the world. The source of water for a number of important rivers, the Himalayan region is vital to the billions of inhabitants of the Asian continent. Over the last fifty years, the climate in the region has warmed more rapidly than anywhere else at the same latitude. Causes and effects, and the geographical details of these alarming warming trends are as yet not fully known. One way of assessing the effects of climate change in this area is to measure the change in glacier volume in the region, but estimates made on the basis of different techniques have not been conclusive to date, and remain difficult to reconcile. We examine the temporal behavior of the mass flux integrated over four distinct groupings of Tibetan glaciers using satellite gravimetry from the Gravity Recovery and Climate Experiment (GRACE). We use a technique known as spatio-spectral localization using spherical Slepian functions to convert global spherical harmonic expansions of the time-dependent geopotential into monthly estimates of mass changes over the Tibetan Plateau. Subsequent reductions are aimed at interpreting this mass change as due to gains or losses in ice mass. We find that (ice) mass has been decreasing on the Tibetan Plateau between 2002 and 2016 but with significant spatial variability throughout the region. Specifically, in the regions of Himalaya, Pamir, Qilian, and Tien Shan, glaciers have been losing ice mass at a rate of −11±3, −1±2, +8±2, and −6±1 Gt/yr, respectively, over the last decade.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"1 1","pages":"83 - 97"},"PeriodicalIF":1.3,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88096122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract In geodesy, hypothesis testing is applied to a wide area of applications e.g. outlier detection, deformation analysis or, more generally, model optimisation. Due to the possible far-reaching consequences of a decision, high statistical test power of such a hypothesis test is needed. The Neyman-Pearson lemma states that under strict assumptions the often-applied likelihood ratio test has highest statistical test power and may thus fulfill the requirement. The application, however, is made more difficult as most of the decision problems are non-linear and, thus, the probability density function of the parameters does not belong to the well-known set of statistical test distributions. Moreover, the statistical test power may change, if linear approximations of the likelihood ratio test are applied. The influence of the non-linearity on hypothesis testing is investigated and exemplified by the planar coordinate transformations. Whereas several mathematical equivalent expressions are conceivable to evaluate the rotation parameter of the transformation, the decisions and, thus, the probabilities of type 1 and 2 decision errors of the related hypothesis testing are unequal to each other. Based on Monte Carlo integration, the effective decision errors are estimated and used as a basis of valuation for linear and non-linear equivalents.
{"title":"Hypothesis Testing in Non-Linear Models Exemplified by the Planar Coordinate Transformations","authors":"R. Lehmann, M. Lösler","doi":"10.1515/jogs-2018-0009","DOIUrl":"https://doi.org/10.1515/jogs-2018-0009","url":null,"abstract":"Abstract In geodesy, hypothesis testing is applied to a wide area of applications e.g. outlier detection, deformation analysis or, more generally, model optimisation. Due to the possible far-reaching consequences of a decision, high statistical test power of such a hypothesis test is needed. The Neyman-Pearson lemma states that under strict assumptions the often-applied likelihood ratio test has highest statistical test power and may thus fulfill the requirement. The application, however, is made more difficult as most of the decision problems are non-linear and, thus, the probability density function of the parameters does not belong to the well-known set of statistical test distributions. Moreover, the statistical test power may change, if linear approximations of the likelihood ratio test are applied. The influence of the non-linearity on hypothesis testing is investigated and exemplified by the planar coordinate transformations. Whereas several mathematical equivalent expressions are conceivable to evaluate the rotation parameter of the transformation, the decisions and, thus, the probabilities of type 1 and 2 decision errors of the related hypothesis testing are unequal to each other. Based on Monte Carlo integration, the effective decision errors are estimated and used as a basis of valuation for linear and non-linear equivalents.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"43 1","pages":"114 - 98"},"PeriodicalIF":1.3,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75060911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract The aim of this paper is to present a quantitative analysis of the adequacy of the main currently existing combined Global Geopotential Models (GGMs) for modeling normal-geoid heights throughout Brazil. As major advances have been reached since mid-2016 in the combined GGMs elaboration and development, the main objective of this analysis is to verify if, in fact, the most recent models present superior or equivalent performance to the most performant previous models. The analysis was based on comparisons between normal-geoid height values obtained fromGNSS/leveling solutions and values calculated from GGMs XGM2016, GOCO05C, EIGEN-6C4 and EGM2008, according to different geopotential functionals - geoid height and height anomaly - and in different degrees of development, always through the relative method. This procedure was applied to 997 stations which carry information of both ellipsoidal and normal-orthometric heights, located all over Brazil. As a main result, it was observed the superior performance of the recent combined GGMs, GOCO05C and XGM2016, when compared to the older models, EIGEN-6C4 and EGM2008, when all of them are developed up to degree 720, the maximum degree of the recent models; and a approximate equality of results when all of the models are used in their individual maximum degrees.
{"title":"Evaluation of recent combined global geopotential models in Brazil","authors":"E. Nicacio, R. Dalazoana, S. Freitas","doi":"10.1515/jogs-2018-0008","DOIUrl":"https://doi.org/10.1515/jogs-2018-0008","url":null,"abstract":"Abstract The aim of this paper is to present a quantitative analysis of the adequacy of the main currently existing combined Global Geopotential Models (GGMs) for modeling normal-geoid heights throughout Brazil. As major advances have been reached since mid-2016 in the combined GGMs elaboration and development, the main objective of this analysis is to verify if, in fact, the most recent models present superior or equivalent performance to the most performant previous models. The analysis was based on comparisons between normal-geoid height values obtained fromGNSS/leveling solutions and values calculated from GGMs XGM2016, GOCO05C, EIGEN-6C4 and EGM2008, according to different geopotential functionals - geoid height and height anomaly - and in different degrees of development, always through the relative method. This procedure was applied to 997 stations which carry information of both ellipsoidal and normal-orthometric heights, located all over Brazil. As a main result, it was observed the superior performance of the recent combined GGMs, GOCO05C and XGM2016, when compared to the older models, EIGEN-6C4 and EGM2008, when all of them are developed up to degree 720, the maximum degree of the recent models; and a approximate equality of results when all of the models are used in their individual maximum degrees.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"10 1","pages":"72 - 82"},"PeriodicalIF":1.3,"publicationDate":"2018-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73195857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract This study provides additional information about the impact of atmospheric pressure on sea level variations. The observed regularity in sea level atmospheric pressure depends mainly on the latitude and verified to be dominantly random closer to the equator. It was demonstrated that almost all the annual and semiannual sea level variations at 27 globally distributed tide gauge stations can be attributed to the regional/local atmospheric forcing as an inverted barometric effect. Statistically significant non-linearities were detected in the regional atmospheric pressure series, which in turn impacted other sea level variations as compounders in tandem with the lunar nodal forcing, generating lunar sub-harmonics with multidecadal periods. It was shown that random component of regional atmospheric pressure tends to cluster at monthly intervals. The clusters are likely to be caused by the intraannual seasonal atmospheric temperature changes,which may also act as random beats in generating sub-harmonics observed in sea level changes as another mechanism. This study also affirmed that there are no statistically significant secular trends in the progression of regional atmospheric pressures, hence there was no contribution to the sea level trends during the 20th century by the atmospheric pressure.Meanwhile, the estimated nonuniform scale factors of the inverted barometer effects suggest that the sea level atmospheric pressure will bias the sea level trends inferred from satellite altimetry measurements if their impact is accounted for as corrections without proper scaling.
{"title":"The effect of regional sea level atmospheric pressure on sea level variations at globally distributed tide gauge stations with long records","authors":"H. Iz","doi":"10.1515/jogs-2018-0007","DOIUrl":"https://doi.org/10.1515/jogs-2018-0007","url":null,"abstract":"Abstract This study provides additional information about the impact of atmospheric pressure on sea level variations. The observed regularity in sea level atmospheric pressure depends mainly on the latitude and verified to be dominantly random closer to the equator. It was demonstrated that almost all the annual and semiannual sea level variations at 27 globally distributed tide gauge stations can be attributed to the regional/local atmospheric forcing as an inverted barometric effect. Statistically significant non-linearities were detected in the regional atmospheric pressure series, which in turn impacted other sea level variations as compounders in tandem with the lunar nodal forcing, generating lunar sub-harmonics with multidecadal periods. It was shown that random component of regional atmospheric pressure tends to cluster at monthly intervals. The clusters are likely to be caused by the intraannual seasonal atmospheric temperature changes,which may also act as random beats in generating sub-harmonics observed in sea level changes as another mechanism. This study also affirmed that there are no statistically significant secular trends in the progression of regional atmospheric pressures, hence there was no contribution to the sea level trends during the 20th century by the atmospheric pressure.Meanwhile, the estimated nonuniform scale factors of the inverted barometer effects suggest that the sea level atmospheric pressure will bias the sea level trends inferred from satellite altimetry measurements if their impact is accounted for as corrections without proper scaling.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"104 1","pages":"55 - 71"},"PeriodicalIF":1.3,"publicationDate":"2018-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82487930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Eshagh, Filippa Johansson, L. Karlsson, M. Horemuz
Abstract Monitoring deformation of man-made structures is very important to prevent them from a risk of collapse and save lives. Such a process is also used for monitoring change in historical objects, which are deforming continuously with time. An example of this is the Vasa warship, which was under water for about 300 years. The ship was raised from the bottom of the sea and is kept in the Vasa museum in Stockholm. A geodetic network with points on the museum building and the ship’s body has been established and measured for 12 years for monitoring the ship’s deformation. The coordinate time series of each point on the ship and their uncertainties have been estimated epoch-wisely. In this paper, our goal is to statistically analyse the ship’s hull movements. By fitting a quadratic polynomial to the coordinate time series of each point of the hull, its acceleration and velocity are estimated. In addition, their significance is tested by comparing them with their respective estimated errors after the fitting. Our numerical investigations show that the backside of the ship, having highest elevation and slope, has moved vertically faster than the other places by a velocity and an acceleration of about 2 mm/year and 0.1 mm/year2, respectively and this part of the ship is the weakest with a higher risk of collapse. The central parts of the ship are more stable as the ship hull is almost vertical and closer to the floor. Generally, the hull is moving towards its port and downwards
{"title":"A case study on displacement analysis of Vasa warship","authors":"M. Eshagh, Filippa Johansson, L. Karlsson, M. Horemuz","doi":"10.1515/jogs-2018-0006","DOIUrl":"https://doi.org/10.1515/jogs-2018-0006","url":null,"abstract":"Abstract Monitoring deformation of man-made structures is very important to prevent them from a risk of collapse and save lives. Such a process is also used for monitoring change in historical objects, which are deforming continuously with time. An example of this is the Vasa warship, which was under water for about 300 years. The ship was raised from the bottom of the sea and is kept in the Vasa museum in Stockholm. A geodetic network with points on the museum building and the ship’s body has been established and measured for 12 years for monitoring the ship’s deformation. The coordinate time series of each point on the ship and their uncertainties have been estimated epoch-wisely. In this paper, our goal is to statistically analyse the ship’s hull movements. By fitting a quadratic polynomial to the coordinate time series of each point of the hull, its acceleration and velocity are estimated. In addition, their significance is tested by comparing them with their respective estimated errors after the fitting. Our numerical investigations show that the backside of the ship, having highest elevation and slope, has moved vertically faster than the other places by a velocity and an acceleration of about 2 mm/year and 0.1 mm/year2, respectively and this part of the ship is the weakest with a higher risk of collapse. The central parts of the ship are more stable as the ship hull is almost vertical and closer to the floor. Generally, the hull is moving towards its port and downwards","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"37 1","pages":"43 - 54"},"PeriodicalIF":1.3,"publicationDate":"2018-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82234619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Lahtinen, Häkli Pasi, L. Jivall, C. Kempe, K. Kollo, Ksenija Kosenko, P. Pihlak, Dalia Prizginiene, O. Tangen, M. Weber, E. Paršeliūnas, Rimvydas Baniulis, Karolis Galinauskas
Abstract The Nordic Geodetic Commission (NKG) has launched a joint NKG GNSS Analysis Centre that aims to routinely produce high qualityGNSS solutions for the common needs of the NKG and the Nordic and Baltic countries. A consistent and densified velocity field is needed for the constraining of the gla-cial isostatic adjustment (GIA) modelling that is a key component of maintaining the national reference frame realisations in the area. We described the methods of the NKG GNSS Analysis Centre including the defined processing setup for the local analysis centres (LAC) and for the combination centres.We analysed the results of the first 2.5 years (2014.5-2016). The results showed that different subnets were consistent with the combined solution within 1-2 mm level. We observed the so called network effect affecting our reference frame alignment. However, the accuracy of the reference frame alignment was on a few millimetre level in the area of the main interest (Nordic and Baltic Countries). TheNKGGNSS AC was declared fully operational in April 2017.
{"title":"First results of the Nordic and Baltic GNSS Analysis Centre","authors":"S. Lahtinen, Häkli Pasi, L. Jivall, C. Kempe, K. Kollo, Ksenija Kosenko, P. Pihlak, Dalia Prizginiene, O. Tangen, M. Weber, E. Paršeliūnas, Rimvydas Baniulis, Karolis Galinauskas","doi":"10.1515/jogs-2018-0005","DOIUrl":"https://doi.org/10.1515/jogs-2018-0005","url":null,"abstract":"Abstract The Nordic Geodetic Commission (NKG) has launched a joint NKG GNSS Analysis Centre that aims to routinely produce high qualityGNSS solutions for the common needs of the NKG and the Nordic and Baltic countries. A consistent and densified velocity field is needed for the constraining of the gla-cial isostatic adjustment (GIA) modelling that is a key component of maintaining the national reference frame realisations in the area. We described the methods of the NKG GNSS Analysis Centre including the defined processing setup for the local analysis centres (LAC) and for the combination centres.We analysed the results of the first 2.5 years (2014.5-2016). The results showed that different subnets were consistent with the combined solution within 1-2 mm level. We observed the so called network effect affecting our reference frame alignment. However, the accuracy of the reference frame alignment was on a few millimetre level in the area of the main interest (Nordic and Baltic Countries). TheNKGGNSS AC was declared fully operational in April 2017.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"8 1","pages":"34 - 42"},"PeriodicalIF":1.3,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84471729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract It is well known that the success in precise determinations of the gravimetric geoid height (N) and the orthometric height (H) rely on the knowledge of the topographic mass distribution. We show that the residual topographic bias due to an imprecise information on the topographic density is practically the same for N and H, but with opposite signs. This result is demonstrated both for the Helmert orthometric height and for a more precise orthometric height derived by analytical continuation of the external geopotential to the geoid. This result leads to the conclusion that precise gravimetric geoid heights cannot be validated by GNSS-levelling geoid heights in mountainous regions for the errors caused by the incorrect modelling of the topographic mass distribution, because this uncertainty is hidden in the difference between the two geoid estimators.
{"title":"On the topographic bias and density distribution in modelling the geoid and orthometric heights","authors":"L. Sjöberg","doi":"10.1515/jogs-2018-0004","DOIUrl":"https://doi.org/10.1515/jogs-2018-0004","url":null,"abstract":"Abstract It is well known that the success in precise determinations of the gravimetric geoid height (N) and the orthometric height (H) rely on the knowledge of the topographic mass distribution. We show that the residual topographic bias due to an imprecise information on the topographic density is practically the same for N and H, but with opposite signs. This result is demonstrated both for the Helmert orthometric height and for a more precise orthometric height derived by analytical continuation of the external geopotential to the geoid. This result leads to the conclusion that precise gravimetric geoid heights cannot be validated by GNSS-levelling geoid heights in mountainous regions for the errors caused by the incorrect modelling of the topographic mass distribution, because this uncertainty is hidden in the difference between the two geoid estimators.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"111 1","pages":"30 - 33"},"PeriodicalIF":1.3,"publicationDate":"2018-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73304799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract In 1962 A. Bjerhammar introduced the method of analytical continuation in physical geodesy, implying that surface gravity anomalies are downward continued into the topographic masses down to an internal sphere (the Bjerhammar sphere). The method also includes analytical upward continuation of the potential to the surface of the Earth to obtain the quasigeoid. One can show that also the common remove-compute-restore technique for geoid determination includes an analytical continuation as long as the complete density distribution of the topography is not known. The analytical continuation implies that the downward continued gravity anomaly and/or potential are/is in error by the so-called topographic bias, which was postulated by a simple formula of L E Sjöberg in 2007. Here we will numerically test the postulated formula by comparing it with the bias obtained by analytical downward continuation of the external potential of a homogeneous ellipsoid to an inner sphere. The result shows that the postulated formula holds: At the equator of the ellipsoid, where the external potential is downward continued 21 km, the computed and postulated topographic biases agree to less than a millimetre (when the potential is scaled to the unit of metre).
摘要1962年A。Bjerhammar在物理大地测量学中引入了解析延拓方法,这意味着地表重力异常向下延续到地形团块中,直到一个内部球体(Bjerhammar球体)。该方法还包括分析向上延拓势到地球表面,以获得拟椭球面。可以证明,只要不知道地形的完整密度分布,用于大地水准面确定的常见的移除-计算-恢复技术也包括解析延拓。分析延拓表明,由于所谓的地形偏差,向下持续的重力异常和/或位势存在误差,而地形偏差是2007年用L E Sjöberg的简单公式假设的。在这里,我们将通过将假设公式与均匀椭球的外势向内球的解析向下延拓得到的偏差进行比较,从而在数值上验证该公式。结果表明,假设公式成立:在椭球的赤道处,外部电位向下持续21公里,计算和假设的地形偏差一致小于1毫米(当电位被缩放到米的单位时)。
{"title":"A numerical test of the topographic bias","authors":"L. Sjöberg, M. Joud","doi":"10.1515/jogs-2018-0002","DOIUrl":"https://doi.org/10.1515/jogs-2018-0002","url":null,"abstract":"Abstract In 1962 A. Bjerhammar introduced the method of analytical continuation in physical geodesy, implying that surface gravity anomalies are downward continued into the topographic masses down to an internal sphere (the Bjerhammar sphere). The method also includes analytical upward continuation of the potential to the surface of the Earth to obtain the quasigeoid. One can show that also the common remove-compute-restore technique for geoid determination includes an analytical continuation as long as the complete density distribution of the topography is not known. The analytical continuation implies that the downward continued gravity anomaly and/or potential are/is in error by the so-called topographic bias, which was postulated by a simple formula of L E Sjöberg in 2007. Here we will numerically test the postulated formula by comparing it with the bias obtained by analytical downward continuation of the external potential of a homogeneous ellipsoid to an inner sphere. The result shows that the postulated formula holds: At the equator of the ellipsoid, where the external potential is downward continued 21 km, the computed and postulated topographic biases agree to less than a millimetre (when the potential is scaled to the unit of metre).","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"50 1","pages":"14 - 17"},"PeriodicalIF":1.3,"publicationDate":"2018-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86841923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract The Bayesian approach allows an intuitive way to derive the methods of statistics. Probability is defined as a measure of the plausibility of statements or propositions. Three rules are sufficient to obtain the laws of probability. If the statements refer to the numerical values of variables, the so-called random variables, univariate and multivariate distributions follow. They lead to the point estimation by which unknown quantities, i.e. unknown parameters, are computed from measurements. The unknown parameters are random variables, they are fixed quantities in traditional statistics which is not founded on Bayes’ theorem. Bayesian statistics therefore recommends itself for Monte Carlo methods, which generate random variates from given distributions. Monte Carlo methods, of course, can also be applied in traditional statistics. The unknown parameters, are introduced as functions of the measurements, and the Monte Carlo methods give the covariance matrix and the expectation of these functions. A confidence region is derived where the unknown parameters are situated with a given probability. Following a method of traditional statistics, hypotheses are tested by determining whether a value for an unknown parameter lies inside or outside the confidence region. The error propagation of a random vector by the Monte Carlo methods is presented as an application. If the random vector results from a nonlinearly transformed vector, its covariance matrix and its expectation follow from the Monte Carlo estimate. This saves a considerable amount of derivatives to be computed, and errors of the linearization are avoided. The Monte Carlo method is therefore efficient. If the functions of the measurements are given by a sum of two or more random vectors with different multivariate distributions, the resulting distribution is generally not known. TheMonte Carlo methods are then needed to obtain the covariance matrix and the expectation of the sum.
{"title":"Bayesian statistics and Monte Carlo methods","authors":"K. Koch","doi":"10.1515/jogs-2018-0003","DOIUrl":"https://doi.org/10.1515/jogs-2018-0003","url":null,"abstract":"Abstract The Bayesian approach allows an intuitive way to derive the methods of statistics. Probability is defined as a measure of the plausibility of statements or propositions. Three rules are sufficient to obtain the laws of probability. If the statements refer to the numerical values of variables, the so-called random variables, univariate and multivariate distributions follow. They lead to the point estimation by which unknown quantities, i.e. unknown parameters, are computed from measurements. The unknown parameters are random variables, they are fixed quantities in traditional statistics which is not founded on Bayes’ theorem. Bayesian statistics therefore recommends itself for Monte Carlo methods, which generate random variates from given distributions. Monte Carlo methods, of course, can also be applied in traditional statistics. The unknown parameters, are introduced as functions of the measurements, and the Monte Carlo methods give the covariance matrix and the expectation of these functions. A confidence region is derived where the unknown parameters are situated with a given probability. Following a method of traditional statistics, hypotheses are tested by determining whether a value for an unknown parameter lies inside or outside the confidence region. The error propagation of a random vector by the Monte Carlo methods is presented as an application. If the random vector results from a nonlinearly transformed vector, its covariance matrix and its expectation follow from the Monte Carlo estimate. This saves a considerable amount of derivatives to be computed, and errors of the linearization are avoided. The Monte Carlo method is therefore efficient. If the functions of the measurements are given by a sum of two or more random vectors with different multivariate distributions, the resulting distribution is generally not known. TheMonte Carlo methods are then needed to obtain the covariance matrix and the expectation of the sum.","PeriodicalId":44569,"journal":{"name":"Journal of Geodetic Science","volume":"1 1","pages":"18 - 29"},"PeriodicalIF":1.3,"publicationDate":"2018-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84446788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}