Pub Date : 2024-03-27DOI: 10.1007/s10614-024-10586-5
Andrés Navarro-Galera, Juan Lara-Rubio, Pavel Novoa-Hernández, Carlos A. Cruz Corona
In today’s economic landscape, with its increasingly brief economic cycles and ever-changing market conditions, forecasting has become more critical than ever. In the specific case of small and medium-sized enterprises (SMEs), a crucial aspect is to anticipate the state of bankruptcy due to the low life expectancy of this type of company. A requirement that has been recommended by several international organizations such as the European Union, especially because SMEs contribute significantly to job creation and added value and to overcoming the effects of economic crises. Despite the progress in this field, there are economies that have been little or poorly addressed by the literature. This is the case for Spain, an economy where SMEs account for a significant share of its business landscape. To close this gap, this paper addressed the problem of predicting the insolvency of Spanish SMEs from a Machine Learning perspective. Leveraging a dataset encompassing financial and non-financial data from 58,267 Spanish SMEs spanning the period 2009–2020, we adjusted several decision tree models to address two scenarios of practical value in the Spanish context. Additionally, we conducted a thorough analysis of the most influential predictors of insolvency from a financial perspective. To empower Spanish SMEs, we provided them with a free software tool implementing the best models for the considered scenarios. The tool is intended to serve as an additional means to proactively and early assess solvency status.
{"title":"Using Decision Trees to Predict Insolvency in Spanish SMEs: Is Early Warning Possible?","authors":"Andrés Navarro-Galera, Juan Lara-Rubio, Pavel Novoa-Hernández, Carlos A. Cruz Corona","doi":"10.1007/s10614-024-10586-5","DOIUrl":"https://doi.org/10.1007/s10614-024-10586-5","url":null,"abstract":"<p>In today’s economic landscape, with its increasingly brief economic cycles and ever-changing market conditions, forecasting has become more critical than ever. In the specific case of small and medium-sized enterprises (SMEs), a crucial aspect is to anticipate the state of bankruptcy due to the low life expectancy of this type of company. A requirement that has been recommended by several international organizations such as the European Union, especially because SMEs contribute significantly to job creation and added value and to overcoming the effects of economic crises. Despite the progress in this field, there are economies that have been little or poorly addressed by the literature. This is the case for Spain, an economy where SMEs account for a significant share of its business landscape. To close this gap, this paper addressed the problem of predicting the insolvency of Spanish SMEs from a Machine Learning perspective. Leveraging a dataset encompassing financial and non-financial data from 58,267 Spanish SMEs spanning the period 2009–2020, we adjusted several decision tree models to address two scenarios of practical value in the Spanish context. Additionally, we conducted a thorough analysis of the most influential predictors of insolvency from a financial perspective. To empower Spanish SMEs, we provided them with a free software tool implementing the best models for the considered scenarios. The tool is intended to serve as an additional means to proactively and early assess solvency status.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"13 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140322276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-25DOI: 10.1007/s10614-024-10565-w
Abstract
The aim of nearly all empirical studies in economics is to provide scientific evidence that can be used to assess policy relevant cause-and-effect. In the context of the general potential outcomes framework, we review how a causal effect parameter can be rigorously but tractably specified, identified and estimated along with its asymptotic standard error. For cases in which the analytic and computational requirements for calculation of the ASE are challenging, we suggest the use of numerical derivatives (ND). We detail the specific type of ND software required for this purpose, and note that it is offered as a feature in most statistical packages. As an illustration, we analyze the causal effect of wife's high school graduation on family size using the Stata/Mata deriv command. Code for this example is supplied in an appendix.
摘要 几乎所有经济学实证研究的目的都是提供可用于评估政策相关因果关系的科学证据。在一般潜在结果框架的背景下,我们回顾了如何严格而简便地指定、确定和估计因果效应参数及其渐近标准误差。如果计算 ASE 的分析和计算要求具有挑战性,我们建议使用数值导数(ND)。我们将详细介绍为此目的所需的 ND 软件的具体类型,并指出大多数统计软件包都提供该功能。作为示例,我们使用 Stata/Mata deriv 命令分析了妻子高中毕业对家庭规模的因果效应。本例的代码见附录。
{"title":"Standard Errors for Regression-Based Causal Effect Estimates in Economics Using Numerical Derivatives","authors":"","doi":"10.1007/s10614-024-10565-w","DOIUrl":"https://doi.org/10.1007/s10614-024-10565-w","url":null,"abstract":"<h3>Abstract</h3> <p>The aim of nearly all empirical studies in economics is to provide scientific evidence that can be used to assess policy relevant cause-and-effect. In the context of the general potential outcomes framework, we review how a causal effect parameter can be rigorously but tractably specified, identified and estimated along with its asymptotic standard error. For cases in which the analytic and computational requirements for calculation of the ASE are challenging, we suggest the use of numerical derivatives (ND). We detail the specific type of ND software required for this purpose, and note that it is offered as a feature in most statistical packages. As an illustration, we analyze the causal effect of wife's high school graduation on family size using the Stata/Mata <strong>deriv</strong> command. Code for this example is supplied in an appendix.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"29 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140302083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-24DOI: 10.1007/s10614-024-10569-6
Halil Ibrahim Gunduz, Furkan Emirmahmutoglu, M. Eray Yucel
Requirements to understand and forecast the behavior of complex macroeconomic interactions mandate the use of high-dimensional macroeconometric models. The Global Vector Autoregressive (GVAR) modeling technique is very popular among them and it allows researchers and policymakers to take into account both the complex interdependencies that exist between various economic entities and the global economy through the world’s trade and financial channels. However, determining the cross-section unit size while using this approach is not a trivial task. In order to address this issue, we suggest an objective procedure for the detection of the size of the cross-country aggregation in GVAR models. While doing so, we depart from the Akaike Information Criterion (AIC) and propose an analytical modification to it, mainly employing an ad hoc approach without violating Akaike’s main principles. To supplement the theoretical results, small sample performances of those procedures are studied in Monte Carlo experiments as well as implementing our approach on real data. The numerical results suggest that our ad hoc modification of AIC can be used to determine the structure of the cross-section unit dimension in GVAR models, allowing the researchers and policymakers to build parsimonious models.
{"title":"A New Look at Cross-Country Aggregation in the Global VAR Approach: Theory and Monte Carlo Simulation","authors":"Halil Ibrahim Gunduz, Furkan Emirmahmutoglu, M. Eray Yucel","doi":"10.1007/s10614-024-10569-6","DOIUrl":"https://doi.org/10.1007/s10614-024-10569-6","url":null,"abstract":"<p>Requirements to understand and forecast the behavior of complex macroeconomic interactions mandate the use of high-dimensional macroeconometric models. The Global Vector Autoregressive (GVAR) modeling technique is very popular among them and it allows researchers and policymakers to take into account both the complex interdependencies that exist between various economic entities and the global economy through the world’s trade and financial channels. However, determining the cross-section unit size while using this approach is not a trivial task. In order to address this issue, we suggest an objective procedure for the detection of the size of the cross-country aggregation in GVAR models. While doing so, we depart from the Akaike Information Criterion (AIC) and propose an analytical modification to it, mainly employing an ad hoc approach without violating Akaike’s main principles. To supplement the theoretical results, small sample performances of those procedures are studied in Monte Carlo experiments as well as implementing our approach on real data. The numerical results suggest that our ad hoc modification of AIC can be used to determine the structure of the cross-section unit dimension in GVAR models, allowing the researchers and policymakers to build parsimonious models.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"66 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140302208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-21DOI: 10.1007/s10614-024-10576-7
Sanjay Bhattacherjee, Palash Sarkar
The first observation of the paper is that methods for determining proportional representation in electoral systems may be suitable as alternatives to the pro-rata order matching algorithm used in stock exchanges. The main part of our work is to comprehensively consider various well known proportional representation methods and analyse in details their suitability for replacing the pro-rata algorithm. Our analysis consists of a theoretical study as well as simulation studies based on data sampled from a distribution which has been suggested in the literature as models of limit orders. Based on our analysis, we put forward the suggestion that the well known Hamilton’s method is a superior alternative to the pro-rata algorithm for order matching applications.
{"title":"On Using Proportional Representation Methods as Alternatives to Pro-rata Based Order Matching Algorithms in Stock Exchanges","authors":"Sanjay Bhattacherjee, Palash Sarkar","doi":"10.1007/s10614-024-10576-7","DOIUrl":"https://doi.org/10.1007/s10614-024-10576-7","url":null,"abstract":"<p>The first observation of the paper is that methods for determining proportional representation in electoral systems may be suitable as alternatives to the pro-rata order matching algorithm used in stock exchanges. The main part of our work is to comprehensively consider various well known proportional representation methods and analyse in details their suitability for replacing the pro-rata algorithm. Our analysis consists of a theoretical study as well as simulation studies based on data sampled from a distribution which has been suggested in the literature as models of limit orders. Based on our analysis, we put forward the suggestion that the well known Hamilton’s method is a superior alternative to the pro-rata algorithm for order matching applications.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"22 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140202044","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-19DOI: 10.1007/s10614-024-10583-8
Priya Singh, Manoj Jha
Integration of asset preselection with appropriate portfolio optimization techniques can improve the performance of the portfolio optimization models. This paper morphed the potential asset selection and the optimal portfolio construction rather than focusing on one. A large volume of sample data from 25 stocks is used for the experiment from the National Stock Exchange, India, between January 2005 and December 2021. Initially, a 3-step screening approach, an asset selection method is applied to select potential assets. The 3-steps comprise data choice, fundamental screening, and the Long Short Term Memory model anticipating real-time stock prices to shortlist stocks with higher expected returns. The suggested approach is effective in determining the quality of assets. Further, the optimal asset allocation is done by introducing a novel exponentially weighted-mean-variance model. This exponential weighting scheme outperforms the classical Mean-Variance model when applied to the maximum Sharpe ratio model. The proposed model outperforms the five baseline techniques in terms of the Sharpe ratio and average potential returns and risks. Additionally, the proposed model’s resilience across diversified time frames is tested through the incorporation of multiple time windows, demonstrating robustness of the performance.
{"title":"Portfolio Optimization Using Novel EW-MV Method in Conjunction with Asset Preselection","authors":"Priya Singh, Manoj Jha","doi":"10.1007/s10614-024-10583-8","DOIUrl":"https://doi.org/10.1007/s10614-024-10583-8","url":null,"abstract":"<p>Integration of asset preselection with appropriate portfolio optimization techniques can improve the performance of the portfolio optimization models. This paper morphed the potential asset selection and the optimal portfolio construction rather than focusing on one. A large volume of sample data from 25 stocks is used for the experiment from the National Stock Exchange, India, between January 2005 and December 2021. Initially, a 3-step screening approach, an asset selection method is applied to select potential assets. The 3-steps comprise data choice, fundamental screening, and the Long Short Term Memory model anticipating real-time stock prices to shortlist stocks with higher expected returns. The suggested approach is effective in determining the quality of assets. Further, the optimal asset allocation is done by introducing a novel exponentially weighted-mean-variance model. This exponential weighting scheme outperforms the classical Mean-Variance model when applied to the maximum Sharpe ratio model. The proposed model outperforms the five baseline techniques in terms of the Sharpe ratio and average potential returns and risks. Additionally, the proposed model’s resilience across diversified time frames is tested through the incorporation of multiple time windows, demonstrating robustness of the performance.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"52 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140170219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-17DOI: 10.1007/s10614-024-10570-z
Julien Albertini, Stéphane Moyen
This paper provides a general representation of endogenous and threshold-based regime-switching models while also developing an efficient numerical solution method. Regime-switching occurs endogenously when certain variables cross threshold conditions that can themselves be regime-dependent. We illustrate our approach using a RBC model with state-dependent government spending policies. It is shown that regime-switching models involve strong non linearities and discontinuities in the dynamics of the model. However, our numerical solution which relies on simulation and projection methods with regime-dependent policy rules, proves to be accurate and sufficiently fast address these challenging aspects. Several also explore several alternative specifications for the model and the method.
{"title":"A General and Efficient Method for Solving Regime-Switching DSGE Models","authors":"Julien Albertini, Stéphane Moyen","doi":"10.1007/s10614-024-10570-z","DOIUrl":"https://doi.org/10.1007/s10614-024-10570-z","url":null,"abstract":"<p>This paper provides a general representation of endogenous and threshold-based regime-switching models while also developing an efficient numerical solution method. Regime-switching occurs endogenously when certain variables cross threshold conditions that can themselves be regime-dependent. We illustrate our approach using a RBC model with state-dependent government spending policies. It is shown that regime-switching models involve strong non linearities and discontinuities in the dynamics of the model. However, our numerical solution which relies on simulation and projection methods with regime-dependent policy rules, proves to be accurate and sufficiently fast address these challenging aspects. Several also explore several alternative specifications for the model and the method.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"163 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140170467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-16DOI: 10.1007/s10614-024-10571-y
Jie Cheng
Scoring rules are commonly applied to assess the accuracy of density forecasts in both univariate and multivariate settings. In a financial risk management context, we are mostly interested in a particular region of the density: the (left) tail of a portfolio’s return distribution. The dependence structure between returns on different assets (associated with a given portfolio) is usually time-varying and asymmetric. In this paper, we conduct a simulation study to compare the discrimination ability between the well-established scores and their threshold-weighted versions with selected regions. This facilitates a comprehensive comparison of the performance of scoring rules in different settings. Our empirical applications also confirm the importance of weighted-threshold scores for accurate estimates of Value-at-risk and related measures of downside risk.
{"title":"Evaluating Density Forecasts Using Weighted Multivariate Scores in a Risk Management Context","authors":"Jie Cheng","doi":"10.1007/s10614-024-10571-y","DOIUrl":"https://doi.org/10.1007/s10614-024-10571-y","url":null,"abstract":"<p>Scoring rules are commonly applied to assess the accuracy of density forecasts in both univariate and multivariate settings. In a financial risk management context, we are mostly interested in a particular region of the density: the (left) tail of a portfolio’s return distribution. The dependence structure between returns on different assets (associated with a given portfolio) is usually time-varying and asymmetric. In this paper, we conduct a simulation study to compare the discrimination ability between the well-established scores and their threshold-weighted versions with selected regions. This facilitates a comprehensive comparison of the performance of scoring rules in different settings. Our empirical applications also confirm the importance of weighted-threshold scores for accurate estimates of Value-at-risk and related measures of downside risk.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"1 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140153506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-16DOI: 10.1007/s10614-024-10561-0
Luca Gori, Francesco Purificato, Mauro Sodini
The main aim of the present research is to consider a monetary union’s economy consisting of N countries, N fiscal authorities (one for each country) and a single monetary authority. The fiscal authorities want to stabilise output and public debt through the primary government balance, and they can exhibit heterogeneous preferences about the trade-off between output and debt stability. Unlike these, the monetary authority has the aim of price and output stability. They play a non-cooperative policy game, in which they independently and simultaneously choose monetary and fiscal instruments to pursue their goals. In a dynamic setting, each authority must choose its policy instrument prevailing in the next period without knowing—at the end of each period—the choice of other authorities. By assuming static expectations, the present work shows the possibility of several dynamic outcomes. First, there exists one Nash equilibrium representing the optimal level for the macro economy; this equilibrium is stable if the average weight that fiscal authorities assign to output stability is not excessively high; therefore, this result holds even if some authorities are less willing to promote debt stabilisation. Second, in addition to this equilibrium, there exist other Nash equilibria representing steady-state values for macroeconomic variables that differ from the targets adopted by the authorities; these equilibria emerge and are stable if the authorities’ preference for output stability is even greater and with a higher degree of heterogeneity compared to the previous case. Third, the parameters of the model matter to determine the stability properties of the equilibria, and the analysis shows the possibility of nonlinear dynamics.
本研究的主要目的是考虑一个由 N 个国家、N 个财政当局(每个国家一个)和一个货币当局组成的货币联盟经济。财政当局希望通过政府收支基本平衡来稳定产出和公共债务,他们可以在产出和债务稳定之间表现出不同的偏好。与之不同的是,货币当局的目标是稳定价格和产出。他们进行的是一种非合作性政策博弈,在这种博弈中,他们独立地同时选择货币和财政工具来实现自己的目标。在动态环境中,每个当局都必须在不知道其他当局的选择的情况下选择下一期的政策工具。通过假设静态预期,本研究显示了几种动态结果的可能性。首先,存在一个代表宏观经济最优水平的纳什均衡;如果财政当局对产出稳定性的平均权重不是过高,这个均衡就是稳定的;因此,即使有些当局不太愿意促进债务稳定,这个结果也是成立的。其次,除了这一均衡外,还存在其他纳什均衡,代表宏观经济变量的稳态值与当局采用的目标不同;如果当局对产出稳定的偏好比前一种情况更大且异质性更高,这些均衡就会出现并保持稳定。第三,模型参数对确定均衡的稳定性非常重要,分析表明了非线性动态的可能性。
{"title":"Debt Stabilisation and Dynamic Interaction Between Monetary Authority and National Fiscal Authorities","authors":"Luca Gori, Francesco Purificato, Mauro Sodini","doi":"10.1007/s10614-024-10561-0","DOIUrl":"https://doi.org/10.1007/s10614-024-10561-0","url":null,"abstract":"<p>The main aim of the present research is to consider a monetary union’s economy consisting of <i>N</i> countries, <i>N</i> fiscal authorities (one for each country) and a single monetary authority. The fiscal authorities want to stabilise output and public debt through the primary government balance, and they can exhibit heterogeneous preferences about the trade-off between output and debt stability. Unlike these, the monetary authority has the aim of price and output stability. They play a non-cooperative policy game, in which they independently and simultaneously choose monetary and fiscal instruments to pursue their goals. In a dynamic setting, each authority must choose its policy instrument prevailing in the next period without knowing—at the end of each period—the choice of other authorities. By assuming static expectations, the present work shows the possibility of several dynamic outcomes. First, there exists one Nash equilibrium representing the optimal level for the macro economy; this equilibrium is stable if the average weight that fiscal authorities assign to output stability is not excessively high; therefore, this result holds even if some authorities are less willing to promote debt stabilisation. Second, in addition to this equilibrium, there exist other Nash equilibria representing steady-state values for macroeconomic variables that differ from the targets adopted by the authorities; these equilibria emerge and are stable if the authorities’ preference for output stability is even greater and with a higher degree of heterogeneity compared to the previous case. Third, the parameters of the model matter to determine the stability properties of the equilibria, and the analysis shows the possibility of nonlinear dynamics.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"98 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140153603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-15DOI: 10.1007/s10614-024-10580-x
Abstract
Composite distributions where volatility itself is assumed to be a random variable have been used to model stock returns. In this paper, we give details of estimation of these composite distributions when the volatility is assumed to follow an arbitrary distribution and the conditional distribution of stock returns given the volatility follows one of normal, Laplace, uniform, Student’s t, Cauchy, logistic of type I, logistic of type II, logistic of type III, logistic of type IV, generalized normal or skew normal distributions. The details given include estimating equations and observed information matrices. An application to Bitcoin exchange rate data is illustrated. Models taking volatility to follow gamma and Weibull distributions are shown to provide excellent fits.
{"title":"Estimation of Models for Stock Returns","authors":"","doi":"10.1007/s10614-024-10580-x","DOIUrl":"https://doi.org/10.1007/s10614-024-10580-x","url":null,"abstract":"<h3>Abstract</h3> <p>Composite distributions where volatility itself is assumed to be a random variable have been used to model stock returns. In this paper, we give details of estimation of these composite distributions when the volatility is assumed to follow an arbitrary distribution and the conditional distribution of stock returns given the volatility follows one of normal, Laplace, uniform, Student’s <em>t</em>, Cauchy, logistic of type I, logistic of type II, logistic of type III, logistic of type IV, generalized normal or skew normal distributions. The details given include estimating equations and observed information matrices. An application to Bitcoin exchange rate data is illustrated. Models taking volatility to follow gamma and Weibull distributions are shown to provide excellent fits.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"4 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140153518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-15DOI: 10.1007/s10614-024-10574-9
Tingting Sun, Haoyuan Wang, Donglin Wang
Valuation of large portfolios of variable annuities (VAs) is a well-researched area in the actuarial science field. However, the study of producing reliable prediction intervals for prices has received comparatively less attention. Compared to point prediction, the prediction interval can calculate a reasonable price range of VAs and help investors and insurance companies better manage risk to maintain profitability and sustainability. In this study, we address this gap by utilizing five different models in conjunction with bootstrapping techniques to generate robust prediction intervals for variable annuity prices. Our findings show that the Gradient Boosting regression (GBR) model provides the narrowest intervals compared to the other four models. While the Random sample consensus (RANSAC) model has the highest coverage rate, but it has the widest interval. In practical applications, considering the trade-off between coverage rate and interval width, the GBR model would be a preferred choice. Therefore, we recommend using the gradient boosting model with the bootstrap method to calculate the prediction interval of valuation for a large portfolio of variable annuity policies.
{"title":"Robust Prediction Intervals for Valuation of Large Portfolios of Variable Annuities: A Comparative Study of Five Models","authors":"Tingting Sun, Haoyuan Wang, Donglin Wang","doi":"10.1007/s10614-024-10574-9","DOIUrl":"https://doi.org/10.1007/s10614-024-10574-9","url":null,"abstract":"<p>Valuation of large portfolios of variable annuities (VAs) is a well-researched area in the actuarial science field. However, the study of producing reliable prediction intervals for prices has received comparatively less attention. Compared to point prediction, the prediction interval can calculate a reasonable price range of VAs and help investors and insurance companies better manage risk to maintain profitability and sustainability. In this study, we address this gap by utilizing five different models in conjunction with bootstrapping techniques to generate robust prediction intervals for variable annuity prices. Our findings show that the Gradient Boosting regression (GBR) model provides the narrowest intervals compared to the other four models. While the Random sample consensus (RANSAC) model has the highest coverage rate, but it has the widest interval. In practical applications, considering the trade-off between coverage rate and interval width, the GBR model would be a preferred choice. Therefore, we recommend using the gradient boosting model with the bootstrap method to calculate the prediction interval of valuation for a large portfolio of variable annuity policies.</p>","PeriodicalId":50647,"journal":{"name":"Computational Economics","volume":"69 1","pages":""},"PeriodicalIF":2.0,"publicationDate":"2024-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140153634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}