首页 > 最新文献

CompSciRN: Other Machine Learning (Topic)最新文献

英文 中文
Deep MVA: Deep Learning for Margin Valuation Adjustment of Callable Products 深度MVA:可赎回产品保证金估值调整的深度学习
Pub Date : 2020-06-22 DOI: 10.2139/ssrn.3634059
Anup Aryan, Allan Cowan
Regulatory initial margin (IM) is being implemented across the financial industry in accordance with BCBS-IOSCO requirements. The regulations target uncleared over-the-counter derivative trading and, among other issues, aim to mitigate counter party credit risk by defining a comprehensive set of rules for initial margin between trading parties. Computing the funding costs of IM, or Margin Valuation Adjustment (MVA), is a major challenge for xVA systems as it requires the future projection of dynamic IM positions. This is particularly challenging for callable products, such as Bermudan swaptions which are complex to price and require path wise exercise tracking in exposure simulations. Brute force simulation of future IM is not feasible due to the excessive computational demands of model calibration and numerical pricing methods. Approximate MVA methods, such as regression techniques, are difficult to design due to the high-dimensionality of the problem. In this paper, we propose a method based on Deep Neural Networks to approximate the Bermudan swaption pricing function and sensitivities. We exploit neural network's high-dimensionality and universal approximation properties to train networks based on prices and sensitivities generated from existing numerical pricing models. The trained neural networks are then used for extremely fast IM simulation where computationally intense numerical methods are replaced by optimized and hardware accelerated neural network inference. We demonstrate that the neural network models deliver exceptional performance, capable of pricing Bermudan Swaption MVAs over 100,000 times faster than traditional approaches while maintaining a high degree of accuracy.
根据BCBS-IOSCO的要求,监管初始保证金(IM)正在整个金融行业实施。这些规定针对的是未清算的场外衍生品交易,除其他问题外,还旨在通过为交易各方之间的初始保证金制定一套全面的规则,降低交易对手的信用风险。计算IM的资金成本,或保证金估值调整(MVA),是xVA系统的主要挑战,因为它需要动态IM头寸的未来预测。这对于可调用的产品尤其具有挑战性,例如百慕大掉期交易,这类产品定价复杂,并且需要在风险模拟中进行路径明智的运动跟踪。由于模型校准和数值定价方法的计算量过大,未来IM的蛮力模拟是不可行的。近似MVA方法,如回归技术,由于问题的高维性而难以设计。本文提出了一种基于深度神经网络的百慕大互换定价函数和灵敏度近似方法。我们利用神经网络的高维性和通用近似特性来训练基于现有数值定价模型生成的价格和敏感性的网络。然后将训练好的神经网络用于极快的IM仿真,其中计算强度高的数值方法被优化和硬件加速的神经网络推理所取代。我们证明,神经网络模型提供了卓越的性能,能够比传统方法快10万倍以上定价百慕大Swaption的MVAs,同时保持高度的准确性。
{"title":"Deep MVA: Deep Learning for Margin Valuation Adjustment of Callable Products","authors":"Anup Aryan, Allan Cowan","doi":"10.2139/ssrn.3634059","DOIUrl":"https://doi.org/10.2139/ssrn.3634059","url":null,"abstract":"Regulatory initial margin (IM) is being implemented across the financial industry in accordance with BCBS-IOSCO requirements. The regulations target uncleared over-the-counter derivative trading and, among other issues, aim to mitigate counter party credit risk by defining a comprehensive set of rules for initial margin between trading parties. Computing the funding costs of IM, or Margin Valuation Adjustment (MVA), is a major challenge for xVA systems as it requires the future projection of dynamic IM positions. This is particularly challenging for callable products, such as Bermudan swaptions which are complex to price and require path wise exercise tracking in exposure simulations. Brute force simulation of future IM is not feasible due to the excessive computational demands of model calibration and numerical pricing methods. Approximate MVA methods, such as regression techniques, are difficult to design due to the high-dimensionality of the problem. In this paper, we propose a method based on Deep Neural Networks to approximate the Bermudan swaption pricing function and sensitivities. We exploit neural network's high-dimensionality and universal approximation properties to train networks based on prices and sensitivities generated from existing numerical pricing models. The trained neural networks are then used for extremely fast IM simulation where computationally intense numerical methods are replaced by optimized and hardware accelerated neural network inference. We demonstrate that the neural network models deliver exceptional performance, capable of pricing Bermudan Swaption MVAs over 100,000 times faster than traditional approaches while maintaining a high degree of accuracy.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116438524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Decision-Driven Regularization: Harmonizing the Predictive and Prescriptive 决策驱动的规范化:协调预测性和规定性
Pub Date : 2020-06-09 DOI: 10.2139/ssrn.3623006
G. Loke, Qinshen Tang, Yangge Xiao
Joint prediction and optimization problems are common in many business applications ranging from customer relationship management and marketing to revenue and retail operations management. These problems involve a first-stage learning model, where outcomes are predicted from features, and a second-stage decision process, which selects the optimal decisions based on these outcomes. In practice, these two stages are conducted separately, but is sub-optimal. In this work, we propose a novel model that solves both parts as a whole, but is computationally tractable under many circumstances. Specifically, we introduce the notion of a regularizer that measures the value of a predictive model in terms of the cost incurred in the decision process. We term this decision-driven regularization, and it is centred on the premise that the bias-variance trade-off in the learning problem is not transformed linearly by the subsequent decision problem. Additionally, this accounts for the ambiguity in the definition of the cost function, which we identify. We prove key properties of our model, namely, that it is consistent, robust to wrong estimation, and has bounded bias. We also examine special cases under which we draw links to existing models in the literature, propose hybrid models and are able to describe their effectiveness using our framework as a theoretical basis. In our numerical experiments, we illustrate the behaviour of our model, and its performance against other models in the literature.
联合预测和优化问题在许多业务应用程序中很常见,从客户关系管理和市场营销到收入和零售运营管理。这些问题涉及第一阶段的学习模型,从特征中预测结果,以及第二阶段的决策过程,根据这些结果选择最佳决策。实际上,这两个阶段是分开进行的,但不是最优的。在这项工作中,我们提出了一个新的模型,将这两个部分作为一个整体来解决,但在许多情况下计算上是可处理的。具体来说,我们引入了一个正则化器的概念,它根据决策过程中产生的成本来衡量预测模型的价值。我们称之为决策驱动的正则化,它集中在学习问题中的偏差-方差权衡不被后续决策问题线性转换的前提下。此外,这也解释了我们所识别的成本函数定义的模糊性。我们证明了模型的关键性质,即它是一致的,对错误估计具有鲁棒性,并且具有有界偏差。我们还研究了一些特殊情况,在这些情况下,我们与文献中的现有模型建立了联系,提出了混合模型,并能够使用我们的框架作为理论基础来描述它们的有效性。在我们的数值实验中,我们说明了我们的模型的行为,以及它与文献中其他模型的表现。
{"title":"Decision-Driven Regularization: Harmonizing the Predictive and Prescriptive","authors":"G. Loke, Qinshen Tang, Yangge Xiao","doi":"10.2139/ssrn.3623006","DOIUrl":"https://doi.org/10.2139/ssrn.3623006","url":null,"abstract":"Joint prediction and optimization problems are common in many business applications ranging from customer relationship management and marketing to revenue and retail operations management. These problems involve a first-stage learning model, where outcomes are predicted from features, and a second-stage decision process, which selects the optimal decisions based on these outcomes. In practice, these two stages are conducted separately, but is sub-optimal. In this work, we propose a novel model that solves both parts as a whole, but is computationally tractable under many circumstances. Specifically, we introduce the notion of a regularizer that measures the value of a predictive model in terms of the cost incurred in the decision process. We term this decision-driven regularization, and it is centred on the premise that the bias-variance trade-off in the learning problem is not transformed linearly by the subsequent decision problem. Additionally, this accounts for the ambiguity in the definition of the cost function, which we identify. We prove key properties of our model, namely, that it is consistent, robust to wrong estimation, and has bounded bias. We also examine special cases under which we draw links to existing models in the literature, propose hybrid models and are able to describe their effectiveness using our framework as a theoretical basis. In our numerical experiments, we illustrate the behaviour of our model, and its performance against other models in the literature.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122099030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
DataGene: A Framework for Dataset Similarity DataGene:数据集相似度的框架
Pub Date : 2020-06-05 DOI: 10.2139/ssrn.3619626
Derek Snow
DataGene is developed to identify data set similarity between real and synthetic datasets as well as train, test, and validation datasets. For many modelling and software development tasks there is a need for datasets to have share similar characteristics. This has traditionally been achieved with visualizations, DataGene seeks to replace these visual methods with a range of novel quantitative methods. Please see the GitHub repository to inspect and install the Python code.
开发DataGene是为了识别真实数据集和合成数据集以及训练、测试和验证数据集之间的数据集相似性。对于许多建模和软件开发任务,需要数据集具有共享的相似特征。这在传统上是通过可视化实现的,DataGene试图用一系列新颖的定量方法取代这些可视化方法。请参阅GitHub存储库来检查和安装Python代码。
{"title":"DataGene: A Framework for Dataset Similarity","authors":"Derek Snow","doi":"10.2139/ssrn.3619626","DOIUrl":"https://doi.org/10.2139/ssrn.3619626","url":null,"abstract":"DataGene is developed to identify data set similarity between real and synthetic datasets as well as train, test, and validation datasets. For many modelling and software development tasks there is a need for datasets to have share similar characteristics. This has traditionally been achieved with visualizations, DataGene seeks to replace these visual methods with a range of novel quantitative methods. Please see the GitHub repository to inspect and install the Python code.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116811292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of Big Data in Smart Agriculture 大数据在智慧农业中的应用
Pub Date : 2020-05-27 DOI: 10.2139/ssrn.3611514
B. Rani, M. Kumari, Kumari Sobha ,, Pinki Kumari, Jyotipragyan Majhi, Subham Chakraborty
Big data is such a field where we collect and store the related data from agriculture. In this paper we researched about advanced technology of agriculture with big data, Smart Farming/Organic Farming which is lacking in India. There are many farmers in India who are deprived of advanced technology.In which we want to provide true knowledge to the farmers through soil, irrigation, environment, pesticides and genetic engineering so that the economy of the farmer can be improved so that more and more people are connected with agriculture.
大数据就是这样一个领域,我们从农业中收集和存储相关数据。在本文中,我们研究了印度缺乏的大数据先进农业技术,智能农业/有机农业。印度有许多农民被剥夺了先进技术。我们希望通过土壤、灌溉、环境、农药和基因工程向农民提供真正的知识,从而提高农民的经济水平,让越来越多的人与农业联系起来。
{"title":"Application of Big Data in Smart Agriculture","authors":"B. Rani, M. Kumari, Kumari Sobha ,, Pinki Kumari, Jyotipragyan Majhi, Subham Chakraborty","doi":"10.2139/ssrn.3611514","DOIUrl":"https://doi.org/10.2139/ssrn.3611514","url":null,"abstract":"Big data is such a field where we collect and store the related data from agriculture. In this paper we researched about advanced technology of agriculture with big data, Smart Farming/Organic Farming which is lacking in India. There are many farmers in India who are deprived of advanced technology.In which we want to provide true knowledge to the farmers through soil, irrigation, environment, pesticides and genetic engineering so that the economy of the farmer can be improved so that more and more people are connected with agriculture.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115518868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Data Science: The Impact of Machine Learning 数据科学:机器学习的影响
Pub Date : 2020-05-25 DOI: 10.2139/ssrn.3674357
S. Islam
In this paper, I prove my promise that Machine Learning is one of the most important parts to provide tools and methods to go deeper and nurture the data properly. The most amazing part is to analyze the large chunks of data in a very precise way, and high-value predictions that can guide better decisions and smart actions in real-time without human intervention. I give an overview over different proposed structures of Data Science and mention the impact of Machine learning such as algorithms, model evaluation and selection, pipeline. I also indicate all misconceptions when neglecting Machine learning reasoning.
在本文中,我证明了我的承诺,即机器学习是提供工具和方法以更深入地培养数据的最重要部分之一。最神奇的是,它能够以非常精确的方式分析大量数据,并做出高价值的预测,从而在没有人为干预的情况下,实时指导更好的决策和智能行动。我概述了数据科学的不同结构,并提到了机器学习的影响,如算法、模型评估和选择、管道。我还指出了忽略机器学习推理时的所有误解。
{"title":"Data Science: The Impact of Machine Learning","authors":"S. Islam","doi":"10.2139/ssrn.3674357","DOIUrl":"https://doi.org/10.2139/ssrn.3674357","url":null,"abstract":"In this paper, I prove my promise that Machine Learning is one of the most important parts to provide tools and methods to go deeper and nurture the data properly. The most amazing part is to analyze the large chunks of data in a very precise way, and high-value predictions that can guide better decisions and smart actions in real-time without human intervention. I give an overview over different proposed structures of Data Science and mention the impact of Machine learning such as algorithms, model evaluation and selection, pipeline. I also indicate all misconceptions when neglecting Machine learning reasoning.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131891432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gazing into the Future: Using Ensemble Techniques to Forecast Company Fundamentals 展望未来:使用集合技术预测公司基本面
Pub Date : 2020-04-19 DOI: 10.2139/ssrn.3580018
Steven Downey
Quantitative factor portfolios generally use historical company fundamental data in portfolio construction. What if we could forecast, with a small margin of error, the forward-looking company fundamentals? Using best practices from the science of forecasting and machine learning techniques, namely Random Forests and Gradient Boosting, I try to build a value composite model to sort portfolios based on forecasted fundamentals. I use the in-sample data to train the models to predict forward looking earnings, free cash flow, EBITDA, and Net Operating Profit After Taxes. The combined value portfolio out of sample did not produce statistically significant outperformance verses the equal weight portfolio (as a comparison to the long only value composite) or versus cash (for the long/short portfolio).
定量因子投资组合通常在投资组合构建中使用公司的历史基本面数据。如果我们能在很小的误差范围内预测出前瞻性的公司基本面,那会怎么样?利用预测科学和机器学习技术的最佳实践,即随机森林和梯度增强,我试图建立一个价值组合模型,根据预测的基本面对投资组合进行排序。我使用样本内数据来训练模型来预测前瞻性收益、自由现金流、EBITDA和税后净营业利润。样本外的组合价值投资组合与同等权重的投资组合(与只做多的价值组合相比)或与现金(多/空投资组合)相比,在统计上没有显著的表现。
{"title":"Gazing into the Future: Using Ensemble Techniques to Forecast Company Fundamentals","authors":"Steven Downey","doi":"10.2139/ssrn.3580018","DOIUrl":"https://doi.org/10.2139/ssrn.3580018","url":null,"abstract":"Quantitative factor portfolios generally use historical company fundamental data in portfolio construction. What if we could forecast, with a small margin of error, the forward-looking company fundamentals? Using best practices from the science of forecasting and machine learning techniques, namely Random Forests and Gradient Boosting, I try to build a value composite model to sort portfolios based on forecasted fundamentals. I use the in-sample data to train the models to predict forward looking earnings, free cash flow, EBITDA, and Net Operating Profit After Taxes. The combined value portfolio out of sample did not produce statistically significant outperformance verses the equal weight portfolio (as a comparison to the long only value composite) or versus cash (for the long/short portfolio).","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130404064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine Learning Portfolio Allocation 机器学习投资组合分配
Pub Date : 2020-03-02 DOI: 10.2139/ssrn.3546294
Michael Pinelis, D. Ruppert
We find economically and statistically significant gains when using machine learning for portfolio allocation between the market index and risk-free asset. Optimal portfolio rules for time-varying expected returns and volatility are implemented with two Random Forest models. One model is employed in forecasting the sign probabilities of the excess return with payout yields. The second is used to construct an optimized volatility estimate. Reward-risk timing with machine learning provides substantial improvements over the buy-and-hold in utility, risk-adjusted returns, and maximum drawdowns. This paper presents a new theoretical basis and unifying framework for machine learning applied to both return- and volatility-timing.
当使用机器学习在市场指数和无风险资产之间进行投资组合分配时,我们发现在经济上和统计上都有显著的收益。用两个随机森林模型实现了时变预期收益和波动率的最优投资组合规则。一个模型用于预测超额收益与支付收益率的符号概率。第二部分用于构造一个优化的波动率估计。与效用、风险调整回报和最大回收量相比,机器学习的风险奖励时机提供了实质性的改进。本文提出了一个新的理论基础和统一框架的机器学习应用于回报和波动时序。
{"title":"Machine Learning Portfolio Allocation","authors":"Michael Pinelis, D. Ruppert","doi":"10.2139/ssrn.3546294","DOIUrl":"https://doi.org/10.2139/ssrn.3546294","url":null,"abstract":"We find economically and statistically significant gains when using machine learning for portfolio allocation between the market index and risk-free asset. Optimal portfolio rules for time-varying expected returns and volatility are implemented with two Random Forest models. One model is employed in forecasting the sign probabilities of the excess return with payout yields. The second is used to construct an optimized volatility estimate. Reward-risk timing with machine learning provides substantial improvements over the buy-and-hold in utility, risk-adjusted returns, and maximum drawdowns. This paper presents a new theoretical basis and unifying framework for machine learning applied to both return- and volatility-timing.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"80 9","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132579839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Data Anonymisation, Outlier Detection and Fighting Overfitting with Restricted Boltzmann Machines 基于受限玻尔兹曼机的数据匿名化、离群值检测与抗过拟合
Pub Date : 2020-01-27 DOI: 10.2139/ssrn.3526436
A. Kondratyev, Christian Schwarz, Blanka Horvath
We propose a novel approach to the anonymisation of datasets through non-parametric learning of the underlying multivariate distribution of dataset features and generation of the new synthetic samples from the learned distribution. The main objective is to ensure equal (or better) performance of the classifiers and regressors trained on synthetic datasets in comparison with the same classifiers and regressors trained on the original data. The ability to generate unlimited number of synthetic data samples from the learned distribution can be a remedy in fighting overtting when dealing with small original datasets. When the synthetic data generator is trained as an autoencoder with the bottleneck information compression structure we can also expect to see a reduced number of outliers in the generated datasets, thus further improving the generalization capabilities of the classifiers trained on synthetic data. We achieve these objectives with the help of the Restricted Boltzmann Machine, a special type of generative neural network that possesses all the required properties of a powerful data anonymiser.
我们提出了一种新的数据集匿名方法,通过对数据集特征的潜在多元分布进行非参数学习,并从学习分布中生成新的合成样本。主要目标是确保在合成数据集上训练的分类器和回归器与在原始数据上训练的相同分类器和回归器相比具有相同(或更好)的性能。在处理小型原始数据集时,从学习分布生成无限数量的合成数据样本的能力可以作为对抗覆盖的补救措施。当合成数据生成器被训练成具有瓶颈信息压缩结构的自编码器时,我们还可以期望看到生成的数据集中的异常值数量减少,从而进一步提高在合成数据上训练的分类器的泛化能力。我们在受限玻尔兹曼机的帮助下实现了这些目标,受限玻尔兹曼机是一种特殊类型的生成神经网络,它拥有强大的数据匿名器所需的所有属性。
{"title":"Data Anonymisation, Outlier Detection and Fighting Overfitting with Restricted Boltzmann Machines","authors":"A. Kondratyev, Christian Schwarz, Blanka Horvath","doi":"10.2139/ssrn.3526436","DOIUrl":"https://doi.org/10.2139/ssrn.3526436","url":null,"abstract":"We propose a novel approach to the anonymisation of datasets through non-parametric learning of the underlying multivariate distribution of dataset features and generation of the new synthetic samples from the learned distribution. The main objective is to ensure equal (or better) performance of the classifiers and regressors trained on synthetic datasets in comparison with the same classifiers and regressors trained on the original data. The ability to generate unlimited number of synthetic data samples from the learned distribution can be a remedy in fighting overtting when dealing with small original datasets. When the synthetic data generator is trained as an autoencoder with the bottleneck information compression structure we can also expect to see a reduced number of outliers in the generated datasets, thus further improving the generalization capabilities of the classifiers trained on synthetic data. We achieve these objectives with the help of the Restricted Boltzmann Machine, a special type of generative neural network that possesses all the required properties of a powerful data anonymiser.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131812088","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Differential Machine Learning 微分机器学习
Pub Date : 2020-01-03 DOI: 10.2139/ssrn.3591734
Antoine Savine, B. Huge
Differential machine learning (ML) extends supervised learning, with models trained on examples of not only inputs and labels, but also differentials of labels to inputs. Differential ML is applicable in all situations where high quality first order derivatives wrt training inputs are available. In the context of financial Derivatives risk management, pathwise differentials are efficiently computed with automatic adjoint differentiation (AAD). Differential ML, combined with AAD, provides extremely effective pricing and risk approximations. We can produce fast pricing analytics in models too complex for closed form solutions, extract the risk factors of complex transactions and trading books, and effectively compute risk management metrics like reports across a large number of scenarios, backtesting and simulation of hedge strategies, or capital regulations. The article focuses on differential deep learning (DL), arguably the strongest application. Standard DL trains neural networks (NN) on punctual examples, whereas differential DL teaches them the shape of the target function, resulting in vastly improved performance, illustrated with a number of numerical examples, both idealized and real world. In the online appendices, we apply differential learning to other ML models, like classic regression or principal component analysis (PCA), with equally remarkable results. This paper is meant to be read in conjunction with its companion GitHub repo https://github.com/differential-machine-learning, where we posted a TensorFlow implementation, tested on Google Colab, along with examples from the article and additional ones. We also posted appendices covering many practical implementation details not covered in the paper, mathematical proofs, application to ML models besides neural networks and extensions necessary for a reliable implementation in production.
微分机器学习(ML)扩展了监督学习,模型不仅在输入和标签的例子上训练,而且在标签到输入的微分上训练。微分ML适用于所有高质量的一阶导数训练输入可用的情况。在金融衍生品风险管理的背景下,利用自动伴随微分(AAD)有效地计算路径差分。差分ML与AAD相结合,提供了极其有效的定价和风险近似。我们可以在过于复杂的模型中生成快速定价分析,对于封闭形式的解决方案来说,提取复杂交易和交易账簿的风险因素,并有效地计算风险管理指标,如跨大量场景的报告,对冲策略的回测和模拟,或资本监管。本文主要关注差分深度学习(DL),这可以说是最强大的应用。标准深度学习在准时的例子上训练神经网络(NN),而微分深度学习则教会它们目标函数的形状,从而大大提高了性能,用许多理想化和现实世界的数值例子来说明。在在线附录中,我们将差分学习应用于其他ML模型,如经典回归或主成分分析(PCA),结果同样显著。这篇文章是为了与它的伙伴GitHub repo https://github.com/differential-machine-learning一起阅读,我们在那里发布了一个TensorFlow实现,在Google Colab上进行了测试,以及文章中的示例和其他示例。我们还发布了附录,涵盖了论文中未涉及的许多实际实现细节,数学证明,除了神经网络和在生产中可靠实现所需的扩展之外,还包括ML模型的应用。
{"title":"Differential Machine Learning","authors":"Antoine Savine, B. Huge","doi":"10.2139/ssrn.3591734","DOIUrl":"https://doi.org/10.2139/ssrn.3591734","url":null,"abstract":"Differential machine learning (ML) extends supervised learning, with models trained on examples of not only inputs and labels, but also differentials of labels to inputs. Differential ML is applicable in all situations where high quality first order derivatives wrt training inputs are available. In the context of financial Derivatives risk management, pathwise differentials are efficiently computed with automatic adjoint differentiation (AAD). Differential ML, combined with AAD, provides extremely effective pricing and risk approximations. We can produce fast pricing analytics in models too complex for closed form solutions, extract the risk factors of complex transactions and trading books, and effectively compute risk management metrics like reports across a large number of scenarios, backtesting and simulation of hedge strategies, or capital regulations. The article focuses on differential deep learning (DL), arguably the strongest application. Standard DL trains neural networks (NN) on punctual examples, whereas differential DL teaches them the shape of the target function, resulting in vastly improved performance, illustrated with a number of numerical examples, both idealized and real world. In the online appendices, we apply differential learning to other ML models, like classic regression or principal component analysis (PCA), with equally remarkable results. This paper is meant to be read in conjunction with its companion GitHub repo https://github.com/differential-machine-learning, where we posted a TensorFlow implementation, tested on Google Colab, along with examples from the article and additional ones. We also posted appendices covering many practical implementation details not covered in the paper, mathematical proofs, application to ML models besides neural networks and extensions necessary for a reliable implementation in production.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126581989","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Objective-Aligned Regression for Two-Stage Linear Programs 两阶段线性规划的目标对准回归
Pub Date : 2019-10-14 DOI: 10.2139/ssrn.3469897
Alexander S. Estes, Jean-Philippe P. Richard
We study an approach to regression that we call objective-aligned fitting, which is applicable when the regression model is used to predict uncertain parameters of some objective problem. Rather than minimizing a typical loss function, such as squared error, we approximately minimize the objective value of the resulting solutions to the nominal optimization problem. While previous work on objective-aligned fitting has tended to focus on uncertainty in the objective function, we consider the case in which the nominal optimization problem is a two-stage linear program with uncertainty in the right-hand side. We define the objective-aligned loss function for the problem and prove structural properties concerning this loss function. Since the objective-aligned loss function is generally non-convex, we develop a convex approximation. We propose a method for fitting a linear regression model to the convex approximation of the objective-aligned loss. Computational results indicate that this procedure can lead to higher-quality solutions than existing regression procedures.
本文研究了一种称为目标拟合的回归方法,它适用于回归模型对某些客观问题的不确定参数的预测。而不是最小化典型的损失函数,如平方误差,我们近似地最小化标称优化问题的结果解的目标值。虽然之前关于目标对齐拟合的工作倾向于关注目标函数的不确定性,但我们考虑的情况是,标称优化问题是一个两阶段线性规划,右侧不确定性。我们定义了目标对准的损失函数,并证明了该损失函数的结构性质。由于目标对准的损失函数通常是非凸的,我们开发了一个凸近似。我们提出了一种拟合线性回归模型到目标对准损失的凸逼近的方法。计算结果表明,该方法可以得到比现有回归方法更高质量的解。
{"title":"Objective-Aligned Regression for Two-Stage Linear Programs","authors":"Alexander S. Estes, Jean-Philippe P. Richard","doi":"10.2139/ssrn.3469897","DOIUrl":"https://doi.org/10.2139/ssrn.3469897","url":null,"abstract":"We study an approach to regression that we call objective-aligned fitting, which is applicable when the regression model is used to predict uncertain parameters of some objective problem. Rather than minimizing a typical loss function, such as squared error, we approximately minimize the objective value of the resulting solutions to the nominal optimization problem. While previous work on objective-aligned fitting has tended to focus on uncertainty in the objective function, we consider the case in which the nominal optimization problem is a two-stage linear program with uncertainty in the right-hand side. We define the objective-aligned loss function for the problem and prove structural properties concerning this loss function. Since the objective-aligned loss function is generally non-convex, we develop a convex approximation. We propose a method for fitting a linear regression model to the convex approximation of the objective-aligned loss. Computational results indicate that this procedure can lead to higher-quality solutions than existing regression procedures.","PeriodicalId":406435,"journal":{"name":"CompSciRN: Other Machine Learning (Topic)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131389884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
CompSciRN: Other Machine Learning (Topic)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1