{"title":"修正多重共线性问题的高维回归模型的一种新的收缩方法","authors":"Z. Ghareeb, Suhad Ali Shaheed Al-Temimi","doi":"10.21533/pen.v11i3.3550","DOIUrl":null,"url":null,"abstract":"This research seeks to present new method of shrinking variables to select some basic variables from large data sets. This new shrinkage estimator is a modification of (Ridge and Adaptive Lasso) shrinkage regression method in the presence of the mixing parameter that was calculated in the Elastic-Net. The Proposed estimator is called (Improved Mixed Shrinkage Estimator (IMSHE)) to handle the problem of multicollinearity. In practice, it is difficult to achieve the required accuracy and efficiency when dealing with a big data set, especially in the case of multicollinearity problem between the explanatory variables. By using Basic shrinkage methods (Lasso, Adaptive Lasso and Elastic Net) and comparing their results with the New shrinkage method (IMSH) was applied to a set of obesity -related data containing (52) variables for a sample of (112) observations. All shrinkage methods have also been compared for efficiency through Mean Square Error (MSE) criterion and Cross Validation Parameter (CVP). The results showed that the best shrinking parameter among the four methods (Lasso, Adaptive Lasso, Elastic Net and IMSH) was for the IMSH shrinkage method, as it corresponds to the lowest (MSE) based on the cross-validation parameter test (CVP). The new proposed method IMSH achieved the optimal shrinking parameter (λ = 0.6932827) according to the (CVP) test, that leads to have minimum value of mean square error (MSE) equal (0.2576002). The results showed when the value of the regularization parameter increases, the value of the shrinkage parameter decreases to become equal to zero, so the ideal number of variables after shrinkage is (p=6)","PeriodicalId":37519,"journal":{"name":"Periodicals of Engineering and Natural Sciences","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A new shrinkage method for higher dimensions regression model to remedy of multicollinearity problem\",\"authors\":\"Z. Ghareeb, Suhad Ali Shaheed Al-Temimi\",\"doi\":\"10.21533/pen.v11i3.3550\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This research seeks to present new method of shrinking variables to select some basic variables from large data sets. This new shrinkage estimator is a modification of (Ridge and Adaptive Lasso) shrinkage regression method in the presence of the mixing parameter that was calculated in the Elastic-Net. The Proposed estimator is called (Improved Mixed Shrinkage Estimator (IMSHE)) to handle the problem of multicollinearity. In practice, it is difficult to achieve the required accuracy and efficiency when dealing with a big data set, especially in the case of multicollinearity problem between the explanatory variables. By using Basic shrinkage methods (Lasso, Adaptive Lasso and Elastic Net) and comparing their results with the New shrinkage method (IMSH) was applied to a set of obesity -related data containing (52) variables for a sample of (112) observations. All shrinkage methods have also been compared for efficiency through Mean Square Error (MSE) criterion and Cross Validation Parameter (CVP). The results showed that the best shrinking parameter among the four methods (Lasso, Adaptive Lasso, Elastic Net and IMSH) was for the IMSH shrinkage method, as it corresponds to the lowest (MSE) based on the cross-validation parameter test (CVP). The new proposed method IMSH achieved the optimal shrinking parameter (λ = 0.6932827) according to the (CVP) test, that leads to have minimum value of mean square error (MSE) equal (0.2576002). The results showed when the value of the regularization parameter increases, the value of the shrinkage parameter decreases to become equal to zero, so the ideal number of variables after shrinkage is (p=6)\",\"PeriodicalId\":37519,\"journal\":{\"name\":\"Periodicals of Engineering and Natural Sciences\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Periodicals of Engineering and Natural Sciences\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.21533/pen.v11i3.3550\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"Engineering\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Periodicals of Engineering and Natural Sciences","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21533/pen.v11i3.3550","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Engineering","Score":null,"Total":0}
引用次数: 0
摘要
本研究试图提出一种新的收缩变量的方法,从大数据集中选择一些基本变量。这种新的收缩估计器是对(Ridge and Adaptive Lasso)收缩回归方法的改进,在弹性网中计算混合参数的情况下。所提出的估计器被称为(改进的混合收缩估计器(IMSHE))来处理多重共线性问题。在实践中,处理大数据集时很难达到所需的准确性和效率,尤其是在解释变量之间存在多重共线性问题的情况下。通过使用基本收缩方法(Lasso、Adaptive Lasso和Elastic Net)并将其结果与新收缩方法(IMSH)进行比较,将其应用于一组肥胖相关数据,该数据包含(112)个观察样本的(52)个变量。还通过均方误差(MSE)准则和交叉验证参数(CVP)对所有收缩方法的效率进行了比较。结果表明,在四种方法(Lasso、Adaptive Lasso、Elastic Net和IMSH)中,IMSH收缩方法的收缩参数最好,因为它对应于基于交叉验证参数测试(CVP)的最低MSE。根据(CVP)检验,新提出的方法IMSH获得了最佳收缩参数(λ=0.6932827),从而使均方误差(MSE)的最小值等于(0.2576002)。结果表明,当正则化参数的值增加时,收缩参数的值减小为等于零,因此收缩后的理想变量数为(p=6)
A new shrinkage method for higher dimensions regression model to remedy of multicollinearity problem
This research seeks to present new method of shrinking variables to select some basic variables from large data sets. This new shrinkage estimator is a modification of (Ridge and Adaptive Lasso) shrinkage regression method in the presence of the mixing parameter that was calculated in the Elastic-Net. The Proposed estimator is called (Improved Mixed Shrinkage Estimator (IMSHE)) to handle the problem of multicollinearity. In practice, it is difficult to achieve the required accuracy and efficiency when dealing with a big data set, especially in the case of multicollinearity problem between the explanatory variables. By using Basic shrinkage methods (Lasso, Adaptive Lasso and Elastic Net) and comparing their results with the New shrinkage method (IMSH) was applied to a set of obesity -related data containing (52) variables for a sample of (112) observations. All shrinkage methods have also been compared for efficiency through Mean Square Error (MSE) criterion and Cross Validation Parameter (CVP). The results showed that the best shrinking parameter among the four methods (Lasso, Adaptive Lasso, Elastic Net and IMSH) was for the IMSH shrinkage method, as it corresponds to the lowest (MSE) based on the cross-validation parameter test (CVP). The new proposed method IMSH achieved the optimal shrinking parameter (λ = 0.6932827) according to the (CVP) test, that leads to have minimum value of mean square error (MSE) equal (0.2576002). The results showed when the value of the regularization parameter increases, the value of the shrinkage parameter decreases to become equal to zero, so the ideal number of variables after shrinkage is (p=6)