Variable Selection of Lasso and Large Model

被引:1
作者
Xia, Huiyi [1 ]
机构
[1] Chizhou Univ, Sch Big Data & Artificial Intelligence, Chizhou 240007, Peoples R China
关键词
Input variables; Estimation; Computational modeling; Data models; Complexity theory; Biological system modeling; Software engineering; Chatbots; Variable selection; Lasso; AIC; forward stagewise; complexity; ChatGPT;
D O I
10.1109/ACCESS.2023.3312015
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In order to clarify the variable selection of Lasso, Lasso is compared with two other variable selection methods AIC and forward stagewise. First, the variable selection of Lasso was compared with that of AIC, and it was discovered that Lasso has a wider application range than AIC. The data simulation shows the variable selection of Lasso under orthonormal design is consistent with AIC, Lasso under orthonormal design can be solved by using the stepwise selection algorithm. The removed variables of Lasso appear again under nonorthonormal design, the variable selection of Lasso under nonorthonormal design isn't consistent with AIC. We continue to compare the variable selection of Lasso and forward stagewise. Based on the analysis of these studies, it is pointed out that the variable selection of Lasso is complex. An infinite number of parameters enable the design matrix to achieve orthonormalization, so that the solution of Lasso can be found with the stepwise selection algorithm, which may be the reason for the success of the large model represented by ChatGPT.
引用
收藏
页码:96514 / 96521
页数:8
相关论文
共 13 条
[1]   Two-Step Lasso Estimation of the Spatial Weights Matrix [J].
Ahrens, Achim ;
Bhattacharjee, Arnab .
ECONOMETRICS, 2015, 3 (01) :128-155
[2]   Penalized Regression and Risk Prediction in Genome-Wide Association Studies [J].
Austin, Erin ;
Pan, Wei ;
Shen, Xiaotong .
STATISTICAL ANALYSIS AND DATA MINING, 2013, 6 (04) :315-328
[3]  
Borboudakis G, 2019, J MACH LEARN RES, V20
[4]  
Breaux H. J., Tech. Rep
[5]   BETTER SUBSET REGRESSION USING THE NONNEGATIVE GARROTE [J].
BREIMAN, L .
TECHNOMETRICS, 1995, 37 (04) :373-384
[6]   Least angle regression - Rejoinder [J].
Efron, B ;
Hastie, T ;
Johnstone, I ;
Tibshirani, R .
ANNALS OF STATISTICS, 2004, 32 (02) :494-499
[7]   Variable selection in functional additive regression models [J].
Febrero-Bande, Manuel ;
Gonzalez-Manteiga, Wenceslao ;
Oviedo de la Fuente, Manuel .
COMPUTATIONAL STATISTICS, 2019, 34 (02) :469-487
[8]   A New Analysis Strategy for DesignsWith Complex Aliasing [J].
Kane, Andrew ;
Mandal, Abhyuday .
AMERICAN STATISTICIAN, 2020, 74 (03) :274-281
[9]   Variable selection for survival data with a class of adaptive elastic net techniques [J].
Khan, Md Hasinur Rahaman ;
Shaw, J. Ewart H. .
STATISTICS AND COMPUTING, 2016, 26 (03) :725-741
[10]   VSOLassoBag: a variable-selection oriented LASSO bagging algorithm for biomarker discovery in omic-based translational research [J].
Liang, Jiaqi ;
Wang, Chaoye ;
Zhang, Di ;
Xie, Yubin ;
Zeng, Yanru ;
Li, Tianqin ;
Zuo, Zhixiang ;
Ren, Jian ;
Zhao, Qi .
JOURNAL OF GENETICS AND GENOMICS, 2023, 50 (03) :151-162