Estimation Under Model Misspecification With Fake Features

被引:9
作者
Hellkvist, Martin [1 ]
Ozcelikkale, Ayca [1 ]
Ahlen, Anders [1 ]
机构
[1] Uppsala Univ, Dept Elect Engn, S-75121 Uppsala, Sweden
基金
瑞典研究理事会;
关键词
Estimation; Data models; Dictionaries; Covariance matrices; Noise measurement; Noise level; Focusing; Model uncertainty; model mismatch; robustness; LINEAR-MODEL; REGULARIZATION; REGRESSION; BOUNDS;
D O I
10.1109/TSP.2023.3237174
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We consider estimation under model misspecification where there is a model mismatch between the underlying system, which generates the data, and the model used during estimation. We propose a model misspecification framework which enables a joint treatment of the model misspecification types of having fake features as well as incorrect covariance assumptions on the unknowns and the noise. We present a decomposition of the output error into components that relate to different subsets of the model parameters corresponding to underlying, fake and missing features. Here, fake features are features which are included in the model but are not present in the underlying system. Under this framework, we characterize the estimation performance and reveal trade-offs between the number of samples, number of fake features, and the possibly incorrect noise level assumption. In contrast to existing work focusing on incorrect covariance assumptions or missing features, fake features is a central component of our framework. Our results show that fake features can significantly improve the estimation performance, even though they are not correlated with the features in the underlying system. In particular, we show that the estimation error can be decreased by including more fake features in the model, even to the point where the model is overparametrized, i.e., the model contains more unknowns than observations.
引用
收藏
页码:47 / 60
页数:14
相关论文
共 37 条
[1]  
Abu-Mostafa Y. S., 2012, LEARNING DATA
[2]   GENERATION OF RANDOM ORTHOGONAL MATRICES [J].
ANDERSON, TW ;
OLKIN, I ;
UNDERHILL, LG .
SIAM JOURNAL ON SCIENTIFIC AND STATISTICAL COMPUTING, 1987, 8 (04) :625-629
[3]   Benign overfitting in linear regression [J].
Bartlett, Peter L. ;
Long, Philip M. ;
Lugosi, Gabor ;
Tsigler, Alexander .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2020, 117 (48) :30063-30070
[4]   Two Models of Double Descent for Weak Features [J].
Belkin, Mikhail ;
Hsu, Daniel ;
Xu, Ji .
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (04) :1167-1180
[5]   Reconciling modern machine-learning practice and the classical bias-variance trade-off [J].
Belkin, Mikhail ;
Hsu, Daniel ;
Ma, Siyuan ;
Mandal, Soumik .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (32) :15849-15854
[6]   Compressed Sensing with Basis Mismatch: Performance Bounds and Sparse-Based Estimator [J].
Bernhardt, Stephanie ;
Boyer, Remy ;
Marcos, Sylvie ;
Larzabal, Pascal .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2016, 64 (13) :3483-3494
[7]   HOW MANY VARIABLES SHOULD BE ENTERED IN A REGRESSION EQUATION [J].
BREIMAN, L ;
FREEDMAN, D .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1983, 78 (381) :131-136
[8]   Sensitivity to Basis Mismatch in Compressed Sensing [J].
Chi, Yuejie ;
Scharf, Louis L. ;
Pezeshki, Ali ;
Calderbank, A. Robert .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2011, 59 (05) :2182-2195
[9]   On the mean and variance of the generalized inverse of a singular Wishart matrix [J].
Cook, R. Dennis ;
Forzani, Liliana .
ELECTRONIC JOURNAL OF STATISTICS, 2011, 5 :146-158
[10]  
d'Ascoli S, 2021, ADV NEUR IN, V34