OPTIMAL CLASSIFICATION FOR FUNCTIONAL DATA

被引:2
作者
Wang, Shuoyang [1 ]
Shang, Zuofeng [2 ]
Cao, Guanqun [3 ]
Liu, Jun S. [4 ]
机构
[1] Univ Louisville, Dept Bioinformat & Biostat, Louisville, KY 40202 USA
[2] New Jersey Inst Technol, Dept Math Sci, Newark, NJ 07102 USA
[3] Michigan State Univ, Dept Stat & Probabil, E Lansing, MI 48824 USA
[4] Harvard Univ, Dept Stat, Cambridge, MA 02138 USA
关键词
Functional classification; functional deep neural network; functional quadratic discriminant analysis; Gaussian process; minimax excess misclassification risk; DEEP NEURAL-NETWORKS; MEAN FUNCTION; DISCRIMINATION; MINIMAX; CLASSIFIERS; RATES;
D O I
10.5705/ss.202022.0057
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
A central topic in functional data analysis is how to design an optimal decision rule, based on training samples, to classify a data function. We exploit the optimal classification problem in which the data functions are Gaussian processes. We derive sharp convergence rates for the minimax excess misclassification risk both when the data functions are fully observed and when they are discretely observed. We explore two easily implementable classifiers, based on a discriminant analysis and on a deep neural network, respectively, which both achieve optimality in Gaussian settings. Our deep neural network classifier is new in the literature, and demonstrates outstanding performance, even when the data functions are nonGaussian. For discretely observed data, we discover a novel critical sampling frequency that governs the sharp convergence rates. The proposed classifiers perform favorably in finite-sample applications, shown in comparisons with other functional classifiers in simulations and one real-data application.
引用
收藏
页码:1545 / 1564
页数:20
相关论文
共 32 条
[1]  
Anderson T.W., 2003, An introduction to multivariate statistical analysis, Vthird
[2]   Functional Logistic Discrimination Via Regularized Basis Expansions [J].
Araki, Yuko ;
Konishi, Sadanori ;
Kawano, Shuichi ;
Matsui, Hidetoshi .
COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2009, 38 (16-17) :2944-2957
[3]   ON DEEP LEARNING AS A REMEDY FOR THE CURSE OF DIMENSIONALITY IN NONPARAMETRIC REGRESSION [J].
Bauer, Benedikt ;
Kohler, Michael .
ANNALS OF STATISTICS, 2019, 47 (04) :2261-2285
[4]   On the Use of Reproducing Kernel Hilbert Spaces in Functional Classification [J].
Berrendero, Jose R. ;
Cuevas, Antonio ;
Torrecilla, Jose L. .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2018, 113 (523) :1210-1218
[5]   High dimensional linear discriminant analysis: optimality, adaptive algorithm and missing data [J].
Cai, T. Tony ;
Zhang, Linjun .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2019, 81 (04) :675-705
[6]   Minimax and Adaptive Prediction for Functional Linear Regression [J].
Cai, T. Tony ;
Yuan, Ming .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2012, 107 (499) :1201-1216
[7]   OPTIMAL ESTIMATION OF THE MEAN FUNCTION BASED ON DISCRETELY SAMPLED FUNCTIONAL DATA: PHASE TRANSITION [J].
Cai, T. Tony ;
Yuan, Ming .
ANNALS OF STATISTICS, 2011, 39 (05) :2330-2355
[8]   Optimal Bayes classifiers for functional data and density ratios [J].
Dai, Xiongtao ;
Mueller, Hans-Georg ;
Yao, Fang .
BIOMETRIKA, 2017, 104 (03) :545-560
[9]   Componentwise classification and clustering of functional data [J].
Delaigle, A. ;
Hall, P. ;
Bathia, N. .
BIOMETRIKA, 2012, 99 (02) :299-313
[10]   Classification Using Censored Functional Data [J].
Delaigle, Aurore ;
Hall, Peter .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2013, 108 (504) :1269-1283