Linear Sequence Discriminant Analysis: A Model-Based Dimensionality Reduction Method for Vector Sequences
被引:27
作者:
Su, Bing
论文数: 0引用数: 0
h-index: 0
机构:
Tsinghua Univ, Dept Elect Engn, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R ChinaTsinghua Univ, Dept Elect Engn, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
Su, Bing
[1
]
Ding, Xiaoqing
论文数: 0引用数: 0
h-index: 0
机构:
Tsinghua Univ, Dept Elect Engn, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R ChinaTsinghua Univ, Dept Elect Engn, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
Ding, Xiaoqing
[1
]
机构:
[1] Tsinghua Univ, Dept Elect Engn, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
来源:
2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV)
|
2013年
关键词:
HANDWRITING RECOGNITION;
WORD RECOGNITION;
D O I:
10.1109/ICCV.2013.115
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Dimensionality reduction for vectors in sequences is challenging since labels are attached to sequences as a whole. This paper presents a model-based dimensionality reduction method for vector sequences, namely linear sequence discriminant analysis (LSDA), which attempts to find a subspace in which sequences of the same class are projected together while those of different classes are projected as far as possible. For each sequence class, an HMM is built from states of which statistics are extracted. Means of these states are linked in order to form a mean sequence, and the variance of the sequence class is defined as the sum of all variances of component states. LSDA then learns a transformation by maximizing the separability between sequence classes and at the same time minimizing the within-sequence class scatter DTW distance between mean sequences is used to measure the separability between sequence classes. We show that the optimization problem can be approximately transformed into an eigen decomposition problem. LDA can be seen as a special case of LSDA by considering non-sequential vectors as sequences of length one. The effectiveness of the proposed LSDA is demonstrated on two individual sequence datasets from UCI machine learning repository as well as two concatenate sequence datasers. APTI Arabic printed text database and IFN/ENIT Arabic handwriting database.