Auditory-model based robust feature selection for speech recognition
Author(s) -
Christos Koniaris,
M. Kuropatwinski,
W. Bastiaan Kleijn
Publication year - 2010
Publication title -
the journal of the acoustical society of america
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.619
H-Index - 187
eISSN - 1520-8524
pISSN - 0001-4966
DOI - 10.1121/1.3284545
Subject(s) - computer science , robustness (evolution) , mel frequency cepstrum , speech recognition , pattern recognition (psychology) , euclidean distance , artificial intelligence , feature selection , discriminant , dimensionality reduction , linear discriminant analysis , feature (linguistics) , frequency domain , feature extraction , computer vision , biochemistry , chemistry , linguistics , philosophy , gene
It is shown that robust dimension-reduction of a feature set for speech recognition can be based on a model of the human auditory system. Whereas conventional methods optimize classification performance, the proposed method exploits knowledge implicit in the auditory periphery, inheriting its robustness. Features are selected to maximize the similarity of the Euclidean geometry of the feature domain and the perceptual domain. Recognition experiments using mel-frequency cepstral coefficients (MFCCs) confirm the effectiveness of the approach, which does not require labeled training data. For noisy data the method outperforms commonly used discriminant-analysis based dimension-reduction methods that rely on labeling. The results indicate that selecting MFCCs in their natural order results in subsets with good performance.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom