[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Reordering Features with Weights Fusion in Multiclass and Multiple-Kernel Speech Emotion Recognition

Published: 01 January 2017 Publication History

Abstract

The selection of feature subset is a crucial aspect in speech emotion recognition problem. In this paper, a Reordering Features with Weights Fusion (RFWF) algorithm is proposed for selecting more effective and compact feature subset. The RFWF algorithm fuses the weights reflecting the relevance, complementarity, and redundancy between features and classes comprehensively and implements the reordering of features to construct feature subset with excellent emotional recognizability. A binary-tree structured multiple-kernel SVM classifier is adopted in emotion recognition. And different feature subsets are selected in different nodes of the classifier. The highest recognition accuracy of the five emotions in Berlin database is 90.549% with only 15 features selected by RFWF. The experimental results show the effectiveness of RFWF in building feature subset and the utilization of different feature subsets for specified emotions can improve the overall recognition performance.

References

[1]
L. Chen, X. Mao, Y. Xue, and L. L. Cheng, “Speech emotion recognition: features and classification models,” Digital Signal Processing, vol. 22, no. 6, pp. 1154–1160, 2012.
[2]
S. Chandaka, A. Chatterjee, and S. Munshi, “Support vector machines employing cross-correlation for emotional speech recognition,” Measurement: Journal of the International Measurement Confederation, vol. 42, no. 4, pp. 611–618, 2009.
[3]
C.-C. Lee, E. Mower, C. Busso, S. Lee, and S. Narayanan, “Emotion recognition using a hierarchical binary decision tree approach,” Speech Communication, vol. 53, no. 9-10, pp. 1162–1171, 2011.
[4]
J. Yuan, L. Chen, T. Fan, and J. Jia, “Dimension reduction of speech emotion feature based on weighted linear discriminate analysis,” in Image Processing and Pattern Recognition, vol. 8, pp. 299–308, International Journal of Signal Processing, 2015.
[5]
Y. Saeys, I. Inza, and P. Larrañaga, “A review of feature selection techniques in bioinformatics,” Bioinformatics, vol. 23, no. 19, pp. 2507–2517, 2007.
[6]
H. Hua Yang and J. Moody, “Data visualization and feature selection: new algorithms for nongaussian data,” Advances in Neural Information Processing Systems, vol. 12, pp. 687–693, 1999.
[7]
D. Gharavian, M. Sheikhan, A. Nazerieh, and S. Garoucy, “Speech emotion recognition using FCBF feature selection method and GA-optimized fuzzy ARTMAP neural network,” Neural Computing and Applications, vol. 21, no. 8, pp. 2115–2126, 2012.
[8]
H. Peng, F. Long, and C. Ding, “Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 8, pp. 1226–1238, 2005.
[9]
D. Lin and X. Tang, “Conditional infomax learning: an integrated framework for feature extraction and fusion,” in Proceedings of the 9th European Conference on Computer Vision, pp. 68–82, Graz, Austria, 2006.
[10]
P. E. Meyer, C. Schretter, and G. Bontempi, “Information-theoretic feature selection in microarray data using variable complementarity,” IEEE Journal on Selected Topics in Signal Processing, vol. 2, no. 3, pp. 261–274, 2008.
[11]
X. Zhao, S. Zhang, and B. Lei, “Robust emotion recognition in noisy speech via sparse representation,” Neural Computing and Applications, vol. 24, no. 7-8, pp. 1539–1553, 2014.
[12]
A. Mencattini, E. Martinelli, G. Costantini, M. Todisco, B. Basile, M. Bozzali, and C. Di Natale, “Speech emotion recognition using amplitude modulation parameters and a combined feature selection procedure,” Knowledge-Based Systems, vol. 63, pp. 68–81, 2014.
[13]
D. Ververidis, C. Kotropoulos, and I. Pitas, “Automatic emotional speech classification,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, pp. I-593–596, Quebec, CA, USA, 2004.
[14]
J. Liu, C. Chen, J. Bu et al., “Speech emotion recognition based on a fusion of all-class and pairwise-class feature selection,” in Proceedings of the ICCS, pp. 168–175, Beijing, China, 2007.
[15]
X. Xu, Y. Li, X. Xu, Z. Wen, H. Che, S. Liu, and J. Tao, “Survey on discriminative feature selection for speech emotion recognition,” in Proceedings of the 9th International Symposium on Chinese Spoken Language Processing, ISCSLP 2014, pp. 345–349, Singapore, Singapore, September 2014.
[16]
L. Tian, X. Jiang, and Z. Hou, “Statistical study on the diversity of pitch parameters in multilingual speech,” Control and Decision, vol. 20, no. 11, pp. 1311–1313, 2005.
[17]
F. Burkhardt, A. Paeschke, M. Rolfes et al., “A database of German emotional speech,” in Proceedings of the 9th European Conference on Speech Communication and Technology, pp. 1517–1520, Lisbon, Portugal, 2005.
[18]
J. R. Vergara and P. A. Estévez, “A review of feature selection methods based on mutual information,” Neural Computing and Applications, vol. 24, no. 1, pp. 175–186, 2014.
[19]
C.-Y. Yeh, W.-P. Su, and S.-J. Lee, “An efficient multiple-kernel learning for pattern classification,” Expert Systems with Applications, vol. 40, no. 9, pp. 3491–3499, 2013.
[20]
G. R. G. Lanckriet, N. Cristianini, P. L. Bartlett et al., “Learning the kernel matrix with semidefinite programming,” Machine Learning Research, vol. 5, no. 1, pp. 27–72, 2004.
[21]
X. Jiang, K. Xia, X. Xia, and B. Zu, “Speech emotion recognition using semi-definite programming multiple-kernel SVM,” Journal of Beijing University of Posts and Telecommunications, vol. 38, no. S1, pp. 67–71, 2015.
[22]
B. Yang and M. Lugger, “Emotion recognition from speech signals using new harmony features,” Signal Processing, vol. 90, no. 5, pp. 1415–1423, 2010.
[23]
Y. Jin, P. Song, W. Zheng, and L. Zhao, “Novel feature fusion method for speech emotion recognition based on multiple kernel learning,” Journal of Southeast University, vol. 29, no. 2, pp. 129–133, 2013.

Index Terms

  1. Reordering Features with Weights Fusion in Multiclass and Multiple-Kernel Speech Emotion Recognition
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Please enable JavaScript to view thecomments powered by Disqus.

          Information & Contributors

          Information

          Published In

          cover image Journal of Electrical and Computer Engineering
          Journal of Electrical and Computer Engineering  Volume 2017, Issue
          2017
          596 pages
          ISSN:2090-0147
          EISSN:2090-0155
          Issue’s Table of Contents
          This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

          Publisher

          Hindawi Limited

          London, United Kingdom

          Publication History

          Published: 01 January 2017

          Qualifiers

          • Research-article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • 0
            Total Citations
          • 0
            Total Downloads
          • Downloads (Last 12 months)0
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 12 Jan 2025

          Other Metrics

          Citations

          View Options

          View options

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media