Optimal prototype selection for speech emotion recognition using fuzzy k-important nearest neighbour Online publication date: Mon, 12-Sep-2016
by Zhen Xing Zhang; Joon Shik Lim; Zhao Cai Jiang; Chun Jie Zhou; Shao Jing Li
International Journal of Communication Networks and Distributed Systems (IJCNDS), Vol. 17, No. 2, 2016
Abstract: Speech emotion recognition has been a popular topic of affective computing. Accuracy in speech emotion recognition depends on selecting the optimal prototype. In this paper, a new 2-D emotional speech recognition model based on a fuzzy k-important nearest neighbour (FKINN) and neuro-fuzzy network is described. In the FKINN algorithm, an important nearest neighbour selection rule is introduced. The neuro-fuzzy network applies a bounded sum of weighted fuzzy membership functions (BSWFM). During the training process, BSWFM calculates the Takagi-Sugeno defuzzification values for the 2-D visual model. The emotional speech signals used in this work were obtained from the Berlin emotional speech database. The proposed new model achieves 83.5% overall classification accuracy with the 2-D emotional speech recognition model. The classification accuracies of anger, happiness, sadness, and neutral were 94.1%, 65.9%, 81.1%, and 87.5%, respectively.
Existing subscribers:
Go to Inderscience Online Journals to access the Full Text of this article.
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Communication Networks and Distributed Systems (IJCNDS):
Login with your Inderscience username and password:
Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.
If you still need assistance, please email subs@inderscience.com