Indian Journal of Science and Technology
DOI: 10.17485/ijst/2016/v9i39/95585
Year: 2016, Volume: 9, Issue: 39, Pages: 1-5
Original Article
Puja Ramesh Chaudhari and John Sahaya Rani Alex*
School of Electronics Engineering, VIT University, Chennai - 632014, Tamil Nadu, India; [email protected]
[email protected].
*Author for correspondence
John Sahaya Rani Alex
School of Electronics Engineering
Email:[email protected].
Background/Objective: Speech is one of the modes for Human Computer Interface (HCI). Speech contains message to convey as well as the speaker characteristics such as speaker identity and emotional state of the speaker. Recently, researchers are taking more interest in the emotional parameters of speech signals which helps to improve the functionality of HCI. This research focus on selecting features which helps to identify the emotion of the speaker. Methods/Statistical Analysis: Mel Frequency Cepstrum Coefficient (MFCC), Linear Prediction Cepstrum Coefficient (LPCC) and Perceptual Linear Predictive (PLP) methods are used to extract the features. Each emotion is modeled as one Hidden Markov Model (HMM) using Hidden Markov Tool Kit (HTK tool kit). The Beagle Bone Black (BBB) board is chosen for the implementation because of the form factor. Findings: The results indicate that MFCC features gives 100% accuracy for surprise emotion, PLP features gives 100% accuracy for anger emotion and LPCC features give 100% accuracy for fear emotion. Conclusion/Improvement: A hybrid feature extraction method should be devised to detect all emotions with 100% accuracy.
Keywords: BBB, Emotion recognition, HCI, HMM, LPCC, MFCC
Subscribe now for latest articles and news.