• Nie Znaleziono Wyników

Speech Emotion Recognition Using Voiced Segment Selection Algorithm

N/A
N/A
Protected

Academic year: 2021

Share "Speech Emotion Recognition Using Voiced Segment Selection Algorithm"

Copied!
3
0
0

Pełen tekst

(1)

Delft University of Technology

Speech Emotion Recognition Using Voiced Segment Selection Algorithm

Gu, Yu; Postma, Eric; Lin, Hai Xiang; van den Herik, Jaap

DOI

10.3233/978-1-61499-672-9-1682

Publication date

2016

Document Version

Final published version

Published in

ECAI 2016 - 22nd European Conference on Artificial Intelligence

Citation (APA)

Gu, Y., Postma, E., Lin, H. X., & van den Herik, J. (2016). Speech Emotion Recognition Using Voiced

Segment Selection Algorithm. In G. A. Kaminka, M. Fox, P. Bouquet, E. Hüllermeier, V. Dignum, F. Dignum,

& F. van Harmelen (Eds.), ECAI 2016 - 22nd European Conference on Artificial Intelligence (pp.

1682-1683). (Frontiers in Artificial Intelligence and Applications; Vol. 285). IOS Press.

https://doi.org/10.3233/978-1-61499-672-9-1682

Important note

To cite this publication, please use the final published version (if applicable).

Please check the document version above.

Copyright

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons. Takedown policy

Please contact us and provide details if you believe this document breaches copyrights. We will remove access to the work immediately and investigate your claim.

This work is downloaded from Delft University of Technology.

(2)

Speech Emotion Recognition Using Voiced Segment

Selection Algorithm

Yu Gu

1

and Eric Postma

2

and Hai-Xiang Lin

3

and Jaap van den Herik

4

Abstract. Speech emotion recognition (SER) poses one of the ma-jor challenges in human-machine interaction. We propose a new al-gorithm, the Voiced Segment Selection (VSS) alal-gorithm, which can produce an accurate segmentation of speech signals. The VSS al-gorithm deals with the voiced signal segment as the texture image processing feature which is different from the traditional method. It uses the Log-Gabor filters to extract the voiced and unvoiced fea-tures from spectrogram to make the classification. The finding shows that the VSS method is a more accurate algorithm for voiced seg-ment detection. Therefore, it has potential to improve performance of emotion recognition from speech.

1

Introduction

A speech always consists of voiced and unvoiced parts. When vocal chords vibrate to pronounce vowels, voiced segments are generated. In contrast to unvoiced segments, voiced segments show periodic and prosodic signals. Unvoiced segments show irregular signals, gener-ated by the influence of narrow vocal tract. Emotion is an important component of the information contained in a speech. Emotional in-formation in speech is represented in a variety of prosodic types and is also mainly contained in the voiced parts. That is why researchers always focus on the voiced parts of a speech in emotion recognition. The outline of the paper is as follows. Section 2 describes the methodology of VSS method in detail. In Section 3, the set-up of the comparative evaluation of the VSS method is described. Exper-imental results and discussions are presented in Section 4. Finally, Section 5 provides concluding remarks.

2

Formulation of VSS Method

There are two key components in the VSS method which are the spectrogram and the log-Gabor filter.

The VSS method deals with the voiced activity detection as a clas-sification issue. In contrast to existing machine learning algorithm for VAD which normally used the acoustic features or statistical fea-tures for the classification, we propose a novel kind of feature which is extracted by log-Gabor filter from the spectrogarm for this issue. We use two-dimensional Gabor filters which are locally correspond-ing to the orientations of energy bands in the spectrogram.

There-1 Tilburg center of Communication and cognition, Tilburg University,

Tilburg, Netherlands, email: y.gu 1@uvt.nl

2 Tilburg center of Communication and cognition, Tilburg University,

Tilburg, Netherlands

3Institute of Applied Mathematics, Delft University of Technology, Delft,

Netherlands

4Leiden Institute of Advanced Computer Science, Leiden University, Leiden,

Netherlands

fore, by convolving the spectrogram with a Gabor filter of a given spatial frequency and orientation, the convolved spectrogram repre-sents spectro-temporal patterns with the associated spatial frequency (width) and orientation, respectively. We use the log-Gabor filters with combination of scale and orientation for convolving whole spec-trogram to obtain the texture information. Because all the texture pat-terns are generated by voiced parts in the spectrogram, these convo-lution results from the filters are sensitive to the voiced segment for speech signal. Each combination of different scale and orientation will result in one filter-image. The energy of each filter-iamge will be averaged by the time sequence and be stored as a vector data. And finally, all the vectors which calculated by each filters will be assem-ble as a matrix. The Support vector machine will be used to classify the voiced and unvoiced segment by using the convolution feature matrix from spectrogram.

3

Experiment of the VSS method

This section describes the implementation which is used in the ex-periments and evaluation details.

The evaluation of the VSS method was performed on two corpora: the Mandarin Affective Speech (MAS) corpus (MAS, 2007) and the Berlin Database of Emotional Speech (Emo-DB, 2007). More details about the corpus can be found via the Linguistic Data Consortium website5.

The experimental procedure consisted of three steps: (i) spectro-gram generation, (ii) spectrospectro-gram convolution using Log-Gabor fil-ters, (iii) voiced and unvoiced segment classification. In the follow-ing, the details of each of these steps will be outlined.

Table 1. The value of parameters for Log-Gabor filter

name of parameter value of parameter spectrogram time resolutiont 0

scaleNsand orientationNo 12 12

segmented Patch 8

minwavelength 3

mult 1.35

sigmaOnf 0.8

(i) The speech signal was visualized in spectrogram. Each auditory signal (utterance) was transformed into a spectrogram using Matlab’s spectral analysis function. (ii) Each spectrogram is convolved by a Log-Gabor filter with a number different parameter values of scale

5http://catalog.ldc.upenn.edu/LDC2007S09

ECAI 2016 G.A. Kaminka et al. (Eds.) © 2016 The Authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution Non-Commercial License 4.0 (CC BY-NC 4.0). doi:10.3233/978-1-61499-672-9-1682

(3)

and orientation.NsandNodenote the scale and the orientation re-spectively. Therefore, if the number of scales and the orientations are equal toNsandN0respectively, then,Ns∗Nobank filter images are generated. The parameters of log-Gabor filters as shown in Table 1 will be applied to make the convolution to the spectrogram. The con-volution values within each image is averaged by the time sequence and store as a vector. Thus, each vector will have1 ∗ 512 value. (iii) Step (ii) will produceNs∗ Noconvolution image, and the total num-ber of the speech record isNt. ThereforeNs∗No∗Ntvectors which we could achieve after the whole convolution. Combining the energy value vectors into matrices features which we can use for training an SVM classifier.

The performance evaluation consists of two parts: a comparison in accuracy of voiced part detection between the VSS method exiting voiced detection methods and the performance in emotion recogni-tion by the VSS method. First, we compare the VSS method with another prevalent VAD method in the voiced segment detection. Three major algorithms were duplicated: the a) The Energy and ZRC method, b) the statistical likelihood ratio method [1] and c) the Deep belief network method [2].

E&Z LR ML VSS Algorithm 75 80 85 90 Performance

Experiment on AMF Database

(a) AMS database

E&ZCR LR ML VSS Algorithm 76 78 80 82 84 86 88 90 92 94 Performance

Experiment on Berlin Emotion Database

(b) Berlin database

Figure 1. Comparison of the voiced part accuracy performances obtained on both database.

Figure 1 show the comparison of the voice accuracy between the three major methods and our VSS method by using the AMS and DB database respectively. We can clearly observe that our method was significantly better than the traditional ZCR and LR method. There are three possible explanations are as follows. Firstly, the spectro-gram contains various kinds of most occurring acoustic phenomena, such as hamonicity, formants, vertical edges which are all the crucial characters for indicating the voice parts. And the formant and ver-tical edges are exactly the start and end point for the voice signal. Therefore, it can detect all acoustic phenomena which in other words to recognize the voice speech in a robust manner. Secondly, tradi-tional methods have limitation due to manually set thresholds, the value of threshold was not always optimized. Compare to that, the SVM has a strong ability for the high dimensional feature learning that for the voice and unvoiced discrimination which can be less in-fluence than manually and more accurate. Last reason for the higher accuracy is due to that, the log-Gabor filters has strong ability to distinguish the phonetic phenomena from noise background in spec-trogram. Even through the noise in speech which can bring a huge number of corresponded peaks to the spectrogram. The reason which can be explained this is because the output of Log-Gabor filters is achieved from integrating the entire 2-D spectrogram information, which make the 2-D filterbank more robust ability to the noise. More-over, the DBN method slightly outperforms the VSS method.

How-ever, for the deep learning it requires huge computation consumption. Therefore, we conclude that the VSS method is a useful and practical method to improve the voice detection accuracy than the traditional method.

4

Experiment of Speech Emotion Recognition

In the speech emotion recognition experiment, the acoustic features were extracted from the voiced segments based on the VSS method. The performance of speech emotion recognition was also compared with and without VSS procedure before the feature extraction. To avoid over-fitting due to the PCA and the SVM parameter optimiza-tion, the evaluation was performed using a nested10 × 10-fold cross validation,. 90.5 91 91.5 92 92.5 93 93.5 94 94.5 95

Acoustic process AMS VSS Process on AMS

Algorithm

Performance

(a) AMS database

85 86 87 88 89 90 91 92 93

Acoustic process on DB VSS Process on DB

Algorithm

Performance

(b) Berlin database

Figure 2. Comparison of the recognition performances obtained from both database.

The classification was applied in both corpus of MAS and Emo-DB. Figure 2 visualizes the classification performances with and without VSS before feature extraction. Both of the features were base on the state-on-art baseline. The comparative evaluation of employ-ing VSS method against without VSS method. It demonstrates that the performance of speech emotion recognition with VSS method led to a non-overlapping improvement. The results showed that VSS method for voiced detection is more effective in extracting features, because most of the acoustic features are related to the voiced parts.

5

Conclusions

We proposed a novel VSS algorithm which uses the log-Gabor filter to extract the features from the spectrogram for the voiced segment classification. Experiment results showed that the VSS algorithm can be a useful and practical method for the voiced activity detection. Furthermore, the performance in speech emotion recognition shows that the classification rate is improved for both Chinese and Ger-man speech. It convinced us the VSS is an useful compliment for the speech emotion recognition.

REFERENCES

[1] Youngjoo Suh and Hoirin Kim, ‘Multiple acoustic model-based discrim-inative likelihood ratio weighting for voice activity detection.’, IEEE

Sig-nal Process. Lett., 19(8), 507–510, (2012).

[2] Xiao-Lei Zhang and Ji Wu, ‘Deep belief networks based voice activity detection’, Audio, Speech, and Language Processing, IEEE Transactions

on, 21(4), 697–710, (2013).

Cytaty

Powiązane dokumenty

The classifiers selected for emotion class happiness have similar performance with 71 training samples in the case of Berlin data set and 48 training samples for

Przekłady Literatur Słowiańskich 7/2,

In this paper, results of dimension reduction in feature space for thyroid ultrasound images using the heuristic identification of noisy variables, testing the significance

W aspekcie „genetycznym” można by zatem mówić o relacji „inwersja” – „manie- ryzm”, powołując się na biografie: Leonarda, Michała Anioła, Szekspira (tych przy-

To właśnie w nim, obok modelu duszpasterstwa salezjańskiego opartego na dynamice wspólnoty wychowawczo-duszpasterskiej, znalazły się doktryna i założenia

Osłabły zatem — choć w nierównym stopniu — wszystkie cechy czyniące w poprzednim okre­ sie ze świadomości rewolucyjnej formę świadomości społecznej, zaczął się proces

Regarding the role of emotion in the relationship with dark triad traits and also the hypothesis of the existence of a relationship between dark triad traits and deficiency

The key aim of the study was to better understand the motivational power of the future L2 selves on students’ L2 performance. In order to do this, the study.. applied Dörnyei’s