Texto completo
Vista Previa |
PDF (Portable Document Format)
- Se necesita un visor de ficheros PDF, como GSview, Xpdf o Adobe Acrobat Reader
Descargar (288kB) | Vista Previa |
ORCID: https://orcid.org/0000-0003-3877-0089, Lebai Lutfi, Syaheerah Binti, Lucas Cuesta, Juan Manuel, Macías Guarasa, Javier, Montero Martínez, Juan Manuel
ORCID: https://orcid.org/0000-0002-7908-5400, San Segundo Hernández, Rubén
ORCID: https://orcid.org/0000-0001-9659-5464 and Pardo Muñoz, José Manuel
ORCID: https://orcid.org/0000-0002-1009-590X
(2009).
Acoustic Emotion Recognition using Dynamic Bayesian Networks and Multi-Space Distributions.
En: "10th Annual Conference of the International Speech Communication Association, Interspeech 2009", 06/09/2009 - 10/09/2009, Brighton, UK.
| Título: | Acoustic Emotion Recognition using Dynamic Bayesian Networks and Multi-Space Distributions |
|---|---|
| Autor/es: |
|
| Tipo de Documento: | Ponencia en Congreso o Jornada (Artículo) |
| Título del Evento: | 10th Annual Conference of the International Speech Communication Association, Interspeech 2009 |
| Fechas del Evento: | 06/09/2009 - 10/09/2009 |
| Lugar del Evento: | Brighton, UK |
| Título del Libro: | Proceedings of 10th Annual Conference of the International Speech Communication Association, Interspeech 2009 |
| Fecha: | 2009 |
| Materias: | |
| ODS: | |
| Palabras Clave Informales: | automatic emotion recognition, multi-space probability distribution, dynamic bayesian networks, emotion challenge. |
| Escuela: | E.T.S.I. Telecomunicación (UPM) |
| Departamento: | Ingeniería Electrónica |
| Licencias Creative Commons: | Reconocimiento - Sin obra derivada - No comercial |
Vista Previa |
PDF (Portable Document Format)
- Se necesita un visor de ficheros PDF, como GSview, Xpdf o Adobe Acrobat Reader
Descargar (288kB) | Vista Previa |
In this paper we describe the acoustic emotion recognition system built at the Speech Technology Group of the Universidad Politecnica de Madrid (Spain) to participate in the INTERSPEECH 2009 Emotion Challenge. Our proposal is based on the use of a Dynamic Bayesian Network (DBN) to deal with the temporal modelling of the emotional speech information. The selected features (MFCC, F0, Energy and their variants) are modelled as different streams, and the F0 related ones are integrated under a Multi Space Distribution (MSD) framework, to properly model its dual nature (voiced/unvoiced). Experimental evaluation on the challenge test set, show a 67.06%and 38.24% of unweighted recall for the 2 and 5-classes tasks respectively. In the 2-class case, we achieve similar results compared with the baseline, with a considerable less number of features. In the 5-class case, we achieve a statistically significant 6.5% relative improvement
| ID de Registro: | 5575 |
|---|---|
| Identificador DC: | https://oa.upm.es/5575/ |
| Identificador OAI: | oai:oa.upm.es:5575 |
| URL Oficial: | http://www.isca-speech.org/archive/interspeech_200... |
| Depositado por: | Memoria Investigacion |
| Depositado el: | 23 Dic 2010 08:44 |
| Ultima Modificación: | 20 Abr 2016 14:21 |
Publicar en el Archivo Digital desde el Portal Científico