Proposing a speech to gesture translation architecture for Spanish deaf people.

San Segundo Hernández, Rubén ORCID: https://orcid.org/0000-0001-9659-5464, Montero Martínez, Juan Manuel ORCID: https://orcid.org/0000-0002-7908-5400, Macías Guarasa, Javier, Córdoba Herralde, Ricardo de ORCID: https://orcid.org/0000-0002-7136-9636, Ferreiros López, Javier ORCID: https://orcid.org/0000-0001-8834-3080 and Pardo Muñoz, José Manuel ORCID: https://orcid.org/0000-0002-1009-590X (2008). Proposing a speech to gesture translation architecture for Spanish deaf people.. "Journal of Visual Languages and Computing", v. 19 (n. 5); pp. 523-538. ISSN 1045-926X. https://doi.org/10.1016/j.jvlc.2007.06.002.

Descripción

Título: Proposing a speech to gesture translation architecture for Spanish deaf people.
Autor/es:
Tipo de Documento: Artículo
Título de Revista/Publicación: Journal of Visual Languages and Computing
Fecha: Octubre 2008
ISSN: 1045-926X
Volumen: 19
Número: 5
Materias:
ODS:
Palabras Clave Informales: Animation, animated agents, lifelike characters, Spanish Sign Language, speech to gesture translation, gesture complexity
Escuela: E.T.S.I. Telecomunicación (UPM)
Departamento: Ingeniería Electrónica
Licencias Creative Commons: Reconocimiento - Sin obra derivada - No comercial

Texto completo

[thumbnail of INVE_MEM_2008_52867.pdf]
Vista Previa
PDF (Portable Document Format) - Se necesita un visor de ficheros PDF, como GSview, Xpdf o Adobe Acrobat Reader
Descargar (1MB) | Vista Previa

Resumen

This article describes an architecture for translating speech into Spanish Sign Language (SSL). The architecture proposed is made up of four modules: speech recognizer, semantic analysis, gesture sequence generation and gesture playing. For the speech recognizer and the semantic analysis modules, we use software developed by IBM and CSLR (Center for Spoken Language Research at University of Colorado), respectively. Gesture sequence generation and gesture animation are the modules on which we have focused our main effort. Gesture sequence generation uses semantic concepts (obtained from the semantic analysis) associating them with several SSL gestures. This association is carried out based on a number of generation rules. For gesture animation, we have developed an animated agent (virtual representation of a human person) and a strategy for reducing the effort in gesture animation. This strategy consists of making the system automatically generate all agent positions necessary for the gesture animation. In this process, the system uses a few main agent positions (two or three per second) and some interpolation strategies, both issues previously generated by the service developer (the person who adapts the architecture proposed in this paper to a specific domain). Related to this module, we propose a distance between agent positions and a measure of gesture complexity. This measure can be used to analyze the gesture perception versus its complexity. With the architecture proposed, we are not trying to build a domain independent translator but a system able to translate speech utterances into gesture sequences in a restricted domain: railway, flights or weather information.

Más información

ID de Registro: 2140
Identificador DC: https://oa.upm.es/2140/
Identificador OAI: oai:oa.upm.es:2140
URL Portal Científico: https://portalcientifico.upm.es/es/ipublic/item/5483282
Identificador DOI: 10.1016/j.jvlc.2007.06.002
URL Oficial: http://www.sciencedirect.com/science/journal/10459...
Depositado por: Memoria Investigacion
Depositado el: 02 Feb 2010 11:01
Ultima Modificación: 12 Nov 2025 00:00