Utilize este identificador para referenciar este registo: http://hdl.handle.net/10071/25552
Autoria: Freitas, J.
Teixeira, A.
Dias, J.
Editor: Nicoletta Calzolari, Khalid Choukri, Thierry Declerck, Hrafn Loftsson, Bente Maegaard, Joseph Mariani, Asuncion Moreno, Jan Odijk and Stelios Piperidis
Data: 2014
Título próprio: Multimodal corpora for silent speech interaction
Título e volume do livro: Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC 2014)
Paginação: 4507 - 4511
Título do evento: 9th International Conference on Language Resources and Evaluation, LREC 2014
Referência bibliográfica: Freitas, J., Teixeira, A., & Dias, J. (2014). Multimodal corpora for silent speech interaction. Em N. Calzolari, K. Choukri, T. Declerck, H. Loftsson, B. Maegaard, J. Mariani, A. Moreno, J. Odijk, & S. Piperidis (Eds.), Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC 2014) (pp. 4507-4511). European Language Resources Association (ELRA)
ISBN: 978-2-9517408-8-4
Palavras-chave: Silent speech
Multimodal HCI
Data collection
Resumo: A Silent Speech Interface (SSI) allows for speech communication to take place in the absence of an acoustic signal. This type of interface is an alternative to conventional Automatic Speech Recognition which is not adequate for users with some speech impairments or in the presence of environmental noise. The work presented here produces the conditions to explore and analyze complex combinations of input modalities applicable in SSI research. By exploring non-invasive and promising modalities, we have selected the following sensing technologies used in human-computer interaction: Video and Depth input, Ultrasonic Doppler sensing and Surface Electromyography. This paper describes a novel data collection methodology where these independent streams of information are synchronously acquired with the aim of supporting research and development of a multimodal SSI. The reported recordings were divided into two rounds: a first one where the acquired data was silently uttered and a second round where speakers pronounced the scripted prompts in an audible and normal tone. In the first round of recordings, a total of 53.94 minutes were captured where 30.25% was estimated to be silent speech. In the second round of recordings, a total of 30.45 minutes were obtained and 30.05% of the recordings were audible speech.
Arbitragem científica: yes
Acesso: Acesso Aberto
Aparece nas coleções:ISTAR-CRI - Comunicações a conferências internacionais

Ficheiros deste registo:
Ficheiro Descrição TamanhoFormato 
conferenceobject_71799.pdfVersão Editora670,81 kBAdobe PDFVer/Abrir


FacebookTwitterDeliciousLinkedInDiggGoogle BookmarksMySpaceOrkut
Formato BibTex mendeley Endnote Logotipo do DeGóis Logotipo do Orcid 

Todos os registos no repositório estão protegidos por leis de copyright, com todos os direitos reservados.