Publicación: Multimodal Data Fusion of Electromyography and Acoustic Signals for Thai Syllable Recognition
dc.contributor.author | Sae Jong, Nida | |
dc.contributor.author | García Seco de Herrera, Alba | |
dc.date.accessioned | 2025-03-27T08:56:38Z | |
dc.date.available | 2025-03-27T08:56:38Z | |
dc.date.issued | 2020-10-27 | |
dc.description | Esta es la versión aceptada del artículo. La versión registrada fue publicada por primera vez en IEEE Journal of Biomedical and Health Informatics, vol. 25, no. 6, pp. 1997-2006, está disponible en línea en el sitio web del editor: IEEE Xplore, https://doi.org/10.1109/JBHI.2020.3034158. This is the accepted version of the article. The registered version was first published in IEEE Journal of Biomedical and Health Informatics, vol. 25, no. 6, pp. 1997-2006, is available online at the publisher's website: IEEE Xplore, https://doi.org/10.1109/JBHI.2020.3034158. | |
dc.description.abstract | Speech disorders such as dysarthria are common and frequent after suffering a stroke. Speech rehabilitation performed by a speech-language pathologist is needed to improve and recover. However, in Thailand, there is a shortage of speech-language pathologists. In this paper, we present a syllable recognition system, which can be deployable in a speech rehabilitation system to provide support to the limited speech-language pathologists available. The proposed system is based on a multimodal fusion of acoustic signal and surface electromyography (sEMG) collected from facial muscles. Multimodal data fusion is studied to improve signal collection under noisy situations while reducing the number of electrodes needed. The signals are simultaneously collected while articulating 12 Thai syllables designed for rehabilitation exercises. Several features are extracted from sEMG signals and five channels are studied. The best combination of features and channels is chosen to be fused with the mel-frequency cepstral coefficients extracted from the acoustic signal. The feature vector from each signal source is projected by spectral regression extreme learning machine and concatenated. Data from seven healthy subjects were collected for evaluation purposes. Results show that the multimodal fusion outperforms the use of a single signal source achieving up to ~ 98% of accuracy. In other words, an accuracy improvement up to 5% can be achieved when using the proposed multimodal fusion. Moreover, its low standard deviations in classification accuracy compared to those from the unimodal fusion indicate the improvement in the robustness of the syllable recognition. | en |
dc.description.version | versión final | |
dc.identifier.citation | N. S. Jong, A. G. S. de Herrera and P. Phukpattaranont, "Multimodal Data Fusion of Electromyography and Acoustic Signals for Thai Syllable Recognition," in IEEE Journal of Biomedical and Health Informatics, vol. 25, no. 6, pp. 1997-2006, June 2021, https://doi.org/10.1109/JBHI.2020.3034158. | |
dc.identifier.doi | https://doi.org/10.1109/JBHI.2020.3034158 | |
dc.identifier.issn | 2168-2194; eISSN: 2168-2208 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14468/26372 | |
dc.journal.issue | 6 | |
dc.journal.title | IEEE Journal of Biomedical and Health Informatics | |
dc.journal.volume | 25 | |
dc.language.iso | en | |
dc.page.final | 2006 | |
dc.page.initial | 1997 | |
dc.publisher | IEEE_ Institute of Electrical and Electronics Engineers | |
dc.relation.center | E.T.S. de Ingeniería Informática | |
dc.relation.department | Lenguajes y Sistemas Informáticos | |
dc.rights | info:eu-repo/semantics/openAccess | |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/4.0/deed.es | |
dc.subject | 12 Matemáticas::1203 Ciencia de los ordenadores ::1203.17 Informática | |
dc.subject.keywords | Acoustic signal | en |
dc.subject.keywords | electromyography | en |
dc.subject.keywords | feature-level fusion | en |
dc.subject.keywords | multimodal fusion | en |
dc.subject.keywords | speech recognition | en |
dc.title | Multimodal Data Fusion of Electromyography and Acoustic Signals for Thai Syllable Recognition | en |
dc.type | artículo | es |
dc.type | journal article | en |
dspace.entity.type | Publication | |
relation.isAuthorOfPublication | 33e1cf81-6a46-4cc6-828f-1c0f2a7e7497 | |
relation.isAuthorOfPublication.latestForDiscovery | 33e1cf81-6a46-4cc6-828f-1c0f2a7e7497 |
Archivos
Bloque original
1 - 1 de 1
Cargando...
- Nombre:
- GarciaSecoDeHerrera_Alba_ThaiSyllableRecognit.pdf
- Tamaño:
- 977.7 KB
- Formato:
- Adobe Portable Document Format
Bloque de licencias
1 - 1 de 1
No hay miniatura disponible
- Nombre:
- license.txt
- Tamaño:
- 3.62 KB
- Formato:
- Item-specific license agreed to upon submission
- Descripción: