Investigations on Speaking Mode Discrepancies in EMG-based Speech Recognition
by , ,
Abstract:
In this paper we present our recent study on the impact of speaking mode variabilities on speech recognition by surface electromyography (EMG). Surface electromyography captures the electric potentials of the human articulatory muscles, which enables a user to communicate naturally without making any audible sound. Our previous experiments have shown that the EMG signal varies greatly between different speaking modes, like audibly uttered speech and silently articulated speech. In this study we extend our previous research and quantify the impact of different speaking modes by investigating the amount of mode-specific leaves in phonetic decision trees. We show that this measure correlates highly with discrepancies in the spectral energy of the EMG signal, as well as with differences in the performance of a recognizer on different speaking modes. We furthermore present how EMG signal adaptation by spectral mapping decreases the effect of the speaking mode.
Reference:
Investigations on Speaking Mode Discrepancies in EMG-based Speech Recognition (Matthias Janke, Michael Wand, Tanja Schultz), In 12th Annual Conference of the International Speech Communication Association, 2011. (Interspeech 2011)
Bibtex Entry:
@inproceedings{janke2011investigations,
  title={Investigations on Speaking Mode Discrepancies in EMG-based Speech Recognition},
  year={2011},
  note={Interspeech 2011},
  booktitle={12th Annual Conference of the International Speech Communication Association},
  url={https://www.csl.uni-bremen.de/cms/images/documents/publications/WandJankeSchultz_Interspeech2011.pdf},
  abstract={In this paper we present our recent study on the impact of speaking mode variabilities on speech recognition by surface electromyography (EMG). Surface electromyography captures the electric potentials of the human articulatory muscles, which enables a user to communicate naturally without making any audible sound. Our previous experiments have shown that the EMG signal varies greatly between different speaking modes, like audibly uttered speech and silently articulated speech. In this study we extend our previous research and quantify the impact of different speaking modes by investigating the amount of mode-specific leaves in phonetic decision trees. We show that this measure correlates highly with discrepancies in the spectral energy of the EMG signal, as well as with differences in the performance of a recognizer on different speaking modes. We furthermore present how EMG signal adaptation by spectral mapping decreases the effect of the speaking mode.},
  keywords={EMG, EMG-based speech recognition, Silent
Speech Interfaces, phonetic decision tree},
  author={Janke, Matthias and Wand, Michael and Schultz, Tanja}
}