Analyzing Speech and Pronunciation Performance Virtually

Poster #7 ID: 52957 Type: Poster/Demo
  1. Todd Cooper, National Institute of Technology, Toyama College, Japan
  2. Nobuyuki Gomi and Yuki Miyashita, National Institute of Technology, Fukui College, Japan
  3. Akira Tsukada, National Institute of Technology, Toyama College, Japan

Tuesday, June 26 6:00 PM-7:00 PM

Presider:
Douglass Scott, Waseda University, Japan

For the past several years we have been creating a system using a motion-sensing camera. The IPA [Interview and Presentation Assistant] combined facial, speech, and gestural recognition and helped students get personalized feedback based on parameters set by human input. However, while we were doing this, voice recognition systems such as Google, Alexa and Siri started to get really, really good. Because of this, we changed the direction. We are currently developing a system to evaluate speech performance of EFL students in Japan which focuses on the pronunciation of individual sounds, word/sentence stress, and rhythm/intonation. Teachers input a score from an audio clip recorded by the student. This scores are then fed into the system to create an algorithm which in turn can reproduce the teacher’s score to a certain degree without the teacher being present. This research is in its initial stage, but the results show the possibility to create a system which will help students improve their English speaking ability. This will help provide the important one-to-one personalized feedback they need to do so.

Topics

Conference attendees are able to comment on papers, view the full text and slides, and attend live presentations. If you are an attendee, please login to get full access.