TY - JOUR
T1 - Multimodal analysis of public speaking performance by EFL learners
T2 - Applying deep learning to understanding how successful speakers use facial movement
AU - Fuyuno, Miharu
AU - Komiya, Rinko
AU - Saitoh, Takeshi
N1 - Funding Information:
This work was supported by the Japan Society for the Promotion of Science Kakenhi Grant numbers 15K12416 and 16H03079. The authors thank the teachers and participants at the high school for access to the research data, particularly Mr. Kazuhide Shinohara for his continued support and involvement in the research.
Publisher Copyright:
© 2018, The Asian Journal of Applied Linguistics.
PY - 2018
Y1 - 2018
N2 - Although multimodal corpus analysis has been widely practiced in the field of applied linguistics, few studies have investigated performance of English public speaking by EFL learners. Needs for effective public speaking are fundamental in the globalizing society; however, performing public speaking in English is challenging for EFL learners, and objective analysis on factors of eye contact and speech pauses still remain few though such information is crucial in efficient teaching. This study analyses public speaking performance by EFL learners based on data from a multimodal corpus. Data were collected in an annual speech contest at a Japanese high school. Speakers presented English speeches to an audience and judges. The data consist of video and digital audio recordings of performance, as well as speech scripts and evaluation scores by contest judges. Characteristics of speakers' facial movement patterns in regard to spoken contents and the correlation between facial movements and eye movements were examined. Facial and eye movements were detected with motion tracking and the deep learning method. The results indicated that facial direction changes were not synchronized with speech pauses among highly evaluated speakers. Furthermore, the facial direction changes tended to be synchronized with content words in the spoken utterance rather than function words.
AB - Although multimodal corpus analysis has been widely practiced in the field of applied linguistics, few studies have investigated performance of English public speaking by EFL learners. Needs for effective public speaking are fundamental in the globalizing society; however, performing public speaking in English is challenging for EFL learners, and objective analysis on factors of eye contact and speech pauses still remain few though such information is crucial in efficient teaching. This study analyses public speaking performance by EFL learners based on data from a multimodal corpus. Data were collected in an annual speech contest at a Japanese high school. Speakers presented English speeches to an audience and judges. The data consist of video and digital audio recordings of performance, as well as speech scripts and evaluation scores by contest judges. Characteristics of speakers' facial movement patterns in regard to spoken contents and the correlation between facial movements and eye movements were examined. Facial and eye movements were detected with motion tracking and the deep learning method. The results indicated that facial direction changes were not synchronized with speech pauses among highly evaluated speakers. Furthermore, the facial direction changes tended to be synchronized with content words in the spoken utterance rather than function words.
UR - http://www.scopus.com/inward/record.url?scp=85047367965&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85047367965&partnerID=8YFLogxK
M3 - Article
AN - SCOPUS:85047367965
SN - 2308-6262
VL - 5
SP - 117
EP - 129
JO - Asian Journal of Applied Linguistics
JF - Asian Journal of Applied Linguistics
IS - 1
ER -