Overview of the 2018 Spoken CALL Shared Task

Claudia Baur, Andrew Caines, Cathy Chua, Johanna Gerlach, Mengjie Qian, Manny Rayner, Martin Russell, Helmer Strik, Xizi Wei

We present an overview of the second edition of the Spoken CALL Shared Task. Groups competed on a prompt-response task using English-language data collected, through an online CALL game, from Swiss German teens in their second and third years of learning English. Each item consists of a written German prompt and an audio file containing a spoken response. The task is to accept linguistically correct responses and reject linguistically incorrect ones, with “linguistically correct” defined by a gold standard derived from human annotations. Scoring was performed using a metric defined as the ratio of the relative rejection rates on incorrect and correct responses. The second edition received eighteen entries and showed very substantial improvement on the first edition; all entries were better than the best entry from the first edition and the best score was about four times higher. We present the task, the resources, the results, a discussion of the metrics used and an analysis of what makes items challenging. In particular, we present quantitative evidence suggesting that incorrect responses are much more difficult to process than correct responses and that the most significant factor in making a response challenging is its distance from the closest training example.

 DOI: 10.21437/Interspeech.2018-97

Cite as: Baur, C., Caines, A., Chua, C., Gerlach, J., Qian, M., Rayner, M., Russell, M., Strik, H., Wei, X. (2018) Overview of the 2018 Spoken CALL Shared Task. Proc. Interspeech 2018, 2354-2358, DOI: 10.21437/Interspeech.2018-97.

  author={Claudia Baur and Andrew Caines and Cathy Chua and Johanna Gerlach and Mengjie Qian and Manny Rayner and Martin Russell and Helmer Strik and Xizi Wei},
  title={Overview of the 2018 Spoken CALL Shared Task},
  booktitle={Proc. Interspeech 2018},