Skip to main content
eScholarship
Open Access Publications from the University of California

Comparing unsupervised speech learning directly to human performance inspeech perception

Abstract

We compare the performance of humans (English and Frenchlisteners) versus an unsupervised speech model in a perceptionexperiment (ABX discrimination task). Although the ABXtask has been used for acoustic model evaluation in previousresearch, the results have not, until now, been compared di-rectly with human behaviour in an experiment. We show that astandard, well-performing model (DPGMM) has better accu-racy at predicting human responses than the acoustic baseline.The model also shows a native language effect, better resem-bling native listeners of the language on which it was trained.However, the native language effect shown by the models isdifferent than the one shown by the human listeners, and, no-tably, the models do not show the same overall patterns ofvowel confusions.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View