Deep acoustic models represent linguistic information based on massive amounts of data. Unfortunately, for regional languages and dialects such resources are mostly not available. However, deep acoustic models might have learned linguistic information that transfers to low-resource languages. In this study, we evaluate whether this is the case through the task of distinguishing low-resource (Dutch) regional varieties. By extracting embeddings from the hidden layers of various wav2vec 2.0 models (including new models which are pre-trained and/or fine-tuned on Dutch) and using dynamic time warping, we compute pairwise pronunciation differences averaged over 10 words for over 100 individual dialects from four (regional) languages. We then clus...
How do Dutch and Korean listeners use acoustic–phonetic information when learning words in an artifi...
In this paper, we propose two methods for automatically obtaining hypotheses about pronunciation va...
Despite the success of deep learning in speech recognition, multi-dialect speech recognition remains...
Deep acoustic models represent linguistic information based on massive amounts of data. Unfortunatel...
Languages are fundamental to human communication and serve as a means to express social and cultural...
Under-resourced speech recognizers may benefit from data in languages other than the target language...
For purposes of automated speech recognition in under-resourced environments, techniques used to sha...
Contemporary Dutch dialects are compared using the most recent Dutch dialect source available: the G...
Over the past decades, speech recognition has dramatically improved in a large variety of applicatio...
The development of a speech recognition system requires at least three resources: a large labeled sp...
Recent research on the TIMIT corpus suggests that longer-length acoustic models are more appropriat...
In this paper, we propose two methods for automatically obtaining hypotheses about pronunciation var...
In this paper we present an experiment aimed at improving automatic phonetic transcription of Dutch ...
In this paper we present an experiment aimed at improving automatic phonetic transcription of Dutch ...
Structuralists famously observed that language is "un systeme oil tout se tient" (Meillet, 1903, p.4...
How do Dutch and Korean listeners use acoustic–phonetic information when learning words in an artifi...
In this paper, we propose two methods for automatically obtaining hypotheses about pronunciation va...
Despite the success of deep learning in speech recognition, multi-dialect speech recognition remains...
Deep acoustic models represent linguistic information based on massive amounts of data. Unfortunatel...
Languages are fundamental to human communication and serve as a means to express social and cultural...
Under-resourced speech recognizers may benefit from data in languages other than the target language...
For purposes of automated speech recognition in under-resourced environments, techniques used to sha...
Contemporary Dutch dialects are compared using the most recent Dutch dialect source available: the G...
Over the past decades, speech recognition has dramatically improved in a large variety of applicatio...
The development of a speech recognition system requires at least three resources: a large labeled sp...
Recent research on the TIMIT corpus suggests that longer-length acoustic models are more appropriat...
In this paper, we propose two methods for automatically obtaining hypotheses about pronunciation var...
In this paper we present an experiment aimed at improving automatic phonetic transcription of Dutch ...
In this paper we present an experiment aimed at improving automatic phonetic transcription of Dutch ...
Structuralists famously observed that language is "un systeme oil tout se tient" (Meillet, 1903, p.4...
How do Dutch and Korean listeners use acoustic–phonetic information when learning words in an artifi...
In this paper, we propose two methods for automatically obtaining hypotheses about pronunciation va...
Despite the success of deep learning in speech recognition, multi-dialect speech recognition remains...