Submitted to Interspeech 2021. arXiv admin note: text overlap with arXiv:2011.11588International audienceWe present the Zero Resource Speech Challenge 2021, which asks participants to learn a language model directly from audio, without any text or labels. The challenge is based on the Libri-light dataset, which provides up to 60k hours of audio from English audio books without any associated text. We provide a pipeline baseline system consisting on an encoder based on contrastive predictive coding (CPC), a quantizer ($k$-means) and a standard language model (BERT or LSTM). The metrics evaluate the learned representations at the acoustic (ABX discrimination), lexical (spot-the-word), syntactic (acceptability judgment) and semantic levels (si...
Most recent speech recognition models rely on large supervised datasets, which are unavailable for m...
Automatic speech recognition for our most widely used languages has recently seen substantial impro...
International audienceRecent work in spoken language modeling shows the possibility of learning a la...
Submitted to Interspeech 2021. arXiv admin note: text overlap with arXiv:2011.11588International aud...
14 pages, including references and supplementary materialInternational audienceWe introduce a new un...
International audienceWe introduce Generative Spoken Language Modeling, the task of learning the aco...
International audienceRecent progress in self-supervised or unsupervised machine learning has opened...
AbstractThis paper reports on the results of the Zero Resource Speech Challenge 2015, the first unif...
International audienceWe present the Zero Resource Speech Challenge 2019, which proposes to build a ...
International audienceWe introduce a new collection of spoken English audio suitable for training sp...
We introduce a new collection of spoken English audio suitable for training speech recognition syste...
The Interspeech 2015 Zero Resource Speech Challenge aims at discovering subword and word units from ...
International audienceWe present the Zero Resource Speech Challenge 2020, which aims at learning spe...
We summarize the accomplishments of a multi-disciplinary work-shop exploring the computational and s...
International audienceRecent work on unsupervised contrastive learning of speech representation has ...
Most recent speech recognition models rely on large supervised datasets, which are unavailable for m...
Automatic speech recognition for our most widely used languages has recently seen substantial impro...
International audienceRecent work in spoken language modeling shows the possibility of learning a la...
Submitted to Interspeech 2021. arXiv admin note: text overlap with arXiv:2011.11588International aud...
14 pages, including references and supplementary materialInternational audienceWe introduce a new un...
International audienceWe introduce Generative Spoken Language Modeling, the task of learning the aco...
International audienceRecent progress in self-supervised or unsupervised machine learning has opened...
AbstractThis paper reports on the results of the Zero Resource Speech Challenge 2015, the first unif...
International audienceWe present the Zero Resource Speech Challenge 2019, which proposes to build a ...
International audienceWe introduce a new collection of spoken English audio suitable for training sp...
We introduce a new collection of spoken English audio suitable for training speech recognition syste...
The Interspeech 2015 Zero Resource Speech Challenge aims at discovering subword and word units from ...
International audienceWe present the Zero Resource Speech Challenge 2020, which aims at learning spe...
We summarize the accomplishments of a multi-disciplinary work-shop exploring the computational and s...
International audienceRecent work on unsupervised contrastive learning of speech representation has ...
Most recent speech recognition models rely on large supervised datasets, which are unavailable for m...
Automatic speech recognition for our most widely used languages has recently seen substantial impro...
International audienceRecent work in spoken language modeling shows the possibility of learning a la...