While Automatic Speech Recognition (ASR) models have shown significant advances with the introduction of unsupervised or self-supervised training techniques, these improvements are still only limited to a subsection of languages and speakers. Transfer learning enables the adaptation of large-scale multilingual models to not only low-resource languages but also to more specific speaker groups. However, fine-tuning on data from new domains is usually accompanied by a decrease in performance on the original domain. Therefore, in our experiments, we examine how well the performance of large-scale ASR models can be approximated for smaller domains, with our own dataset of German Senior Voice Commands (SVC-de), and how much of the general speech ...
Automatic speech recognition research focuses on training and evaluating on static datasets. Yet, as...
The cross-domain performance of automatic speech recognition (ASR) could be severely hampered due to...
Continual Learning, also known as Lifelong Learning, aims to continually learn from new data as it b...
Adapting a trained Automatic Speech Recognition (ASR) model to new tasks results in catastrophic for...
Automatic speech recognition models are often adapted to improve their accuracy in a new domain. A p...
Adapting Automatic Speech Recognition (ASR) models to new domains results in a deterioration of perf...
Advances in self-supervised learning have significantly reduced the amount of transcribed audio requ...
We introduce the Universal Speech Model (USM), a single large model that performs automatic speech r...
Learning a set of tasks in sequence remains a challenge for artificial neural networks, which, in su...
Self-supervised representation learning (SSRL) has improved the performance on downstream phoneme re...
Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications ...
The performance of the speech recognition systems to translate voice to text is still an issue in la...
In this paper, we investigate the usage of large language models (LLMs) to improve the performance o...
We summarize the results of a host of efforts using giant automatic speech recognition (ASR) models ...
Recent work on self-supervised pre-training focus on leveraging large-scale unlabeled speech data to...
Automatic speech recognition research focuses on training and evaluating on static datasets. Yet, as...
The cross-domain performance of automatic speech recognition (ASR) could be severely hampered due to...
Continual Learning, also known as Lifelong Learning, aims to continually learn from new data as it b...
Adapting a trained Automatic Speech Recognition (ASR) model to new tasks results in catastrophic for...
Automatic speech recognition models are often adapted to improve their accuracy in a new domain. A p...
Adapting Automatic Speech Recognition (ASR) models to new domains results in a deterioration of perf...
Advances in self-supervised learning have significantly reduced the amount of transcribed audio requ...
We introduce the Universal Speech Model (USM), a single large model that performs automatic speech r...
Learning a set of tasks in sequence remains a challenge for artificial neural networks, which, in su...
Self-supervised representation learning (SSRL) has improved the performance on downstream phoneme re...
Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications ...
The performance of the speech recognition systems to translate voice to text is still an issue in la...
In this paper, we investigate the usage of large language models (LLMs) to improve the performance o...
We summarize the results of a host of efforts using giant automatic speech recognition (ASR) models ...
Recent work on self-supervised pre-training focus on leveraging large-scale unlabeled speech data to...
Automatic speech recognition research focuses on training and evaluating on static datasets. Yet, as...
The cross-domain performance of automatic speech recognition (ASR) could be severely hampered due to...
Continual Learning, also known as Lifelong Learning, aims to continually learn from new data as it b...