Neural models are known to be over-parameterized, and recent work has shown that sparse text-to-speech (TTS) models can outperform dense models. Although a plethora of sparse methods has been proposed for other domains, such methods have rarely been applied in TTS. In this work, we seek to answer the question: what are the characteristics of selected sparse techniques on the performance and model complexity? We compare a Tacotron2 baseline and the results of applying five techniques. We then evaluate the performance via the factors of naturalness, intelligibility and prosody, while reporting model size and training time. Complementary to prior research, we find that pruning before or during training can achieve similar performance to prunin...
Distilling state-of-the-art transformer models into lightweight student models is an effective way t...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...
Neuroscience suggests that the sparse behavior of a neural population underlies the mechanisms of th...
Sparsity has become one of the promising methods to compress and accelerate Deep Neural Networks (DN...
The modern paradigm in speech processing has demonstrated the importance of scale and compute for en...
With the dramatically increased number of parameters in language models, sparsity methods have recei...
For personalized speech generation, a neural text-to-speech (TTS) model must be successfully impleme...
The training of sparse neural networks is becoming an increasingly important tool for reducing the ...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
State-of-the-art text-to-speech (TTS) systems have utilized pretrained language models (PLMs) to enh...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
We consider the problem of accurate sparse fine-tuning of large language models (LLMs), that is, fin...
Large Language Models have become the core architecture upon which most modern natural language proc...
Model sparsification in deep learning promotes simpler, more interpretable models with fewer paramet...
Distilling state-of-the-art transformer models into lightweight student models is an effective way t...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...
Neuroscience suggests that the sparse behavior of a neural population underlies the mechanisms of th...
Sparsity has become one of the promising methods to compress and accelerate Deep Neural Networks (DN...
The modern paradigm in speech processing has demonstrated the importance of scale and compute for en...
With the dramatically increased number of parameters in language models, sparsity methods have recei...
For personalized speech generation, a neural text-to-speech (TTS) model must be successfully impleme...
The training of sparse neural networks is becoming an increasingly important tool for reducing the ...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
State-of-the-art text-to-speech (TTS) systems have utilized pretrained language models (PLMs) to enh...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
We consider the problem of accurate sparse fine-tuning of large language models (LLMs), that is, fin...
Large Language Models have become the core architecture upon which most modern natural language proc...
Model sparsification in deep learning promotes simpler, more interpretable models with fewer paramet...
Distilling state-of-the-art transformer models into lightweight student models is an effective way t...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...
Neuroscience suggests that the sparse behavior of a neural population underlies the mechanisms of th...