The increasing of pre-trained models has significantly facilitated the performance on limited data tasks with transfer learning. However, progress on transfer learning mainly focuses on optimizing the weights of pre-trained models, which ignores the structure mismatch between the model and the target task. This paper aims to improve the transfer performance from another angle - in addition to tuning the weights, we tune the structure of pre-trained models, in order to better match the target task. To this end, we propose TransTailor, targeting at pruning the pre-trained model for improved transfer learning. Different from traditional pruning pipelines, we prune and fine-tune the pre-trained model according to the target-aware weight importa...
We theoretically and empirically analyze the phenomenon of transfer learning in overparameterized ma...
We present three related ways of using Transfer Learning to improve feature selection. The three met...
Pre-training language models (LMs) on large-scale unlabeled text data makes the model much easier to...
Inductive transfer learning aims to learn from a small amount of training data for the target task b...
Large, pre-trained models are problematic to use in resource constrained applications. Fortunately, ...
Pretrained models could be reused in a way that allows for improvement in training accuracy. Trainin...
Traditional machine learning makes a basic assumption: the training and test data should be under th...
This empirical research study discusses how much the model’s accuracy changes when adding a new imag...
We propose a novel framework that unifies and extends existing methods of transfer learning (TL) for...
© 2020, Springer Nature Switzerland AG. We propose a novel adaptive transfer learning framework, lea...
Transfer learning algorithms are used when one has sufficient training data for one supervised learn...
Transfer-learning methods aim to improve performance in a data-scarce target domain using a model pr...
We study a fundamental transfer learning process from source to target linear regression tasks, incl...
Transfer learning is a successful technique that significantly improves machine learning algorithms ...
The traditional way of obtaining models from data, inductive learning, has proved itself both in the...
We theoretically and empirically analyze the phenomenon of transfer learning in overparameterized ma...
We present three related ways of using Transfer Learning to improve feature selection. The three met...
Pre-training language models (LMs) on large-scale unlabeled text data makes the model much easier to...
Inductive transfer learning aims to learn from a small amount of training data for the target task b...
Large, pre-trained models are problematic to use in resource constrained applications. Fortunately, ...
Pretrained models could be reused in a way that allows for improvement in training accuracy. Trainin...
Traditional machine learning makes a basic assumption: the training and test data should be under th...
This empirical research study discusses how much the model’s accuracy changes when adding a new imag...
We propose a novel framework that unifies and extends existing methods of transfer learning (TL) for...
© 2020, Springer Nature Switzerland AG. We propose a novel adaptive transfer learning framework, lea...
Transfer learning algorithms are used when one has sufficient training data for one supervised learn...
Transfer-learning methods aim to improve performance in a data-scarce target domain using a model pr...
We study a fundamental transfer learning process from source to target linear regression tasks, incl...
Transfer learning is a successful technique that significantly improves machine learning algorithms ...
The traditional way of obtaining models from data, inductive learning, has proved itself both in the...
We theoretically and empirically analyze the phenomenon of transfer learning in overparameterized ma...
We present three related ways of using Transfer Learning to improve feature selection. The three met...
Pre-training language models (LMs) on large-scale unlabeled text data makes the model much easier to...