Marian (@sshleifer) A new model architecture, MarianMTModel with 1,008+ pretrained weights is available for machine translation in PyTorch. The corresponding MarianTokenizer uses a prepare_translation_batch method to prepare model inputs. All pretrained model names use the following format: Helsinki-NLP/opus-mt-{src}-{tgt} See docs for information on pretrained model discovery and naming, or find your language here AlbertForPreTraining (@jarednielsen) A new model architecture has been added: AlbertForPreTraining in both PyTorch and TensorFlow TF 2.2 compatibility (@mfuntowicz, @jplu) Changes have been made to both the TensorFlow scripts and our internals so that we are compatible with TensorFlow 2.2 TFTrainer now supports new tasks Mult...
New class Pipeline (beta): easily run and use models on down-stream NLP tasks We have added a new cl...
T5 Model (@patrickvonplaten, @thomwolf ) T5 is a powerful encoder-decoder model that formats every N...
Better backward-compatibility for tokenizers following v3.0.0 refactoring Version v3.0.0, included a...
Name change: welcome Transformers Following the extension to TensorFlow 2.0, pytorch-transformers =...
New model architectures: ALBERT, CamemBERT, GPT2-XL, DistilRoberta Four new models have been added i...
Trainer & TFTrainer Version 2.9 introduces a new Trainer class for PyTorch, and its equivalent TFTra...
Transformers: State-of-the-art Natural Language Processing for TensorFlow 2.0 and PyTorch
New model architectures: CTRL, DistilGPT-2 Two new models have been added since release 2.0. CTRL (...
New model architecture: DistilBERT Adding Huggingface's new transformer architecture, DistilBERT des...
FlauBERT, MMBT MMBT was added to the list of available models, as the first multi-modal model to ma...
New Model: BART (added by @sshleifer) Bart is one of the first Seq2Seq models in the library, and ac...
Transformers: State-of-the-art Natural Language Processing for Pytorch, TensorFlow, and JAX
Longformer Longformer (@ibeltagy) Longformer for QA (@patil-suraj + @patrickvonplaten) Longformer f...
Added network.py module (see "Changed") initial tests/test_tweetynet.py module, written for pytest ...
Rust tokenizers (@mfuntowicz, @n1t0 ) Tokenizers for Bert, Roberta, OpenAI GPT, OpenAI GPT2, Transf...
New class Pipeline (beta): easily run and use models on down-stream NLP tasks We have added a new cl...
T5 Model (@patrickvonplaten, @thomwolf ) T5 is a powerful encoder-decoder model that formats every N...
Better backward-compatibility for tokenizers following v3.0.0 refactoring Version v3.0.0, included a...
Name change: welcome Transformers Following the extension to TensorFlow 2.0, pytorch-transformers =...
New model architectures: ALBERT, CamemBERT, GPT2-XL, DistilRoberta Four new models have been added i...
Trainer & TFTrainer Version 2.9 introduces a new Trainer class for PyTorch, and its equivalent TFTra...
Transformers: State-of-the-art Natural Language Processing for TensorFlow 2.0 and PyTorch
New model architectures: CTRL, DistilGPT-2 Two new models have been added since release 2.0. CTRL (...
New model architecture: DistilBERT Adding Huggingface's new transformer architecture, DistilBERT des...
FlauBERT, MMBT MMBT was added to the list of available models, as the first multi-modal model to ma...
New Model: BART (added by @sshleifer) Bart is one of the first Seq2Seq models in the library, and ac...
Transformers: State-of-the-art Natural Language Processing for Pytorch, TensorFlow, and JAX
Longformer Longformer (@ibeltagy) Longformer for QA (@patil-suraj + @patrickvonplaten) Longformer f...
Added network.py module (see "Changed") initial tests/test_tweetynet.py module, written for pytest ...
Rust tokenizers (@mfuntowicz, @n1t0 ) Tokenizers for Bert, Roberta, OpenAI GPT, OpenAI GPT2, Transf...
New class Pipeline (beta): easily run and use models on down-stream NLP tasks We have added a new cl...
T5 Model (@patrickvonplaten, @thomwolf ) T5 is a powerful encoder-decoder model that formats every N...
Better backward-compatibility for tokenizers following v3.0.0 refactoring Version v3.0.0, included a...