New class Pipeline (beta): easily run and use models on down-stream NLP tasks We have added a new class called Pipeline to simply run and use models for several down-stream NLP tasks. A Pipeline is just a tokenizer + model wrapped so they can take human-readable inputs and output human-readable results. The Pipeline will take care of : tokenizing inputs strings => convert in tensors => run in the model => post-process output Currently, we have added the following pipelines with a default model for each: feature extraction (can be used with any pretrained and finetuned models) inputs: strings/list of strings – output: list of floats (last hidden-states of the model for each token) sentiment classification (DistilBert model fine-tuned on SST...
Common Crawl is a considerably large, heterogeneous multilingual corpus comprised of crawled documen...
Perceiver The Perceiver model was released in the previous version: Perceiver Eight new models are r...
Longformer Longformer (@ibeltagy) Longformer for QA (@patil-suraj + @patrickvonplaten) Longformer f...
New Model: BART (added by @sshleifer) Bart is one of the first Seq2Seq models in the library, and ac...
T5 Model (@patrickvonplaten, @thomwolf ) T5 is a powerful encoder-decoder model that formats every N...
New model architectures: ALBERT, CamemBERT, GPT2-XL, DistilRoberta Four new models have been added s...
New model architecture: DistilBERT Adding Huggingface's new transformer architecture, DistilBERT des...
FlauBERT, MMBT MMBT was added to the list of available models, as the first multi-modal model to ma...
The current modus operandi in NLP involves downloading and fine-tuning pre-trained models consisting...
In 2017, Vaswani et al. proposed a new neural network architecture named Transformer. That modern ar...
New model architectures: CTRL, DistilGPT-2 Two new models have been added since release 2.0. CTRL (...
Trainer & TFTrainer Version 2.9 introduces a new Trainer class for PyTorch, and its equivalent TFTra...
Natural language processing (NLP) involves the computer analysis and processing of human languages u...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Rust tokenizers (@mfuntowicz, @n1t0 ) Tokenizers for Bert, Roberta, OpenAI GPT, OpenAI GPT2, Transf...
Common Crawl is a considerably large, heterogeneous multilingual corpus comprised of crawled documen...
Perceiver The Perceiver model was released in the previous version: Perceiver Eight new models are r...
Longformer Longformer (@ibeltagy) Longformer for QA (@patil-suraj + @patrickvonplaten) Longformer f...
New Model: BART (added by @sshleifer) Bart is one of the first Seq2Seq models in the library, and ac...
T5 Model (@patrickvonplaten, @thomwolf ) T5 is a powerful encoder-decoder model that formats every N...
New model architectures: ALBERT, CamemBERT, GPT2-XL, DistilRoberta Four new models have been added s...
New model architecture: DistilBERT Adding Huggingface's new transformer architecture, DistilBERT des...
FlauBERT, MMBT MMBT was added to the list of available models, as the first multi-modal model to ma...
The current modus operandi in NLP involves downloading and fine-tuning pre-trained models consisting...
In 2017, Vaswani et al. proposed a new neural network architecture named Transformer. That modern ar...
New model architectures: CTRL, DistilGPT-2 Two new models have been added since release 2.0. CTRL (...
Trainer & TFTrainer Version 2.9 introduces a new Trainer class for PyTorch, and its equivalent TFTra...
Natural language processing (NLP) involves the computer analysis and processing of human languages u...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Rust tokenizers (@mfuntowicz, @n1t0 ) Tokenizers for Bert, Roberta, OpenAI GPT, OpenAI GPT2, Transf...
Common Crawl is a considerably large, heterogeneous multilingual corpus comprised of crawled documen...
Perceiver The Perceiver model was released in the previous version: Perceiver Eight new models are r...
Longformer Longformer (@ibeltagy) Longformer for QA (@patil-suraj + @patrickvonplaten) Longformer f...