This is the legacy GPT-NeoX relying on old DeeperSpeed (0.3.15). We only recommend using this release under circumstance that you're loading a model based on old DeeperSpeed (e.g. GPT-J, GPT-NeoX20B, the Pythia suite, etc). The primary difference between this release and v2.x is the DeepSpeed version supported. If you're using 2.x, we're assuming that you're using either the latest release of DeepSpeed or DeeperSpeed 2.x.If you use this software, please cite it using these metadata
Pretrained models corresponding to Action-GPT publication. For more details refer https://actiongpt....
These are datasets for the paper: "A Systematic Evaluation of Large Language Models of Code" https...
Floating point 32 bits precision weights for the Resnet-50 v1.5 PyTorch deep learning model. This f...
With GPT-NeoX 2.0, we now support upstream DeepSpeed. This enables the use of new DeepSpeed features...
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library
GPT-Neo is an implementation of model & data-parallel GPT-2 and GPT-3-like models, utilizing Mesh Te...
This is the model used to process the data in the paper Pix2Prof: fast extraction of sequential info...
OpenAI’s continuous efforts to push our knowledge and build on it for natural language processing ha...
This version fixes annotation errors in the AI2D-RST dataset, and updates the PyTorch DataLoader and...
This collection contains all the pre-trained model weights used to produce the results in our paper ...
deepStats is a statistical and dataviz toolbox for deeptools, genomic signals, and more
In this repository we release (yet another) GPT-2 model, that was trained on various texts for Germa...
A flexible package for multimodal-deep-learning to combine tabular data with text and images using W...
Keras GPT Copilot is the first Python package designed to integrate an LLM copilot within the model ...
This repository contains a pre-processed dataset derived from the GNPS public repository of natural ...
Pretrained models corresponding to Action-GPT publication. For more details refer https://actiongpt....
These are datasets for the paper: "A Systematic Evaluation of Large Language Models of Code" https...
Floating point 32 bits precision weights for the Resnet-50 v1.5 PyTorch deep learning model. This f...
With GPT-NeoX 2.0, we now support upstream DeepSpeed. This enables the use of new DeepSpeed features...
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library
GPT-Neo is an implementation of model & data-parallel GPT-2 and GPT-3-like models, utilizing Mesh Te...
This is the model used to process the data in the paper Pix2Prof: fast extraction of sequential info...
OpenAI’s continuous efforts to push our knowledge and build on it for natural language processing ha...
This version fixes annotation errors in the AI2D-RST dataset, and updates the PyTorch DataLoader and...
This collection contains all the pre-trained model weights used to produce the results in our paper ...
deepStats is a statistical and dataviz toolbox for deeptools, genomic signals, and more
In this repository we release (yet another) GPT-2 model, that was trained on various texts for Germa...
A flexible package for multimodal-deep-learning to combine tabular data with text and images using W...
Keras GPT Copilot is the first Python package designed to integrate an LLM copilot within the model ...
This repository contains a pre-processed dataset derived from the GNPS public repository of natural ...
Pretrained models corresponding to Action-GPT publication. For more details refer https://actiongpt....
These are datasets for the paper: "A Systematic Evaluation of Large Language Models of Code" https...
Floating point 32 bits precision weights for the Resnet-50 v1.5 PyTorch deep learning model. This f...