Instruction tuning is instrumental in enabling Large Language Models~(LLMs) to follow user instructions to complete various open-domain tasks. The success of instruction tuning depends on the availability of high-quality instruction data. Owing to the exorbitant cost and substandard quality of human annotation, recent works have been deeply engaged in the exploration of the utilization of powerful closed-source models to generate instruction data automatically. However, these methods carry potential risks arising from the usage requirements of powerful closed-source models, which strictly forbid the utilization of their outputs to develop machine learning models. To deal with this problem, in this work, we explore alternative approaches to ...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Visual instruction tuning is an essential approach to improving the zero-shot generalization capabil...
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks de...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Recent advancements in Large Language Models (LLMs) have expanded the horizons of natural language u...
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension a...
Large language models~(LLMs) are instruction followers, but it can be challenging to find the best i...
Instruction-tuning has become an integral part of training pipelines for Large Language Models (LLMs...
The practice of transferring knowledge from a sophisticated, proprietary large language model (LLM) ...
In the swiftly expanding domain of Natural Language Processing (NLP), the potential of GPT-based mod...
Instruction tuning has emerged as a promising approach to enhancing large language models in followi...
Sparse Mixture-of-Experts (MoE) is a neural architecture design that can be utilized to add learnabl...
To learn text understanding models with millions of parameters one needs massive amounts of data. In...
The integration of visual encoders and large language models (LLMs) has driven recent progress in mu...
Recently, Instruction fine-tuning has risen to prominence as a potential method for enhancing the ze...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Visual instruction tuning is an essential approach to improving the zero-shot generalization capabil...
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks de...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Recent advancements in Large Language Models (LLMs) have expanded the horizons of natural language u...
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension a...
Large language models~(LLMs) are instruction followers, but it can be challenging to find the best i...
Instruction-tuning has become an integral part of training pipelines for Large Language Models (LLMs...
The practice of transferring knowledge from a sophisticated, proprietary large language model (LLM) ...
In the swiftly expanding domain of Natural Language Processing (NLP), the potential of GPT-based mod...
Instruction tuning has emerged as a promising approach to enhancing large language models in followi...
Sparse Mixture-of-Experts (MoE) is a neural architecture design that can be utilized to add learnabl...
To learn text understanding models with millions of parameters one needs massive amounts of data. In...
The integration of visual encoders and large language models (LLMs) has driven recent progress in mu...
Recently, Instruction fine-tuning has risen to prominence as a potential method for enhancing the ze...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Visual instruction tuning is an essential approach to improving the zero-shot generalization capabil...
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks de...