Recent studies have demonstrated the great potential of Large Language Models (LLMs) serving as zero-shot relevance rankers. The typical approach involves making comparisons between pairs or lists of documents. Although effective, these listwise and pairwise methods are not efficient and also heavily rely on intricate prompt engineering. To tackle this problem, we introduce a novel instruction distillation method. The key idea is to distill the pairwise ranking ability of open-sourced LLMs to a simpler but more efficient pointwise ranking. Specifically, given the same LLM, we first rank documents using the effective pairwise approach with complex instructions, and then distill the teacher predictions to the pointwise approach with simpler i...
Making an informed choice of pre-trained language model (LM) is critical for performance, yet enviro...
Deep learning for Information Retrieval (IR) requires a large amount of high-quality query-document ...
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension a...
We consider language modelling (LM) as a multi-label structured prediction task by re-framing traini...
Deploying large language models (LLMs) is challenging because they are memory inefficient and comput...
We propose utilizing n-best reranking to enhance the Sequence-Level Knowledge Distillation (Kim and ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
Recent work has shown that small distilled language models are strong competitors to models that are...
Many information retrieval tasks require large labeled datasets for fine-tuning. However, such datas...
Knowledge Distillation (KD), which transfers the knowledge of a well-trained large model (teacher) t...
Recently, multi-modal content generation has attracted lots of attention from researchers by investi...
Pretrained large language models (LLMs) are widely used in many sub-fields of natural language proce...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Learning-to-Rank (LtR) techniques leverage machine learning algorithms and large amounts of training...
Making an informed choice of pre-trained language model (LM) is critical for performance, yet enviro...
Deep learning for Information Retrieval (IR) requires a large amount of high-quality query-document ...
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension a...
We consider language modelling (LM) as a multi-label structured prediction task by re-framing traini...
Deploying large language models (LLMs) is challenging because they are memory inefficient and comput...
We propose utilizing n-best reranking to enhance the Sequence-Level Knowledge Distillation (Kim and ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
Recent work has shown that small distilled language models are strong competitors to models that are...
Many information retrieval tasks require large labeled datasets for fine-tuning. However, such datas...
Knowledge Distillation (KD), which transfers the knowledge of a well-trained large model (teacher) t...
Recently, multi-modal content generation has attracted lots of attention from researchers by investi...
Pretrained large language models (LLMs) are widely used in many sub-fields of natural language proce...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Learning-to-Rank (LtR) techniques leverage machine learning algorithms and large amounts of training...
Making an informed choice of pre-trained language model (LM) is critical for performance, yet enviro...
Deep learning for Information Retrieval (IR) requires a large amount of high-quality query-document ...
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension a...