HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Distilling Knowledge from Reader to Retriever for Question Answering

Gautier Izacard Edouard Grave

Distilling Knowledge from Reader to Retriever for Question Answering

Abstract

The task of information retrieval is an important component of many natural language processing systems, such as open domain question answering. While traditional methods were based on hand-crafted features, continuous representations based on neural networks recently obtained competitive results. A challenge of using such methods is to obtain supervised data to train the retriever model, corresponding to pairs of query and support documents. In this paper, we propose a technique to learn retriever models for downstream tasks, inspired by knowledge distillation, and which does not require annotated pairs of query and documents. Our approach leverages attention scores of a reader model, used to solve the task based on retrieved documents, to obtain synthetic labels for the retriever. We evaluate our method on question answering, obtaining state-of-the-art results.

Code Repositories

FenQQQ/Fusion-in-decoder
pytorch
Mentioned in GitHub
hackerchenzhuo/LaKo
pytorch
Mentioned in GitHub
facebookresearch/FiD
Official
pytorch
Mentioned in GitHub
lucidrains/marge-pytorch
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
question-answering-on-narrativeqaFiD+Distil
BLEU-1: 35.3
BLEU-4: 7.5
METEOR: 11.1
Rouge-L: 32
question-answering-on-triviaqaFiD+Distil
EM: 72.1

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp