mirror of
https://github.com/deepset-ai/haystack.git
synced 2025-08-28 02:16:32 +00:00
190 lines
6.1 KiB
Markdown
190 lines
6.1 KiB
Markdown
![]() |
<!---
|
||
|
title: "Tutorial 7"
|
||
|
metaTitle: "Generative QA with RAG"
|
||
|
metaDescription: ""
|
||
|
slug: "/docs/tutorial7"
|
||
|
date: "2020-11-12"
|
||
|
id: "tutorial7md"
|
||
|
--->
|
||
|
|
||
|
# Generative QA with "Retrieval-Augmented Generation"
|
||
|
|
||
|
[](https://colab.research.google.com/github/deepset-ai/haystack/blob/master/tutorials/Tutorial7_RAG_Generator.ipynb)
|
||
|
|
||
|
While extractive QA highlights the span of text that answers a query,
|
||
|
generative QA can return a novel text answer that it has composed.
|
||
|
In this tutorial, you will learn how to set up a generative system using the
|
||
|
[RAG model](https://arxiv.org/abs/2005.11401) which conditions the
|
||
|
answer generator on a set of retrieved documents.
|
||
|
|
||
|
### Prepare environment
|
||
|
|
||
|
#### Colab: Enable the GPU runtime
|
||
|
Make sure you enable the GPU runtime to experience decent speed in this tutorial.
|
||
|
**Runtime -> Change Runtime type -> Hardware accelerator -> GPU**
|
||
|
|
||
|
<img src="https://raw.githubusercontent.com/deepset-ai/haystack/master/docs/_src/img/colab_gpu_runtime.jpg">
|
||
|
|
||
|
|
||
|
```python
|
||
|
# Make sure you have a GPU running
|
||
|
!nvidia-smi
|
||
|
```
|
||
|
|
||
|
Here are the packages and imports that we'll need:
|
||
|
|
||
|
|
||
|
```python
|
||
|
# Install the latest release of Haystack in your own environment
|
||
|
#! pip install farm-haystack
|
||
|
|
||
|
# Install the latest master of Haystack
|
||
|
!pip install --upgrade pip
|
||
|
!pip install git+https://github.com/deepset-ai/haystack.git#egg=farm-haystack[colab,faiss]
|
||
|
```
|
||
|
|
||
|
|
||
|
```python
|
||
|
from typing import List
|
||
|
import requests
|
||
|
import pandas as pd
|
||
|
from haystack import Document
|
||
|
from haystack.document_stores import FAISSDocumentStore
|
||
|
from haystack.nodes import RAGenerator, DensePassageRetriever
|
||
|
```
|
||
|
|
||
|
Let's download a csv containing some sample text and preprocess the data.
|
||
|
|
||
|
|
||
|
|
||
|
```python
|
||
|
# Download sample
|
||
|
temp = requests.get(
|
||
|
"https://raw.githubusercontent.com/deepset-ai/haystack/master/tutorials/small_generator_dataset.csv"
|
||
|
)
|
||
|
open("small_generator_dataset.csv", "wb").write(temp.content)
|
||
|
|
||
|
# Create dataframe with columns "title" and "text"
|
||
|
df = pd.read_csv("small_generator_dataset.csv", sep=",")
|
||
|
# Minimal cleaning
|
||
|
df.fillna(value="", inplace=True)
|
||
|
|
||
|
print(df.head())
|
||
|
```
|
||
|
|
||
|
We can cast our data into Haystack Document objects.
|
||
|
Alternatively, we can also just use dictionaries with "text" and "meta" fields
|
||
|
|
||
|
|
||
|
```python
|
||
|
# Use data to initialize Document objects
|
||
|
titles = list(df["title"].values)
|
||
|
texts = list(df["text"].values)
|
||
|
documents: List[Document] = []
|
||
|
for title, text in zip(titles, texts):
|
||
|
documents.append(Document(content=text, meta={"name": title or ""}))
|
||
|
```
|
||
|
|
||
|
Here we initialize the FAISSDocumentStore, DensePassageRetriever and RAGenerator.
|
||
|
FAISS is chosen here since it is optimized vector storage.
|
||
|
|
||
|
|
||
|
```python
|
||
|
# Initialize FAISS document store.
|
||
|
# Set `return_embedding` to `True`, so generator doesn't have to perform re-embedding
|
||
|
document_store = FAISSDocumentStore(faiss_index_factory_str="Flat", return_embedding=True)
|
||
|
|
||
|
# Initialize DPR Retriever to encode documents, encode question and query documents
|
||
|
retriever = DensePassageRetriever(
|
||
|
document_store=document_store,
|
||
|
query_embedding_model="facebook/dpr-question_encoder-single-nq-base",
|
||
|
passage_embedding_model="facebook/dpr-ctx_encoder-single-nq-base",
|
||
|
use_gpu=True,
|
||
|
embed_title=True,
|
||
|
)
|
||
|
|
||
|
# Initialize RAG Generator
|
||
|
generator = RAGenerator(
|
||
|
model_name_or_path="facebook/rag-token-nq",
|
||
|
use_gpu=True,
|
||
|
top_k=1,
|
||
|
max_length=200,
|
||
|
min_length=2,
|
||
|
embed_title=True,
|
||
|
num_beams=2,
|
||
|
)
|
||
|
```
|
||
|
|
||
|
We write documents to the DocumentStore, first by deleting any remaining documents then calling `write_documents()`.
|
||
|
The `update_embeddings()` method uses the retriever to create an embedding for each document.
|
||
|
|
||
|
|
||
|
|
||
|
```python
|
||
|
# Delete existing documents in documents store
|
||
|
document_store.delete_documents()
|
||
|
|
||
|
# Write documents to document store
|
||
|
document_store.write_documents(documents)
|
||
|
|
||
|
# Add documents embeddings to index
|
||
|
document_store.update_embeddings(retriever=retriever)
|
||
|
```
|
||
|
|
||
|
Here are our questions:
|
||
|
|
||
|
|
||
|
```python
|
||
|
QUESTIONS = [
|
||
|
"who got the first nobel prize in physics",
|
||
|
"when is the next deadpool movie being released",
|
||
|
"which mode is used for short wave broadcast service",
|
||
|
"who is the owner of reading football club",
|
||
|
"when is the next scandal episode coming out",
|
||
|
"when is the last time the philadelphia won the superbowl",
|
||
|
"what is the most current adobe flash player version",
|
||
|
"how many episodes are there in dragon ball z",
|
||
|
"what is the first step in the evolution of the eye",
|
||
|
"where is gall bladder situated in human body",
|
||
|
"what is the main mineral in lithium batteries",
|
||
|
"who is the president of usa right now",
|
||
|
"where do the greasers live in the outsiders",
|
||
|
"panda is a national animal of which country",
|
||
|
"what is the name of manchester united stadium",
|
||
|
]
|
||
|
```
|
||
|
|
||
|
Now let's run our system!
|
||
|
The retriever will pick out a small subset of documents that it finds relevant.
|
||
|
These are used to condition the generator as it generates the answer.
|
||
|
What it should return then are novel text spans that form and answer to your question!
|
||
|
|
||
|
|
||
|
```python
|
||
|
# Or alternatively use the Pipeline class
|
||
|
from haystack.pipelines import GenerativeQAPipeline
|
||
|
from haystack.utils import print_answers
|
||
|
|
||
|
pipe = GenerativeQAPipeline(generator=generator, retriever=retriever)
|
||
|
for question in QUESTIONS:
|
||
|
res = pipe.run(query=question, params={"Generator": {"top_k": 1}, "Retriever": {"top_k": 5}})
|
||
|
print_answers(res, details="minimum")
|
||
|
```
|
||
|
|
||
|
## About us
|
||
|
|
||
|
This [Haystack](https://github.com/deepset-ai/haystack/) notebook was made with love by [deepset](https://deepset.ai/) in Berlin, Germany
|
||
|
|
||
|
We bring NLP to the industry via open source!
|
||
|
Our focus: Industry specific language models & large scale QA systems.
|
||
|
|
||
|
Some of our other work:
|
||
|
- [German BERT](https://deepset.ai/german-bert)
|
||
|
- [GermanQuAD and GermanDPR](https://deepset.ai/germanquad)
|
||
|
- [FARM](https://github.com/deepset-ai/FARM)
|
||
|
|
||
|
Get in touch:
|
||
|
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Slack](https://haystack.deepset.ai/community/join) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
|
||
|
|
||
|
By the way: [we're hiring!](https://www.deepset.ai/jobs)
|