Generative QA uses large language models to generate human-like responses to user queries in question answering apps. Instead of simply extracting answers from existing documents, generative systems create a new text based on instructions provided in the prompt. A prompt is a specific instruction given to the model in natural language to help it understand the task and generate an appropriate response.
To build a basic generative QA system, you need a large language model (LLM), like GPT-4, and a simple prompt, such as a user query. LLMs are trained to predict the next word in a sequence and generate answers token by token. Based on the prompt, the model generates an answer.
Generative QA offers several advantages compared to extractive QA:
- Extracting information from multiple sources: The system can extract and combine information from various sources to produce coherent and informative answers from scratch.
- Generating original content: LLMs go beyond extracting existing answers and produce more creative and personalized responses. You can use them to create a system that matches your brand voice or feels like talking to a real human.
- Language flexibility: LLMs are trained on large amounts of text, allowing them to generate natural language responses that include idioms or language variations.
- Reasoning capabilities: LLMs have some reasoning capabilities, which make it possible for them to compare facts or draw conclusions.
Generative QA systems also have limitations that are important to consider:
- Cost: Proprietary LLMs often come with a price tag. The pricing models are based on the number of tokens processed or per query.
- Limited context length: Models have a maximum token limit they can handle. This includes both the prompt and the generated output. When handling long documents, the model may truncate the generated text to fit within the allowed context length.
- Factual accuracy: Models may generate fictional or incorrect outputs with a high level of confidence. This is known as hallucinations. Hallucinations can occur because of biases in training data or the model's inability to differentiate between factual and fictional information.
- Latency: Generative QA systems are generally slower compared to extractive QA systems.
- Output control: LLMs can sometimes generate harmful, inappropriate, or biased content.
- Evaluation: Evaluating generative models remains challenging. Some of the reasons for that include a lack of objective metrics. Evaluation often involves subjective human judgment, and the diverse outputs make establishing a single ground truth difficult.
To address concerns about hallucinations and unreliable answers, you can run a generative QA system on your own data using a retrieval-augmented generation (RAG) approach. This involves limiting the context to a predefined set of documents and adding a Retriever component to your QA system.
In deepset Cloud, you can pass your documents in the prompt. When given a query, the Retriever finds the most relevant documents, injects them into the prompt, and passes them on to the generative model. The model uses these documents to produce the final answer.
To make sure the generated answers are based on the documents you feed to the model, add a ReferencePredictor component to your pipeline. It adds references to each piece of the answer text so that you can verify it's grounded in your data. This helps build a reliable system that users can trust.
Generative QA systems excel in scenarios where you need novel, natural-language answers and not verbatim from existing documents. Some popular applications include:
- Chatbots, AI assistants, and customer support applications offering personalized assistance across various questions.
- Writing aids and content generation apps automating content creation tasks and assisting with content curation.
- Learning assistants in educational applications, providing explanations and summaries of content.
- Translation aids.
Generative QA pipelines in deepset Cloud:
- Can work on your own data (retrieval-augmented generation). You can pass the documents in the prompt and use a Retriever node to filter them.
- Are not chat-based. They're search systems that generate novel answers based on your data or the model's knowledge of the world, depending on the setup.
Chat in deepset Cloud
We've introduced a Chat pipeline template that you can use out of the box and then test it in Playground.
deepset Cloud supports you throughout the process of creating a generative QA system by:
- Providing a pipeline template ready to use as it is.
- Providing Prompt Studio, which is a sandbox environment for prompt engineering.
- Offering a library of tested and curated prompts you can reuse and modify as needed.
- Providing the Groundedness Observability dashboard where you can check if the answers are grounded in the documents.
deepset Cloud uses PromptNode for building RAG pipelines. By default, it uses the free google/flan-t5-base model, but you can replace it with other models, like GPT-4 by OpenAI, Claude by Anthropic, Command by Cohere, and more. See Language Models in deepset Cloud and Using Hosted LLMs in Your Pipelines.
Take a look at Use Case: Generative AI Systems to learn when it's best to use a generative system and what data it needs.
And if you want to see what it looks like in production, test our Generative QA pipeline that runs on 10-K annual reports of the largest companies.
Updated 4 days ago