diff --git a/UseCases/Chat_with_docs_GenAI/Chat_with_docs_OpenSource_GenAI_Python.ipynb b/UseCases/Chat_with_docs_GenAI/Chat_with_docs_OpenSource_GenAI_Python.ipynb index 8c4f9007..4b468379 100644 --- a/UseCases/Chat_with_docs_GenAI/Chat_with_docs_OpenSource_GenAI_Python.ipynb +++ b/UseCases/Chat_with_docs_GenAI/Chat_with_docs_OpenSource_GenAI_Python.ipynb @@ -19,65 +19,65 @@ "id": "fff71661-19b4-423a-867a-7c815b064c81", "metadata": {}, "source": [ - "

Introduction:

\n", + "

Introduction:

\n", "\n", - "

In the Chat with documentation system using Generative AI demo, the combination of RAG, Langchain, and LLM models allows users to ask queries in layman's terms, retrieve relevant information from the Vector store, and generate accurate and concise answers based on the retrieved data. This integration of retrieval-based and generative-based approaches provides a powerful tool for extracting knowledge from structured sources and delivering user-friendly responses.

\n", + "

In the Chat with documentation system using Generative AI demo, the combination of RAG, Langchain, and LLM models allows users to ask queries in layman's terms, retrieve relevant information from the Vector store, and generate accurate and concise answers based on the retrieved data. This integration of retrieval-based and generative-based approaches provides a powerful tool for extracting knowledge from structured sources and delivering user-friendly responses.

\n", "\n", - "

In this demo we will build Chatbot using Panel (for chat UI), LangChain, a powerful library for working with LLMs like GPT-3.5, GPT-4, Bloom, etc. and JumpStart in ClearScape notebooks, a system is built where users can ask business questions in natural English and receive answers with data drawn from the relevant databases.

\n", + "

In this demo we will build Chatbot using Panel (for chat UI), LangChain, a powerful library for working with LLMs like GPT-3.5, GPT-4, Bloom, etc. and JumpStart in ClearScape notebooks, a system is built where users can ask business questions in natural English and receive answers with data drawn from the relevant databases.

\n", "\n", - "

The following diagram illustrates the architecture.

\n", + "

The following diagram illustrates the architecture.

\n", "\n", "
\"architecture\"
\n", "\n", "\n", "
\n", - "

Before going any farther, let's get a better understanding of RAG, LangChain, and LLM.

\n", + "

Before going any farther, let's get a better understanding of RAG, LangChain, and LLM.

\n", "\n", - "
  1. Retrieval-Augmented Generation (RAG):
\n", - "

   RAG is a framework that combines the strengths of retrieval-based and generative-based approaches in question-answering systems.It utilizes both a retrieval model and a generative model to generate high-quality answers to user queries. The retrieval model is responsible for retrieving relevant information from a knowledge source, such as a database or documents. The generative model then takes the retrieved information as input and generates concise and accurate answers in natural language.

\n", + "
  1. Retrieval-Augmented Generation (RAG):
\n", + "

   RAG is a framework that combines the strengths of retrieval-based and generative-based approaches in question-answering systems.It utilizes both a retrieval model and a generative model to generate high-quality answers to user queries. The retrieval model is responsible for retrieving relevant information from a knowledge source, such as a database or documents. The generative model then takes the retrieved information as input and generates concise and accurate answers in natural language.

\n", "\n", "\n", - "

A typical RAG (Retrieval-and-Generation) application has two main components:

\n", + "

A typical RAG (Retrieval-and-Generation) application has two main components:

\n", "\n", - "

Indexing: a pipeline for ingesting data from a source and indexing it. This usually happens offline. The indexing process involves several steps, including loading the data, splitting it into smaller chunks, and storing and indexing the splits. This is often done using a VectorStore and Embeddings model.

\n", + "

Indexing: a pipeline for ingesting data from a source and indexing it. This usually happens offline. The indexing process involves several steps, including loading the data, splitting it into smaller chunks, and storing and indexing the splits. This is often done using a VectorStore and Embeddings model.

\n", " \n", - "

Retrieval and generation: the actual RAG chain, which takes the user query at run time and retrieves the relevant data from the index, then passes that to the model. The retrieval process involves searching the index for the most relevant data based on the user query, and then passing that data to the model for generation.

\n", + "

Retrieval and generation: the actual RAG chain, which takes the user query at run time and retrieves the relevant data from the index, then passes that to the model. The retrieval process involves searching the index for the most relevant data based on the user query, and then passing that data to the model for generation.

\n", "\n", - "

The most common full sequence from raw data to answer looks like:

\n", - "

Indexing

\n", - "