Generative AI Series

Retrieval Augmented Generation(RAG) — Chatbot for Wikipedia with LlamaIndex

Implement the RAG technique using Langchain, and LlamaIndex for conversational chatbot on Wikipedia.

A B Vijay Kumar
3 min readFeb 8, 2024


This blog is an ongoing series on GenerativeAI and is a continuation of the previous blog, which talks about the RAG pattern and how RAG is used to augment prompts and enhance the content and context of an LLM, with specific data.

In this blog, we will build a Q&A chatbot for content fetched from Wikipedia on Star Wars Franchise. We will be using the LlamaIndex reader WikipediaReader

Please go through the following blogs

  1. Prompt Engineering: Retrieval Augmented Generation(RAG)
  2. Retrieval Augmented Generation(RAG) with LlamaIndex
  3. Retrieval Augmented Generation(RAG) with LlamaIndex on a Database (Text2SQL)

Let's walk through the code. I won’t be walking through all the code, as it is very similar to what I had published and explained in my previous blogs. I will just call out specific changes made.

The above code is very similar to what we had done before, except that we are importing a reader WikipediaReader We will use this to parse a specific topic through Wikipedia (Star Wars). This content will be chunked, indexed, and stored as vectors. Please refer to my previous blogs for details.

In the above code, we are initializing the WikipediaReader and providing the topic “Star Wards Movie” and creating a vector store, with that documents

The following code is very similar to what we did in the previous blogs. Please refer to those blogs for explanations

Let's run this with streamlit run



A B Vijay Kumar

IBM Fellow, Master Inventor, Mobile, RPi & Cloud Architect & Full-Stack Programmer