Skip to content

RAG technique for QA tasks with PDF, website and video, leveraging LLMs and the OpenAI API for interactive applications

Notifications You must be signed in to change notification settings

dinhquy-nguyen-1704/chat-with-pdf-website

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

chat-with-pdf-website

1. Introduction

The codebase demonstrates the fundamental application of RAG for question-answering tasks, leveraging the power of open-source Large Language Models (LLMs) from HuggingFace for interacting with PDFs (chat_with_pdf) and utilizing the OpenAI API key for a simple question-answering website (chat_with_website) and a simple question-answering video (chat_with_video) using langchain.

Retrieval-Augmented Generation Workflow

Retrieval-Augmented Generation Workflow

2. Getting Started

git clone https://github.com/dinhquy-nguyen-1704/chat-with-pdf-website.git
cd chat-with-pdf-website
conda create --name chat-with-pdf-website python=3.10
conda activate chat-with-pdf-website
pip install -r requirements.txt

3. Inference

3.1 chat_with_pdf

First, change to the chat_with_pdf directory and create new folder models.

cd chat_with_pdf
mkdir models

Then, download models (LLM and embedding model) you want to use. In my source code, the default LLM is vinallama-2.7b-chat_q5_0.gguf and the default embedding model is all-MiniLM-L6-v2-f16.gguf. You should organize the folder structure as follows:

  • 📁 chat-with-pdf-website
    • 📁 chat_with_pdf
      • 📂 data
        • 📄 your_file.pdf
      • 📁 models
        • all-MiniLM-L6-v2-f16.gguf
        • vinallama-2.7b-chat_q5_0.gguf
      • 📁 vectorstores
      • 🐍 config.py
      • 🐍 create_vector_db.py
      • 🐍 qa_bot.py
      • 🐍 utils.py
    • 📁 chat_with_website
      • 🐍 utils.py
      • 🐍 app.py
    • 📄 README.md
    • 📄 requirements.txt

Delete 2 files index.faiss and index.pkl in vectorstores if you want to use your_file.pdf.

After that, run file create_vector_db.py

python create_vector_db.py

When the above command is completed, two files named index.faiss and index.pkl will appear in the vectorstores.

Now, you can use chatbot to ask questions about the information in the your_file.pdf file in the command line environment.

python qa_bot.py --question "your_question"

3.2 chat_with_website

First, change to the chat_with_website directory

cd chat_with_website

Next, replace the OpenAI API key in the app.py file.

import os
os.environ["OPENAI_API_KEY"] = "sk-..."
api_key = os.getenv("OPENAI_API_KEY")

Now, you can run the app.py file and a Streamlit chatbot interface will appear.

python -m streamlit run app.py

You can paste a link to any website and ask for information related to that website.

Streamlit GUI

Streamlit GUI

3.3 chat_with_video

First, change to the chat_with_video directory

cd chat_with_video

Change OpenAI API key at the first line of gradio.py

API_KEY = "sk-..."

Change youtube_url

loader = YoutubeLoader.from_youtube_url("https://www.youtube.com/watch?v=tcqEUSNCn8I", add_video_info=True)

Finally, run gradio.py, a link to Gradio interface will appear.

python gradio.py

Gradio

Gradio

4. Contact

If you have any questions or feedback, please open an issue in this repository
or send an email to [email protected].

5. Reference

About

RAG technique for QA tasks with PDF, website and video, leveraging LLMs and the OpenAI API for interactive applications

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages