AI-powered search and chat for Uganda's Law Corpus
All code & data used is 100% open-source.
This is the website with the law material that is scrapped. I built it with Nextra.
The dataset is a json file containing all text.
View it here.
I recommend getting familiar with fetching, cleaning, and storing data as outlined in the scraping and embedding scripts below, but feel free to skip those steps and just use the dataset.
LawSuite provides 2 things:
- A search interface for Uganda's law corpus e.g constitution, penal code.
- A chat interface for the information.
Search was created with OpenAI Embeddings (text-embedding-ada-002
).
First, we loop over our website and generate embeddings for each chunk of text.
Then in the app we take the user's search query, generate an embedding, and use the result to find the most similar passages from the book.
The comparison is done using cosine similarity across our database of vectors.
Our database is a Postgres database with the pgvector extension hosted on Supabase.
Results are ranked by similarity score and returned to the user.
Chat builds on top of search. It uses search results to create a prompt that is fed into GPT-3.
This allows for a chat-like experience where the user can ask questions about the book and get answers.
Here's a quick overview of how to run it locally.
- Set up OpenAI
You'll need an OpenAI API key to generate embeddings.
- Set up Supabase
There is a schema.sql file in the root of the repo that you can use to set up the database.
Run that in the SQL editor in Supabase.
I recommend turning on Row Level Security and setting up a service role to use with the app.
Note: You don't have to use Supabase. Use whatever method you prefer to store your data. But I like Supabase and think it's easy to use.
- Clone repo
git clone https://github.com/greatsage-raphael/lawSuite.git
- Install dependencies
npm i
- Set up environment variables
Create a .env.local file in the root of the repo with the following variables:
OPENAI_API_KEY=
NEXT_PUBLIC_SUPABASE_URL=
SUPABASE_SERVICE_ROLE_KEY=
- Run scraping script
npm run scrape
This scrapes the text of the book from The Network State website and saves it to a json file.
- Run embedding script
npm run embed
This reads the json file, generates embeddings for each chunk of text, and saves the results to your database.
- Run app
npm run dev
Thanks to Mckay Wrigley for his work on The Network State GPT. This project would not be possible if he didn't make the project open-source. I highly recommend you check it out.
I would also like to shout out the open source technologies without which this project wouldn't have been possible: Supabase, Nextjs, Nextra. I highly recommend you check them out.
If you have any questions, feel free to reach out to me on Gmail.
New feature incoming: Ability to upload a pdf and have questions about the pdf answered.