A Python project that uses ChromaDB and Ollama to generate answers based on documents. Only uses open source products, and it meant to run locally on your computer.
A client was looking for a RAG solution as part of their products. I did this project as a test case to see how difficult that would be. I had found some articles, but those were typically notebooks, and not meant to run continuously.
This project has two Python scripts:
- load_data.py loads the document that you want to ask questions about into a ChromaDB, and persists it.
- rag_server.py connects to the ChromaDB and answers questions based on the information in the document.
The code is adapted from this article on Medium.
- Ollama
- Python 3.9
- These instructions are for a Mac, I assume it would be similar on any other OS
- Install Ollama on your computer: https://ollama.com/download/mac
- Create a virtual environment based on Python 3.9
- Download a sample document. I used Alice in Wonderland
- Run
python -m pip install -r requirements.txt
Ensure that
loader = DirectoryLoader('./../../data/')
in load_data.py points to the directory that contains your document(s).
- Ensure Ollama is running
- Run
python load_data.py
. This may take a couple of minutes - Run
python rag_server.py
This program only uses one document as its knowledge base, but it is not difficult to extend this to multiple directories, and also multiple formats. For more info, see langchain document loaders.
Feel free to reach out with comments or issues.
Jan Gerritsen
- 0.1
- Initial Release
This project is licensed under the MIT License - see the license.txt file for details