Skip to content

Python project that uses ChromaDB and Ollama to generate answers based on documents. Only uses open source products, and it meant to run locally on your computer.

License

Notifications You must be signed in to change notification settings

jan-gerritsen/local_rag

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Retrieval Augmented Generation (RAG) with Ollama and ChromaDB

A Python project that uses ChromaDB and Ollama to generate answers based on documents. Only uses open source products, and it meant to run locally on your computer.

Description

A client was looking for a RAG solution as part of their products. I did this project as a test case to see how difficult that would be. I had found some articles, but those were typically notebooks, and not meant to run continuously.

This project has two Python scripts:

  • load_data.py loads the document that you want to ask questions about into a ChromaDB, and persists it.
  • rag_server.py connects to the ChromaDB and answers questions based on the information in the document.

The code is adapted from this article on Medium.

Getting Started

Dependencies

  • Ollama
  • Python 3.9

Installing

  • These instructions are for a Mac, I assume it would be similar on any other OS
  • Install Ollama on your computer: https://ollama.com/download/mac
  • Create a virtual environment based on Python 3.9
  • Download a sample document. I used Alice in Wonderland
  • Run python -m pip install -r requirements.txt

Ensure that

loader = DirectoryLoader('./../../data/')

in load_data.py points to the directory that contains your document(s).

Executing program

  • Ensure Ollama is running
  • Run python load_data.py. This may take a couple of minutes
  • Run python rag_server.py

Enhancements

This program only uses one document as its knowledge base, but it is not difficult to extend this to multiple directories, and also multiple formats. For more info, see langchain document loaders.

Help

Feel free to reach out with comments or issues.

Authors

Jan Gerritsen

Version History

  • 0.1
    • Initial Release

License

This project is licensed under the MIT License - see the license.txt file for details

Acknowledgments

About

Python project that uses ChromaDB and Ollama to generate answers based on documents. Only uses open source products, and it meant to run locally on your computer.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages