Skip to content

Mando-03/Mini-RAG-App

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 

Repository files navigation

mini-rag

This is a minimal implementation of the RAG model for question answering.

Requirements

  • Python 3.8 or later

Install Python using MiniConda

  1. Download and install MiniConda from here
  2. Create a new environment using the following command:
$ conda create -n mini-rag python=3.8
  1. Activate the environment:Ll
$ conda activate mini-rag

(Optional) Setup you command line interface for better readability

export PS1="\[\033[01;32m\]\u@\h:\w\n\[\033[00m\]\$ "

(Optional) Run Ollama Local LLM Server using Colab + Ngrok

Installation

Install the required packages

$ pip install -r requirements.txt

Setup the environment variables

$ cp .env.example .env

Set your environment variables in the .env file. Like OPENAI_API_KEY value.

Run Docker Compose Services

$ cd docker
$ cp .env.example .env
  • update .env with your credentials
$ cd docker
$ sudo docker compose up -d

Run the FastAPI server

$ uvicorn main:app --reload --host 0.0.0.0 --port 5000

Releases

No releases published

Packages

No packages published

Languages