This is the first official release of RAG_Ollama_Mac - a lightweight, private, and customizable retrieval-augmented chatbot running entirely on your Mac.
- ๐ค Local RAG Implementation: Complete retrieval-augmented generation system using Ollama
- ๐ PDF Processing: Load and process multiple PDF documents from the data directory
- ๐ฌ Interactive Chat: Clean and intuitive Streamlit-based chat interface
- ๐ Smart Retrieval: ChromaDB-powered vector storage for efficient document retrieval
- ๐ Privacy First: All processing happens locally - no cloud API calls
- Embedding Model:
nomic-embed-text-v2-moefor high-quality document embeddings - LLM:
gemma3nfor intelligent, context-aware responses - Vector Database: ChromaDB for persistent vector storage
- UI Framework: Streamlit for interactive web interface
- โ Complete source code for RAG chatbot
- โ Document loading and processing utilities
- โ Configurable text chunking (800 chars with 80 char overlap)
- โ Vector database management with reset functionality
- โ Interactive chat interface with message history
- โ Comprehensive documentation and setup guide
- โ MIT License
# Clone the repository
git clone https://github.com/eplt/RAG_Ollama_Mac.git
cd RAG_Ollama_Mac
# Create virtual environment
python3 -m venv venv
source venv/bin/activate
# Install dependencies
pip install -r ./src/requirements.txt-
Start Ollama and pull required models:
ollama serve ollama pull gemma3n ollama pull toshk0/nomic-embed-text-v2-moe:Q6_K
-
Load your documents:
python ./src/load_docs.py
-
Launch the chatbot:
streamlit run ./src/UI.py
- Operating System: macOS (optimized for Mac, but works on other platforms)
- Python: 3.7 or higher
- RAM: At least 16GB recommended (24GB+ for optimal performance)
- Ollama: Latest version with required models installed
Key dependencies included in this release:
- langchain (0.3.26)
- langchain-ollama (0.3.5)
- langchain-chroma (0.2.4)
- chromadb (1.0.15)
- streamlit (1.41.1)
- ollama (0.5.1)
See src/requirements.txt for complete list.
- Research: Query academic papers and research documents
- Documentation: Get instant answers from technical documentation
- Books: Ask questions about long-form content
- Privacy-Sensitive Work: Process confidential documents locally
- Support for additional file formats (Markdown, .txt, HTML)
- Multiple LLM model support (phi-3, mistral, llama3)
- Chat history persistence
- Enhanced document management
- User authentication options
This project is based on the excellent work by pruthvirajcyn.
Released under the MIT License. See LICENSE for details.
None reported for this initial release.
For detailed changes, see CHANGELOG.md.
Full Changelog: https://github.com/eplt/RAG_Ollama_Mac/commits/v0.1.0