Skip to content

Latest commit

ย 

History

History
118 lines (84 loc) ยท 3.33 KB

File metadata and controls

118 lines (84 loc) ยท 3.33 KB

Release Notes - v0.1.0

๐ŸŽ‰ Initial Public Release

This is the first official release of RAG_Ollama_Mac - a lightweight, private, and customizable retrieval-augmented chatbot running entirely on your Mac.

โœจ Features

Core Functionality

  • ๐Ÿค– Local RAG Implementation: Complete retrieval-augmented generation system using Ollama
  • ๐Ÿ“„ PDF Processing: Load and process multiple PDF documents from the data directory
  • ๐Ÿ’ฌ Interactive Chat: Clean and intuitive Streamlit-based chat interface
  • ๐Ÿ” Smart Retrieval: ChromaDB-powered vector storage for efficient document retrieval
  • ๐Ÿ”’ Privacy First: All processing happens locally - no cloud API calls

Technical Details

  • Embedding Model: nomic-embed-text-v2-moe for high-quality document embeddings
  • LLM: gemma3n for intelligent, context-aware responses
  • Vector Database: ChromaDB for persistent vector storage
  • UI Framework: Streamlit for interactive web interface

๐Ÿ“‹ What's Included

  • โœ… Complete source code for RAG chatbot
  • โœ… Document loading and processing utilities
  • โœ… Configurable text chunking (800 chars with 80 char overlap)
  • โœ… Vector database management with reset functionality
  • โœ… Interactive chat interface with message history
  • โœ… Comprehensive documentation and setup guide
  • โœ… MIT License

๐Ÿ› ๏ธ Installation

# Clone the repository
git clone https://github.com/eplt/RAG_Ollama_Mac.git
cd RAG_Ollama_Mac

# Create virtual environment
python3 -m venv venv
source venv/bin/activate

# Install dependencies
pip install -r ./src/requirements.txt

๐Ÿš€ Quick Start

  1. Start Ollama and pull required models:

    ollama serve
    ollama pull gemma3n
    ollama pull toshk0/nomic-embed-text-v2-moe:Q6_K
  2. Load your documents:

    python ./src/load_docs.py
  3. Launch the chatbot:

    streamlit run ./src/UI.py

๐Ÿ’ป System Requirements

  • Operating System: macOS (optimized for Mac, but works on other platforms)
  • Python: 3.7 or higher
  • RAM: At least 16GB recommended (24GB+ for optimal performance)
  • Ollama: Latest version with required models installed

๐Ÿ“ฆ Dependencies

Key dependencies included in this release:

  • langchain (0.3.26)
  • langchain-ollama (0.3.5)
  • langchain-chroma (0.2.4)
  • chromadb (1.0.15)
  • streamlit (1.41.1)
  • ollama (0.5.1)

See src/requirements.txt for complete list.

๐ŸŽฏ Use Cases

  • Research: Query academic papers and research documents
  • Documentation: Get instant answers from technical documentation
  • Books: Ask questions about long-form content
  • Privacy-Sensitive Work: Process confidential documents locally

๐Ÿ”ฎ Future Plans

  • Support for additional file formats (Markdown, .txt, HTML)
  • Multiple LLM model support (phi-3, mistral, llama3)
  • Chat history persistence
  • Enhanced document management
  • User authentication options

๐Ÿ™ Acknowledgments

This project is based on the excellent work by pruthvirajcyn.

๐Ÿ“„ License

Released under the MIT License. See LICENSE for details.

๐Ÿ› Known Issues

None reported for this initial release.

๐Ÿ“ Changelog

For detailed changes, see CHANGELOG.md.


Full Changelog: https://github.com/eplt/RAG_Ollama_Mac/commits/v0.1.0