Skip to content

xxxily/hello-ai

Repository files navigation

Hello-AI Logo

Hello-AI

An intelligent, auto-updating directory of high-quality open-source projects.

Docs License GitHub Stars Node.js

English | 中文文档

📊 Project Statistics

Summary of high-quality open-source AI projects collected from the internet:

  • 📁 Total Collected: 17740 projects
  • Active Shown: 8294 projects (updated within the last 6 months)
  • 🏷️ Categories (Active / Total):
    • 🔥 Trending: 30 / 30
    • 🧠 Foundation Models: 133 / 537
    • 🤖 Agents & Orchestration: 1171 / 1487
    • 🔍 RAG & Data Engineering: 380 / 606
    • ☁️ Infrastructure & Deployment: 854 / 1428
    • 🔧 Fine-tuning & Training: 346 / 878
    • 👁️ Multimodal (Audio/Video): 828 / 2634
    • 🛠️ Developer Tools & SDKs: 1822 / 3120
    • 🎨 AI Applications: 775 / 1467
    • 📚 Learning Resources: 1150 / 3992
    • 💻 Desktop & OS Apps: 236 / 326
    • 🦾 Robotics & IoT: 456 / 998
    • 💼 Business & Finance: 155 / 299
  • 📅 Last Updated: 2026-04-14

Overview

Hello-AI is an intelligent, AI-driven navigation hub for open-source AI projects.

In an era of rapid AI evolution, developers often face "information overload" when navigating the vast sea of GitHub repositories. Hello-AI leverages AI Agent automation to discover, evaluate, and organize the most high-quality AI resources from the global open-source community.

✨ Key Features

  • 🤖 Autonomous Smart Maintenance: Unlike traditional link directories maintained manually, Hello-AI uses AI agents for 24/7 autonomous project discovery, quality assessment, and categorization. It's "AI discovering AI."
  • 🗺️ Evolutionary Landscape Map: Deep coverage across foundational models, Agent frameworks, RAG, infrastructure, multimodal apps, and dev tools—organized precisely and intuitively.
  • 🔄 Dynamic Activity Tracking: The system automatically purges stale projects and dynamically updates Star counts and health status, ensuring you only see the most relevant and active repositories.
  • Instant Value Insights: AI automatically generates concise summaries and use-cases for every repository, allowing you to bypass manual research and identify the right tools in seconds.

This is your shortcut to exploring the boundaries of AI and finding the ultimate productivity tools.

🏗️ Architecture & Execution Logic

This project operates entirely through the collaboration of automated scripts and Large Language Models. Below is a visual flowchart demonstrating the complete data lifecycle—from discovery to frontend rendering:

graph TD
    classDef default fill:#f9f9f9,stroke:#e0e0e0,stroke-width:1px;
    classDef database fill:#e1f5fe,stroke:#4fc3f7,stroke-width:2px;
    classDef ai fill:#fff3e0,stroke:#ffb74d,stroke-width:2px;
    classDef process fill:#e8f5e9,stroke:#81c784,stroke-width:2px;
    
    subgraph Discovery [1. Auto-Evolving Discovery Layer]
        T[(topics.json<br/>Knowledge Pool)]:::database -->|Provide un-explored topics| Fetch[GitHub Crawl Script]:::process
        Fetch -.->|Inject newly found topics| T
        Fetch -->|Enqueue High-Star Projects| Q[(pending-projects.json<br/>Queue)]:::database
    end

    subgraph Evaluation [2. AI Batch Evaluation Layer]
        Q -->|Pop N Batch Projects| P[Combined AI Prompt]:::process
        DB[(projects.json<br/>Core Database)]:::database -.->|Inject dynamic categories| P
        P --> LLM((LLM AI Engine<br/>Tag / Summarize / Rate)):::ai
        LLM --> Condition{AI Decides}
    end

    subgraph Storage [3. Isolation Storage Layer]
        Condition -->|Rejected or Poor Quality| R[(rejected-projects/<br/>Audit Trash)]:::database
        Condition -->|Valuable AI Projects| DB
    end

    subgraph Frontend [4. Automated View Layer]
        DB -->|Dynamically Reflect Categories| Vite[VitePress Nav & Sidebar]:::process
        DB -->|Group by Subcategory| Gen[generate-docs.js]:::process
        Vite --> Site([🌍 Hello-AI Website])
        Gen -->|Generate Markdown Pages| Site
    end
Loading

Its core mechanism, data flow, and system architecture details are as follows:

1. Dynamic Auto-Evolving Discovery Layer

  • Topic Mining: Using a predefined seed list in data/topics.json, the crawler iterates over the GitHub API, prioritizing the "least recently explored" topics to search for new repositories with Stars >= 500.
  • Knowledge Base Growth: When unseen topics are detected from newly fetched projects, the system automatically registers them into topics.json as 'Level 2' (secondary exploration targets).
  • Pending Queue: All discovered new repositories flow directly into data/pending-projects.json for validation.

2. Local/Cloud AI Batch Evaluation Engine

  • Concurrent Batch Processing: The core script discover-and-evaluate.js pops a configured number of items (via EVALUATE_BATCH_SIZE) from the pending pool and creates a combined prompt for the LLM. This batch design drastically reduces API frequency limits and reuses token context.
  • Dynamic Category Routing: The system never "hard-codes" categories. Upon each evaluation, it dynamically reads the valid categories and subcategories from data/projects.json and instructs the AI to route projects accordingly.
  • Tagging & Auditing: The AI automatically extracts tags, generates optimal Chinese descriptions, and assigns the project to the most suitable subcategory. If an item is deemed unworthy or un-categorizable by the AI, it gets discarded into an isolation audit log (data/rejected-projects/).
  • Objective Trending List: A daily objective calculation forces the recalculation of the top 30 highest-star, recently updated projects, automatically placing them into the 🔥 Trending category, overriding AI randomness.

3. Automated Frontend Rendering & View Decoupling

  • Adaptive Routing Presentation: Built with VitePress, the Navbar and Sidebar have been rewritten from static mappings. Whenever categories are added or removed from projects.json, the VitePress compiler dynamically analyzes it and renders the UI perfectly, preventing data-to-UI discrepancies.
  • Smart Markdown Folding & Stale Cleanup: generate-docs.js iterates over categories, grouping items by subcategory. It also automatically purges projects that haven't been updated for a long time based on the RECENCY_THRESHOLD_MONTHS setting to maintain high project quality.

4. Automation Pipeline

  • For hands-free, continuous discovery (e.g., avoiding rate-limit drops), you can utilize process daemon scripts like scripts/loop-eval.js which leverage continuous sleep loops. This achieves a permanent closed-loop operation of: Discover -> Buffer -> AI Evaluate -> Static Page Build, endlessly exploring the ocean of open source code.

🚀 Local Deployment & Running Guide

You are completely welcome to run this entire auto-expanding AI knowledge base framework locally! It is very simple to start:

1. Environment & Setup

A Node.js environment is required (v18.x or above is recommended).

git clone https://github.com/xxxily/hello-ai.git
cd hello-ai
npm install

2. Environment Variables Configuration

Copy from the template:

cp .env.example .env

Open .env and adjust the core configurations:

  • GITHUB_TOKEN= (Highly Recommended): Bypasses the strict rate limits applied to anonymous GitHub search API calls.
  • LLM_API_KEY=: Your target LLM API Key (used for analyzing and curating projects).
    • 💡 Zero-Cost Prompt: If you are using a local LLM setup (e.g. Ollama via llama3), you can simply use LLM_API_KEY=local-fallback.
  • LLM_PROVIDER=: Select a built-in provider preset (openai, minimax, deepseek, ollama). When omitted, auto-detected from LLM_BASE_URL or provider-specific API key env vars.
  • LLM_BASE_URL=: LLM endpoint (e.g. https://api.openai.com/v1, or local http://127.0.0.1:11434/v1).
  • LLM_MODEL=: Standard model identity to use (e.g. gpt-4o-mini, MiniMax-M2.5).
  • DISCOVER_BATCH_SIZE / EVALUATE_BATCH_SIZE / UPDATE_STATUS_BATCH_SIZE: Modify limits per pull from GitHub, per LLM prompt, and for status update batching.
  • LOOP_INTERVAL_SECONDS: Configure the base idle time interval between consecutive ai:loop-eval cycles (default: 60s).
  • MAX_PAGES_DEFAULT: Default max pages to explore per topic (default: 5).
  • MAX_PAGES_QUALITY: Max pages for high-quality topics (default: 20).
  • QUALITY_TOPIC_THRESHOLD: Score threshold for high-quality topics (default: 5).
  • AUTO_FETCH_DESC_STARS: Star threshold to proactively fetch missing descriptions (default: 1000).
  • RECENCY_THRESHOLD_MONTHS: Only projects updated within the last N months are kept during doc generation (default: 24, i.e., 2 years).

Supported LLM Providers

The evaluation engine supports any OpenAI-compatible LLM API. Built-in presets make it easy to switch:

Provider LLM_PROVIDER Default Model API Key Env
OpenAI openai gpt-4o-mini OPENAI_API_KEY or LLM_API_KEY
MiniMax minimax MiniMax-M2.5 MINIMAX_API_KEY or LLM_API_KEY
DeepSeek deepseek deepseek-chat DEEPSEEK_API_KEY or LLM_API_KEY
Ollama (local) ollama llama3 N/A (uses local-fallback)

Quick start with MiniMax:

LLM_PROVIDER=minimax
MINIMAX_API_KEY=your-key-here
# Optionally pick a specific model:
# LLM_MODEL=MiniMax-M2.7

3. Run Automation Pipelines

Choose how you want to execute scripts:

  • Single Manual Execution:
    npm run ai:discover-eval
  • Constant Background Daemon (Continuous fetch & evaluate):
    npm run ai:loop-eval
  • Interactive TUI Daemon (Recommended for manual param selection):
    npm run ai:loop-eval-tui
  • Incremental Status Check (Background process to silently check/update GitHub star & health status for evaluated items):
    npm run ai:update-status
  • Re-Evaluate Active Projects (Moves items back to Queue to catch up with latest sub-category mapping):
    npm run ai:re-evaluate-all
  • Consume Queue & Exit (Strictly evaluates queue without pinging GitHub API to avoid rate-limits; auto-exits when empty):
    npm run ai:consume-queue

💡 Advanced CLI Flags

When running npm run ai:discover-eval or its variants, you can append the following flags:

  • --sort-topic-by=quality|time:
    • quality: Prioritize exploring topics with the highest quality score (based on already included projects).
    • time: Prioritize exploring the least recently explored topics.
  • --topic-order=asc|desc: Sorting direction (default: Desc for Quality, Asc for Time).
  • --consume-only: Evaluates local queue items only, skipping new GitHub searches.
  • --resume: Resumes discovery from the last saved topic and page.
  • --update-only: Mass updates existing projects without LLM evaluation.
  • --init-topics: (One-time) Re-initializes topic quality scores based on existing data in projects.json.

4. Dynamic Pages Generation & Local Preview

Once your AI has finished evaluating and saving into the core databank, preview the outcome locally!

# Iterates projects.json database, groups it into subcategories and auto-renders individual Markdown documents
npm run ai:generate-docs

# Boot local VitePress web server (live reloading with newly tracked projects)
npm run docs:dev

# Prepend distribution artifacts for production site deployment (under docs folder)
npm run docs:build

🌟 Explore AI Projects

Please use the navigation bar or sidebar of this site to browse this oasis of open-source code meticulously curated for you by our AI agent. Our crawlers and scoring models regularly sweep GitHub for trending projects, keeping this directory fresh!

📚 View Online: https://hello-ai.anzz.top

Discussion Groups

AI chat groups, some of which offer direct chat experiences with AI, and a platform to exchange ideas with like-minded individuals.

Join Telegram Group Join WeChat Group (Note: Join AI group)

Please specify your intent to join the WeChat group to avoid unwanted invitations and information harassment.
Telegram group link: https://t.me/auto_gpt_ai

trackgit-views

About

It's not AI that takes away your job, but the people who master the use of AI tools. The most deadly attack is a dimension-reducing strike: destroying you has nothing to do with you - from "The Three-Body Problem". 中文说明: 抢走你工作的不是AI,而是掌握使用AI工具的人。 降维打击最为致命:毁灭你,与你何干《三体》

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors