BobAi ba28246178 Add LLM synthesis feature with smart model selection and increase default results to 10
🧠 NEW: LLM Synthesis Feature
- Intelligent analysis of RAG search results using Ollama LLMs
- Smart model selection: Qwen3 → Qwen2.5 → Mistral → Llama3.2
- Prioritizes efficient models (1.5B-3B parameters) for best performance
- Structured output: summary, key findings, code patterns, suggested actions
- Confidence scoring for result reliability
- Graceful fallback with setup instructions if Ollama unavailable

📊 Enhanced Search Experience
- Increased default search results from 5 to 10 across all components
- Updated demo script to show all 8 results with richer previews
- Better user experience with more comprehensive result sets

🎯 New CLI Options
- Added --synthesize/-s flag: rag-mini search project "query" --synthesize
- Zero-configuration setup - automatically detects best available model
- Never downloads models - only uses what's already installed

🧪 Tested with qwen3:1.7b
- Confirmed excellent performance with 1.7B parameter model
- Professional-grade analysis including security recommendations
- Fast response times with quality RAG context

Perfect for users who already have Ollama - transforms FSS-Mini-RAG
from search tool into AI-powered code assistant\!
2025-08-12 17:12:51 +10:00

FSS-Mini-RAG

A lightweight, educational RAG system that actually works
Built for beginners who want results, and developers who want to understand how RAG really works

FSS-Mini-RAG Icon

How It Works

graph LR
    Files[📁 Your Code] --> Index[🔍 Index]
    Index --> Chunks[✂️ Smart Chunks]
    Chunks --> Embeddings[🧠 Semantic Vectors]
    Embeddings --> Database[(💾 Vector DB)]
    
    Query[❓ "user auth"] --> Search[🎯 Hybrid Search]
    Database --> Search
    Search --> Results[📋 Ranked Results]
    
    style Files fill:#e3f2fd
    style Results fill:#e8f5e8
    style Database fill:#fff3e0

What This Is

FSS-Mini-RAG is a distilled, lightweight implementation of a production-quality RAG (Retrieval Augmented Generation) search system. Born from 2 years of building, refining, and tuning RAG systems - from enterprise-scale solutions handling 14,000 queries/second to lightweight implementations that anyone can install and understand.

The Problem This Solves: Most RAG implementations are either too simple (poor results) or too complex (impossible to understand and modify). This bridges that gap.

Quick Start (2 Minutes)

# 1. Install everything
./install_mini_rag.sh

# 2. Start using it
./rag-tui                    # Friendly interface for beginners
# OR
./rag-mini index ~/my-project     # Direct CLI for developers
./rag-mini search ~/my-project "authentication logic"

That's it. No external dependencies, no configuration required, no PhD in computer science needed.

What Makes This Different

For Beginners

  • Just works - Zero configuration required
  • Multiple interfaces - TUI for learning, CLI for speed
  • Educational - Shows you CLI commands as you use the TUI
  • Solid results - Finds code by meaning, not just keywords

For Developers

  • Hackable - Clean, documented code you can actually modify
  • Configurable - YAML config for everything, or change the code directly
  • Multiple embedding options - Ollama, ML models, or hash-based
  • Production patterns - Streaming, batching, error handling, monitoring

For Learning

  • Complete technical documentation - How chunking, embedding, and search actually work
  • Educational tests - See the system in action with real examples
  • No magic - Every decision explained, every component documented

Usage Examples

Find Code by Concept

./rag-mini search ~/project "user authentication"
# Finds: login functions, auth middleware, session handling, password validation

Natural Language Queries

./rag-mini search ~/project "error handling for database connections"
# Finds: try/catch blocks, connection pool error handlers, retry logic

Development Workflow

./rag-mini index ~/new-project              # Index once
./rag-mini search ~/new-project "API endpoints"   # Search as needed
./rag-mini status ~/new-project            # Check index health

Installation Options

./install_mini_rag.sh
# Handles Python setup, dependencies, optional AI models

Experimental: Copy & Run (May Not Work)

# Copy folder anywhere and try to run directly
./rag-mini index ~/my-project
# Auto-setup will attempt to create environment
# Falls back with clear instructions if it fails

Manual Setup

python3 -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt

Note: The experimental copy & run feature is provided for convenience but may fail on some systems. If you encounter issues, use the full installer for reliable setup.

System Requirements

  • Python 3.8+ (installer checks and guides setup)
  • Optional: Ollama (for best search quality - installer helps set up)
  • Fallback: Works without external dependencies (uses built-in embeddings)

Project Philosophy

This implementation prioritizes:

  1. Educational Value - You can understand and modify every part
  2. Practical Results - Actually finds relevant code, not just keyword matches
  3. Zero Friction - Works out of the box, configurable when needed
  4. Real-world Patterns - Production techniques in beginner-friendly code

What's Inside

  • Hybrid embedding system - Ollama → ML → Hash fallbacks
  • Smart chunking - Language-aware code parsing
  • Vector + keyword search - Best of both worlds
  • Streaming architecture - Handles large codebases efficiently
  • Multiple interfaces - TUI, CLI, Python API, server mode

Next Steps

  • New users: Run ./rag-mini for guided experience
  • Developers: Read TECHNICAL_GUIDE.md for implementation details
  • Contributors: See CONTRIBUTING.md for development setup

Documentation

License

MIT - Use it, learn from it, build on it.


Built by someone who got frustrated with RAG implementations that were either too simple to be useful or too complex to understand. This is the system I wish I'd found when I started.

Description
No description provided
Readme MIT 1.3 MiB
Languages
Python 85.3%
Shell 8.7%
PowerShell 4%
Batchfile 1.8%
Makefile 0.2%