Fss-Rag-Mini/.mini-rag/config.yaml
FSSCoding 8e67c76c6d Fix model visibility and config transparency for users
CRITICAL UX FIXES for beginners:

Model Display Issues Fixed:
- TUI now shows ACTUAL configured model, not hardcoded model
- CLI status command shows configured vs actual model with mismatch warnings
- Both TUI and CLI use identical model selection logic (no more inconsistency)

Config File Visibility Improved:
- Config file location prominently displayed in TUI configuration menu
- CLI status shows exact config file path (.mini-rag/config.yaml)
- Added clear documentation in config file header about model settings
- Users can now easily find and edit YAML file for direct configuration

User Trust Restored:
-  Shows 'Using configured: qwen3:1.7b' when config matches reality
- ⚠️ Shows 'Model mismatch!' when config differs from actual
- Config changes now immediately visible in status displays

No more 'I changed the config but nothing happened' confusion!
2025-08-15 22:17:08 +10:00

66 lines
2.2 KiB
YAML

# FSS-Mini-RAG Configuration
#
# 🔧 EDIT THIS FILE TO CUSTOMIZE YOUR RAG SYSTEM
#
# This file controls all behavior of your Mini-RAG system.
# Changes take effect immediately - no restart needed!
#
# 💡 IMPORTANT: To change the AI model, edit the 'synthesis_model' line below
#
# Common model options:
# synthesis_model: auto # Let system choose best available
# synthesis_model: qwen3:0.6b # Ultra-fast (500MB)
# synthesis_model: qwen3:1.7b # Balanced (1.4GB) - recommended
# synthesis_model: qwen3:4b # High quality (2.5GB)
#
# See docs/GETTING_STARTED.md for detailed explanations
# Text chunking settings
chunking:
max_size: 2000 # Maximum characters per chunk
min_size: 150 # Minimum characters per chunk
strategy: semantic # 'semantic' (language-aware) or 'fixed'
# Large file streaming settings
streaming:
enabled: true
threshold_bytes: 1048576 # Files larger than this use streaming (1MB)
# File processing settings
files:
min_file_size: 50 # Skip files smaller than this
exclude_patterns:
- "node_modules/**"
- ".git/**"
- "__pycache__/**"
- "*.pyc"
- ".venv/**"
- "venv/**"
- "build/**"
- "dist/**"
include_patterns:
- "**/*" # Include all files by default
# Embedding generation settings
embedding:
preferred_method: ollama # 'ollama', 'ml', 'hash', or 'auto'
ollama_model: nomic-embed-text
ollama_host: localhost:11434
ml_model: sentence-transformers/all-MiniLM-L6-v2
batch_size: 32 # Embeddings processed per batch
# Search behavior settings
search:
default_top_k: 10 # Default number of top results
enable_bm25: true # Enable keyword matching boost
similarity_threshold: 0.1 # Minimum similarity score
expand_queries: false # Enable automatic query expansion
# LLM synthesis and query expansion settings
llm:
ollama_host: localhost:11434
synthesis_model: qwen3:1.7b # 'auto', 'qwen3:1.7b', etc.
expansion_model: auto # Usually same as synthesis_model
max_expansion_terms: 8 # Maximum terms to add to queries
enable_synthesis: false # Enable synthesis by default
synthesis_temperature: 0.3 # LLM temperature for analysis