3 Commits

Author SHA1 Message Date
92cb600dd6 Fix LLM response formatting and Windows installer robustness
- Preserve whitespace and newlines in streaming responses
- Clean thinking tags from final LLM responses
- Add lazy initialization to _call_ollama method
- Improve Windows installer to handle existing virtual environments
- Add better error reporting for import failures

These fixes address formatting corruption in numbered lists and
improve installer reliability when dependencies already exist.
2025-08-15 14:26:53 +10:00
a84ff94fba Improve UX with streaming tokens, fix model references, and add icon integration
This comprehensive update enhances user experience with several key improvements:

## Enhanced Streaming & Thinking Display
- Implement real-time streaming with gray thinking tokens that collapse after completion
- Fix thinking token redisplay bug with proper content filtering
- Add clear "AI Response:" headers to separate thinking from responses
- Enable streaming by default for better user engagement
- Keep thinking visible for exploration, collapse only for suggested questions

## Natural Conversation Responses
- Convert clunky JSON exploration responses to natural, conversational format
- Improve exploration prompts for friendly, colleague-style interactions
- Update summary generation with better context handling
- Eliminate double response display issues

## Model Reference Updates
- Remove all llama3.2 references in favor of qwen3 models
- Fix non-existent qwen3:3b references, replace with proper model names
- Update model rankings to prioritize working qwen models across all components
- Ensure consistent model recommendations in docs and examples

## Cross-Platform Icon Integration
- Add desktop icon setup to Linux installer with .desktop entry
- Add Windows shortcuts for desktop and Start Menu integration
- Improve installer user experience with visual branding

## Configuration & Navigation Fixes
- Fix "0" option in configuration menu to properly go back
- Improve configuration menu user-friendliness
- Update troubleshooting guides with correct model suggestions

These changes significantly improve the beginner experience while maintaining
technical accuracy and system reliability.
2025-08-15 12:20:06 +10:00
cc99edde79 Add comprehensive Windows compatibility and enhanced LLM setup
- Add Windows installer (install_windows.bat) and launcher (rag.bat)
- Enhance both Linux and Windows installers with intelligent Qwen3 model detection and setup
- Fix installation script continuation issues and improve user guidance
- Update README with side-by-side Linux/Windows commands
- Auto-save model preferences to config.yaml for consistent experience

Makes FSS-Mini-RAG fully cross-platform with zero-friction Windows adoption 🚀
2025-08-15 10:52:44 +10:00