
Personal LLM 2.0
100% offline private AI assistant running 28 open-source LLMs on your hardware. Desktop + Android + Web — no cloud, no Docker, no Ollama.
About the Project
A fully offline private AI assistant that loads open-source Large Language Models directly into your PC's memory. Features 28 pre-configured GGUF models (Phi-4, DeepSeek-R1, Llama 3.2, Qwen 3, CodeLlama, etc.), multi-turn chat with persistent JSON history, RAG document Q&A via ChromaDB + sentence-transformers, 4-layer Context Intelligence Engine (RAG → Recursive Decomposition → Self-Refine → Chain-of-Thought), cloud proxy for mixing local and cloud AI, and a hot model swap system. Full-stack monorepo spanning Python (FastAPI backend), React (Electron desktop), React Native (Expo Android app), and Next.js (marketing site). All data stays on your machine — zero telemetry, zero analytics, zero cloud.
Tech Stack
Related Projects
The Grand Horizon
Luxury boutique hotel template featuring a sophisticated dark slate aesthetic, gold accents, and a timeless sanctuary theme.
Poetry Hub
A modern full-stack web application for poetry enthusiasts with AI-powered daily poetry generation.
DINE24 AI System
A comprehensive AI-powered restaurant management system featuring smart ordering and integrated payments.