Private On-Device RAG for Personal Knowledge That Runs Natively on Mobile
People want to chat with their personal documents (PDFs, notes, health records, financial docs) using AI without uploading anything to the cloud. Desktop solutions exist (Reor, AnythingLLM, Obsidian+Ollama) but mobile is severely underserved. The few mobile options are either just API wrappers to cloud models or require connecting to a home server. A truly on-device mobile RAG app with local inference doesn't exist yet.
The hardware is finally ready. Flagship phones can run Phi-3-mini at usable speeds. The app needs three things: (1) dead-simple document import from camera/files/share sheet, (2) local embedding + vector store on device, (3) a chat UI that cites which document passages it's drawing from. Skip multi-model support at launch. Pick one model, make it fast, and nail the UX.
landscape (4 existing solutions)
Desktop private RAG is a solved problem (Reor, AnythingLLM, Obsidian+Ollama). Mobile private RAG is not. The existing mobile options either require a home server connection or are proof-of-concept quality. Modern phones (Snapdragon 8 Gen 3, Apple A17 Pro) can run 3-7B models at usable speeds, but nobody has built a polished mobile app that combines document ingestion, local embedding, local inference, and a good chat UI into one package.