LLM Wiki llmwiki
winget install --id=nashsu.LLMWiki -e LLM Wiki is a cross-platform desktop application that turns your documents into an organized, interlinked knowledge base — automatically. Instead of traditional RAG (retrieve-and-answer from scratch every time), the LLM incrementally builds and maintains a persistent wiki from your sources. Knowledge is compiled once and kept current, not re-derived on every query. This project is based on Karpathy's LLM Wiki pattern — a methodology for building personal knowledge bases using LLMs. We implemented the core ideas as a full desktop application with significant enhancements. Features - Two-Step Chain-of-Thought Ingest — LLM analyzes first, then generates wiki pages with source traceability and incremental cache - 4-Signal Knowledge Graph — relevance model with direct links, source overlap, Adamic-Adar, and type affinity - Louvain Community Detection — automatic knowledge cluster discovery with cohesion scoring - Graph Insights — surprising connections and knowledge gaps with one-click Deep Research - Vector Semantic Search — optional embedding-based retrieval via LanceDB, supports any OpenAI-compatible endpoint - Persistent Ingest Queue — serial processing with crash recovery, cancel, retry, and progress visualization - Folder Import — recursive folder import preserving directory structure, folder context as LLM classification hint - Deep Research — LLM-optimized search topics, multi-query web search, auto-ingest results into wiki - Async Review System — LLM flags items for human judgment, predefined actions, pre-generated search queries - Chrome Web Clipper — one-click web page capture with auto-ingest into knowledge base