- Installs MCP config to ~/.claude/mcp.json (merges with existing)
- Appends knowledge graph instructions to ~/.claude/CLAUDE.md
- Run once, works in any repo after that
- LLM_BACKEND=openai routes to /v1/chat/completions
- Default: ollama (unchanged)
- For Kiro gateway: LLM_BACKEND=openai OPENAI_URL=http://192.168.86.11:8000 OPENAI_MODEL=claude-haiku-4
- Updated README with new env vars
- Schema versioned with migrations table
- watch.sh runs db migrations on every pull
- watch.sh uses uv run/sync instead of pip
- Fixed parser.py reference in watch.sh
- parser.py renamed to go_parser.py (avoids Python builtin conflict)
- docgen.py was missing from flat structure
- Added pyproject.toml for uv
- Updated .mcp.json to use uv run
- Updated README for uv workflow
- SQLite backend with file/repo/relationship entities
- tree-sitter Go AST parser for deterministic import detection
- Ollama doc generation with concurrent batch processing
- MCP server (FastMCP) for Claude Code integration
- Merge simulation with staleness cascade
- Lazy refresh for stale relationship and repo docs
- CLAUDE.md for agent context