We read all 512K lines of Claude Code's accidentally exposed source. 82 docs, 15 diagrams, every subsystem mapped — from the hidden YOLO safety classifier to multi-agent swarms.
-
Updated
Apr 2, 2026
We read all 512K lines of Claude Code's accidentally exposed source. 82 docs, 15 diagrams, every subsystem mapped — from the hidden YOLO safety classifier to multi-agent swarms.
Mechanistic interpretability of transformer hallucinations via attention flow, residual stream geometry, and head-level attribution analysis.
A reference point for phenomena that have been reported to occur inside AI systems but have no direct mapping into natural language.
Add a description, image, and links to the llm-internals topic page so that developers can more easily learn about it.
To associate your repository with the llm-internals topic, visit your repo's landing page and select "manage topics."