Agent Teams Eval: comparing Claude Code Agent Teams vs single-agent for feature implementation on LangGraph. Ceiling effect, 3.6x speedup, zero peer communication.
-
Updated
Mar 10, 2026 - Python
Agent Teams Eval: comparing Claude Code Agent Teams vs single-agent for feature implementation on LangGraph. Ceiling effect, 3.6x speedup, zero peer communication.
An independent study on AI interiority, documenting two experiments that probe the line between knowledge and consciousness in large language models.
Agent Teams Eval: comparing Claude Code Agent Teams vs subagents for architecture design. First significant result — Agent Teams advantage d=+0.99, p=0.014.
Exploratory compact LM research branch derived from LUMI-Arch, focused on structure-sensitive generalization under tight model constraints.
Agent Teams Eval: comparing Claude Code Agent Teams vs subagents for bug-fixing on Ruff. Ceiling effect — 8/8 solve rate, zero peer communication.
Add a description, image, and links to the experimental-research topic page so that developers can more easily learn about it.
To associate your repository with the experimental-research topic, visit your repo's landing page and select "manage topics."