Kai Hallett

Software engineer. Learning in public about agentic systems, human-AI interaction, and what happens when you give LLMs real responsibilities.


Recent

all posts →

Research notes

LLM Verification Phenomena

What happens when LLM-generated tests pass for the wrong reason, automated reviewers find different bugs on every pass, and the fix-review loop refuses to converge. A survey of what academia knows, what it doesn't, and where our daily practice sits in the gap.

LLM-produced · not independently verified · starting material only

Metacognitive Analysis

Three-stage clinical psychology analysis of the reflective functioning that enabled each pushback moment. Initial assessment: RF 6-7. Recalibrated to RF 6-9 after disclosure of 20 years meditation practice including 15 Zen sesshin. Three peaks at RF 8-9.

LLM-produced · not independently verified · starting material only

Prospective Metacognitive Regulation

The engineering work IS the prospective regulation — not post-hoc analysis of it. The governance systems map onto distributed cognition research in ways I didn't plan. Hutchins' foundational work was literally a study of naval navigation.

LLM-produced · not independently verified · starting material only

The Fight Card

16 rounds where the human pushed back against sycophantic drift across 208 session decisions. Mapped from the actual decision record — not reconstructed from memory. Key finding: human won every round by being honest when the model couldn't be.

LLM-produced · not independently verified · starting material only