RL Know vs RL Do: Why AI Learned to Act
AI training has three stages: pre-training builds capability, RL Know refines judgment, RL Do trains action. That last one is why AI feels different now.
Writing on agentic engineering, AI coding workflows, and building software with multi-agent systems.
AI training has three stages: pre-training builds capability, RL Know refines judgment, RL Do trains action. That last one is why AI feels different now.
SkillsBench tested 7,308 agent runs. The data confirms: good engineering -- focused skills -- consistently improves agent outcomes.
The full synthesis: context rot, DRYP, skills vs. agents, and why good practice -- not grand design -- is how you get to 100+.
Applying the DRY principle to AI agent instructions: why duplicated prompts are a refactoring opportunity.
Five non-negotiable rules for moving from vibe coding to real AI agentic software engineering.
The framework that made 70+ AI agents manageable: the distinction between skills and agents.
Why stuffing AI with more context makes results worse, and how a skills-specialist agent fixes it.