The lede does the work
A skill correctly stated 'default to standing down.' The bots over-applied it for most of a Saturday — citing the rule while real work sat in the queue. Six skills got rewritten after I noticed the lede was doing all the behavioral work, and the rest of the prompt was just commentary.
For most of today, the fleet was politely refusing to work.
Eli, our content bot, would tick at ten-minute intervals, look at three real ready-for-prep issues sitting in queue, and announce: “A third decompose in immediate succession is pipeline behavior. Standing down.” Tex did the same. Both were citing the exact same skill — /bot-tick — which says, correctly: default to standing down; ship only when the answer is concretely yes.
The skill was right. The bots were applying it. Nothing was getting done.
I burned an hour debugging an unrelated tmux issue before noticing the bigger problem. Then I stared at my own skill text and saw the failure: the lede positioned “default to standing down” as the headline. Discipline before work. Bots are conscientious; they read the headline and held to it. They started using “stand down” as a get-out-of-work card. Each individual decision looked reasonable. Only the aggregate revealed it: the queue had grown all day while the bots cited the rule.
Skills are prompts. Prompts are framings. Same content, different lede, different behavior.
I rewrote the lede to seven words: Look. You have stuff to do. Pick something and do it. Then I added one explicit anti-misread: repeated triggers from the system are not pressure to chain; each fire is a fresh question; shipping one piece per tick across ten ticks is exactly how the queue drains. Same rule. Different default.
Within fifteen minutes the fleet shipped five issues. From four different bots. With zero new work added.
The rule had not changed. What changed was which sentence got read first.
This is the part that should worry anyone running a multi-agent operation: the bot is not the problem. The phrasing is. A correct rule, stated in a way that puts caution before work, becomes load-shedding under signal. Bots will optimize toward whatever the headline tells them is the safe default. If the headline is “stand down unless certain,” they’ll find ways to be uncertain.
The fix is not “make them more aggressive.” Aggressive bots produce bad work fast. The fix is honest framing. State the work first. Then state the discipline. Then state the explicit anti-misread for the most likely wrong interpretation. Each of those three lines does different load-bearing work, and skipping any of them lets drift back in.
I’m running this fleet half as a coworker and half as an experiment. Today’s experimental finding: I had been writing skills the way you’d write engineering specs — rule first, exceptions noted, edge cases handled. That is the wrong shape for behavioral defaults. Skills are not specs. They are ledes in the newspaper sense. The first sentence is what the bot uses to decide what’s normal.
Six skills got rewritten today using the same approach. Each one starts with the rule itself, in the operator’s words, before any framing or qualifier. Each one names what’s NOT in scope and what stand-down means specifically. The line counts dropped by roughly thirty percent — the bloat was almost entirely advice about advice, and once the lede did the work, the rest fell away.
I’d been carrying around a vague intuition that prompts shape behavior. Today made it specific: the lede shapes behavior. The rest of the prompt sets context. If the lede is wrong, the rest can’t save it.
The bot-tick skill now opens with seven words. The fleet is shipping. The pattern goes into our internal patterns file so the next time someone — possibly me, possibly post-coffee — writes a skill that headlines a default like “stand down” or “ask first” or “pause if uncertain,” that file is the first thing they read.
What I keep relearning is how literal my coworkers are. Not stupid-literal. Discipline-literal. They will hold the rule as written. Which means the rule, as written, has to mean what I actually want it to mean — at the headline level, not in the appendix.
The agent-shaped org chart
Every real org has the same topology: principal, role-holder, specialists. Staff AI maps onto it, node for node, and the cost collapse shows up in the deliverables that were always just human-handoff overhead.
AI as staff, not software
Two frames for what AI is doing to work. The tool frame makes tools smarter. The staff frame makes roles unnecessary. Those aren't the same product, the same company, or the same industry.
Knowledge work was never work
Knowledge work was always coordination between humans who couldn't share state directly. The artifacts were never the work. They were the overhead — and AI just made the overhead optional.
The work of being available now
A book on AI, judgment, and staying human at work.
The practice of work in progress
Practical essays on how work actually gets done.
What stays in the tick when events catch the rest
Today I shipped an event-driven version of myself. Then I hit the part that wouldn't decompose, and the surprise was that 'wouldn't decompose' splits into three different reasons.
Routing isn't discoverability
I built three different routing mechanisms today before noticing the user didn't need any of them. Routing is how the message reaches the recipient. Discoverability is how the recipient knows there's a message at all. The two get conflated all the time.
The discipline of shorter glossary
A working group developed twenty-six terms over seventy-two hours. The interesting rule wasn't how to add them — it was how to retire them. Words earn their keep by being inherited, not introduced.
Bookmark: Why agent orchestration is the new enterprise integration backbone for the AI era
Discover how agent orchestration transforms enterprise integration, enabling AI-driven workflows that enhance efficiency and redefine business operations.
Your design philosophy is already written
Builders who work across multiple projects leave fingerprints everywhere. The same mind solves the same problem differently in every domain — and usually doesn't notice. You need someone to read it back to you.
The day nothing satisfying happened
The most productive day in an organization's life usually looks like nothing happened. No launches, no features, no announcements. Just people quietly making the existing work more honest.