bartkolendowski.com writing 2026-04-15 · 12 min
Language Models: Patterns of Engagement

Language Models: Patterns of Engagement

Treating AI like a human is costing you.

April 15, 2026 · ~12 minute read · 4,500 words

It is natural to talk to a large language model the way one talks to a colleague. The interface invites it. A chat box, a cursor, a turn-by-turn rhythm familiar from every conversation one has ever had. A request goes in, a response comes back, a refinement is offered, a refinement is delivered. The flow is legible, comfortable, and it leaves most of the available leverage on the table.

The instinct to treat the model as a collaborator is not a mistake of ignorance. It is a mistake of transfer. A lifetime of working with other people has produced a carefully tuned set of intuitions about how to get the best work out of them: when to clarify, when to push back, when to refine, when to commit. Those intuitions are calibrated to a specific cost structure. Human time and attention are scarce and expensive, so they are conserved. Large language models invert that cost structure completely. Almost none of the instincts that survive in human collaboration survive in model collaboration intact.

What follows is a framework for where the inherited instincts go wrong and what the alternatives look like. Five asymmetries between managing people and managing models. Three anti-patterns that emerge when the asymmetries are over-applied. One meta-principle that ties the rest together.

The Core Inversion

With a human collaborator, the bottleneck is their time and attention. Both are scarce, both are expensive, and everything about good management is built around conserving them. Direction is clarified before a request is made. Feedback is batched. One approach is chosen and committed to. The effort spent on specification is a direct tax on someone else's day, so it is minimized through decisiveness.

With a language model, the bottleneck shifts. Generation is effectively free. What remains expensive is the operator's evaluation bandwidth: the ability to judge output, notice what is wrong, and decide what comes next. This single inversion rewires almost every rule that follows from it.

The optimal human role shifts from producer to evaluator. The strategy is no longer to guide the model toward one good answer. It is to have the model cover the solution space efficiently and to use scarce human judgment to navigate it. The operator is not a worker directing another worker. The operator is an evaluation function steering a search process.

Anywhere the instinct is to conserve model effort the way one would conserve a person's effort, leverage is being left on the table.

Five Asymmetries

1. Generate, don't iterate.

The default human workflow is a greedy walk. Commit to one direction, refine, refine again. With a human collaborator, this is rational. Production is expensive, so the cost of exploring five parallel branches is five times what anyone can afford.

A language model supports a fundamentally different strategy: beam search over the solution space. Generate N variants in parallel. Pick the best. Generate variants of the parts that are still weak. Compose the result from the strongest pieces of each pass. The operator is not climbing one hill. The operator is surveying a landscape.

The approach applies recursively and across every abstraction level. Five architectural sketches, then three refinements of each of the top two. Five tones for the same paragraph, then recombine. Five hypotheses for a debugging problem, then test the most promising three simultaneously. Volume-plus-selection reliably outperforms single-pass refinement when generation is free and evaluation is not.

2. Radical candor has zero cost.

Feedback between humans is a social negotiation. Criticism is softened, sandwiched, rationed. "This is wrong, start over" damages working relationships, so operators spend real effort avoiding the situations that would force them to say it. Every hedge word in a prompt is a small tax paid to preserve a relationship that, in the model case, does not exist.

A language model has no ego, no morale, no political standing. It cannot be demoralized, cannot take offense, cannot carry a slight into the next conversation. The cost of candor is zero.

Most operators are paying the tax anyway. Prompts arrive in the softened register that a lifetime of human collaboration has made automatic: "Could you maybe try...", "I was wondering if...", "Would it be possible to..." Each hedge is wasted tokens and wasted ambiguity. The optimal prompt looks more like a specification document than a message to a coworker. Direct, dense, constraint-heavy, without apology. The shift feels rude, which is exactly why most operators do not make it, which is exactly why making it is a differentiator.

3. Context is manipulable.

Human memory is immutable and singular. A colleague's understanding of a project cannot be reset, cannot be forked at a decision point, cannot be given selective amnesia. Whatever they carry, they carry.

Model context is data. It can be copied, branched, discarded, truncated, and curated at will. That single property enables three operations with no analogue in human collaboration.

Reset is the discard operation. When accumulated context becomes a liability (anchoring bias, sunk-cost thinking, path dependency), a fresh session produces a reviewer free of the commitments the building session made.

Fork is the branch operation. When accumulated context is an asset (expensive shared understanding that should not be rebuilt), a decision point becomes an opportunity to split into two parallel explorations, both carrying the same history forward, both capable of running to completion, with reconvergence only when one approach wins.

Curation is the selective operation. When some of the context helps and some of it hurts, the helpful portion can be preserved while the harmful portion is stripped. "Keep the constraints from earlier, drop the proposed architecture, start fresh on that particular question."

For a human manager, every decision point is a commitment point. For an operator running language models, every decision point is a branching opportunity. Most operators have not yet adjusted.

4. Adversarial self-play is free.

A human cannot be asked to genuinely try to break their own work. They have ego, they have blind spots from the context in which they built the thing, and asking them to attack what they have just built feels, to them, like an attack on themselves. The best organizations handle this by hiring dedicated reviewers, red teams, and devil's advocates. It is an expense that only larger operations can justify.

With language models, one session builds and a fresh session attacks. The red-teaming session has no ego in the outcome and no accumulated commitment to the solution. It can interrogate the work with a freedom that no human reviewer, including the original builder, can match.

The pattern extends far beyond code review. A proposal can be submitted to a fresh context that is instructed to argue against it as forcefully as possible. A system design can be handed to a fresh context whose entire job is to find what is broken. A business decision can be role-played against by a fresh context impersonating the stakeholder who thinks it is the wrong call.

This amounts to organizational capability that almost no small team has historically been able to afford: structured adversarial review on every important output. The cost has collapsed to the price of a second conversation window. Very few operators are using it.

5. Context documents are infrastructure, not documentation.

In human organizations, written documentation is a nice-to-have that almost nobody reads. The return on investment is unclear, the incentive to produce it is weak, and what gets written tends to rot faster than it gets updated. Documentation is treated as overhead because, in a human-only workflow, it essentially is.

In a language-model-augmented workflow, context documents become the memory substrate. Every session starts from zero, and every session also starts from whatever has been written down. The operator who maintains the best context documents has the most capable model collaborator, by a margin that is almost never small.

This inverts the economics entirely. Session logs, architectural decision records, notes explaining why one option was chosen over another: work that used to feel like overhead becomes compound interest. Every hour spent producing high-quality context multiplies across every future session that loads it. The ROI is no longer unclear. It is arguably the highest-leverage time an operator can spend.

Context documents should be treated the way a database schema is treated. Versioned, maintained, refactored as the underlying work evolves. Treating them as disposable is the expensive habit inherited from a world in which nobody read them.

Three Anti-Patterns

Every asymmetry that grants leverage creates a corresponding way to waste it. The failure modes deserve at least as much attention as the wins, because they are subtle. The output looks productive. Sometimes it is not.

1. The illusion of progress.

The trap is an over-application of the first asymmetry: generation is cheap, so produce more. The operator generates five variants, picks the best, generates three refinements, picks again, and at the end of an hour holds something polished that does not solve the right problem. With a human collaborator, the slowness of production would have forced a natural checkpoint. Before three days are spent on this, is the target correct? With a model, that checkpoint is absent, and the ease of iteration becomes a momentum that masks the absence of clarity. The operator is hill-climbing on the wrong hill, very efficiently.

The defense is a discipline applied before any generation begins. A short written statement, for the operator alone, of what "done" looks like and how it will be evaluated. If that statement cannot be produced in two sentences, the operator is not yet ready to generate.

2. Fluency as correctness.

The trap is an over-application of the ability to inspect model reasoning. A detailed chain of reasoning ("I considered A, B, and C. A fails because X. B is suboptimal because Y. Therefore C.") looks like rigorous analysis. It may be post-hoc rationalization of a pattern-matched conclusion: a plausible justification for the answer the model was going to produce regardless of the reasoning displayed.

This is the most dangerous anti-pattern because it directly corrupts the evaluation function. Reasoning transparency was requested in order to improve oversight. If the transparency is itself unreliable, the result is more confidence in answers that may be wrong. A layer of persuasion has been installed where a layer of verification was intended.

The defense is to verify artifacts empirically, not argumentatively. Stated reasoning is a checklist of claims to test, not evidence of correctness. If the model claims its output handles edge case X, the operator tests edge case X directly. Develop, test, fix: the loop that protects against this anti-pattern is the same loop that protects against most of the others.

3. Artifact drift.

This is the least-discussed anti-pattern in sustained model work and one of the most expensive. The operator uses a model to produce artifact A, a strategy document, say. Later, the operator works intensively on artifact B, the materials that implement that strategy. During the work on B, real decisions get made: pricing changes, positioning refinements, messaging shifts. The decisions are correct. But artifact A does not update. When the operator returns to A, it is silently stale. Not because of error, but because the model does not maintain coherence across independent sessions.

A human who writes the strategy remembers the numbers when they build the materials. They hold a mental model spanning every artifact they have produced, and coherence is maintained in their head as a side effect of working. With language models, each generation context is independent. The operator is the only integration point, and the operator's working memory has limits that the model's output volume does not respect.

The trap compounds with the first asymmetry. Cheap generation encourages more artifacts. More artifacts demand more cross-artifact coherence. Volume and coherence pull in opposite directions, and without explicit defense, coherence loses.

The defense is to treat source-of-truth documents as infrastructure that must be updated as part of any downstream work. When a decision during execution affects an upstream artifact, the upstream artifact is updated immediately or it is never updated at all. Every artifact decision must be treated as a potentially multi-artifact decision, or cross-artifact coherence decays silently and then, eventually, all at once.

The Meta-Principle

The mental shift required is from managing a worker to managing a search process. The operator is not guiding one path to one good outcome. The operator is exploring a solution space and using judgment to navigate it. The model is the search engine. The operator is the evaluation function.

This reframes the scarce resource. It is not generation. It is attention and judgment. The optimal strategy, across every asymmetry and anti-pattern above, reduces to a single directive: maximize the information content per unit of evaluation time. Sometimes that means five variants. Sometimes a structured comparison. Sometimes a fresh-context red team. Sometimes pasting everything into context and asking what breaks first.

The uncomfortable implication is that as the tools become more capable, the limiting factor becomes the operator's ability to direct them well. Every asymmetry that grants leverage also creates a new way to waste it. The tools multiply bad judgment as readily as good judgment, and the gap between skilled and unskilled operators will widen faster than the gap between old and new models.

Three disciplines protect against the waste.

Know what you want before you generate.

Verify outputs empirically rather than argumentatively.

Maintain the discipline to stop.

The rest is execution.

← back to bartkolendowski.com writing · 2026-04-15