Specialized AI Agents: The Ancient Art of Knowing Less
The sculptor's secret: David was always in the marble. The genius was knowing what to remove

Specialized AI Agents: The Ancient Art of Knowing Less

When you create a 'QA engineer' agent, you're not loading specialized capabilities. You're asking a generalist to forget everything except what matters. The sculptor knew this. Now the engineer must learn it.

You did not hire a QA engineer.

When you spun up that agent with the prompt "You are a senior QA engineer focused on edge cases and security vulnerabilities," you did not instantiate a specialist. You asked a generalist to forget everything except what a QA engineer would notice.

Same model. Same weights. Same capabilities. Different mask.

This distinction matters more than most teams realize. The sculptor knew this. The actor knew this. Now the engineer must learn it: specialization has never been about acquisition. It has always been about removal.

The Paradox of the Empty Stage

There is a story, perhaps apocryphal but certainly true, about Michelangelo explaining his David: "I simply removed everything that wasn't David." The block of marble contained infinite potential figures. The genius was not in adding David but in having the discipline to subtract everything else.

Daniel Day-Lewis practices the same ancient craft. When he becomes Abraham Lincoln, he does not download historical knowledge or install presidential capabilities. What he does is far more radical: he eliminates Daniel Day-Lewis. For months, he refuses to break character. He insists the crew address him as Mr. President. He does not add Lincoln; he subtracts everything that is not Lincoln.

This is precisely what happens when you give an LLM a role prompt.

The model's attention mechanism works like a spotlight on a vast stage. The entire theater exists in darkness: every capability, every pattern, every learned association from training. When you prompt the model to behave as a "senior security engineer," you are not building new scenery. You are positioning the spotlight. The security patterns illuminate. The irrelevant dims.

Same stage. Different light. The oldest trick in the theatrical book, now running on silicon.

What the Stoics Knew About Transformers

A mentor once told me something that reframed how I think about specialization: "Focus does not mean increased attention, or trying harder. It is the elimination of things outside the subject you want to focus on. Focus is not additive. It is a reductive exercise."

You have experienced this. In a crowded restaurant, when you truly focus on what someone is saying, you do not hear better. You become selectively deaf. The clatter of dishes fades. The conversation two tables over dissolves into murmur. You did not gain the power to hear your companion more clearly. You lost the ability to hear anything else.

Marcus Aurelius would have nodded. The Stoics spent considerable effort learning what not to attend to. They understood that wisdom is less about knowing more and more about ignoring well.

When a transformer processes a specialized prompt, the attention weights do not increase in magnitude. They redistribute. Tokens related to the specified role receive higher attention scores. Tokens outside that domain receive lower scores. The model does not try harder; it ignores more specifically.

The mathematics of attention is, quite literally, a focusing operation. And focusing is always subtraction wearing the costume of addition.

The Mask That Reveals

Here is where the parallel deepens into something like philosophy.

We moderns have a confused relationship with masks. We think they conceal. But the Greeks knew better. The word "persona" comes from the masks actors wore; the mask did not hide the actor but revealed the character. The constraint created the clarity.

Large language models contain multitudes. They have ingested the writing patterns of security experts, the reasoning structures of architects, the debugging instincts of senior developers. All of it lives in the weights, compressed and distributed. When you create a specialized agent, you are not adding a persona. You are providing a mask that constrains which patterns surface.

The "QA engineer" prompt does not make the model better at testing. It makes the model worse at everything except testing. And that constraint, paradoxically, produces better testing. You have given it the gift of ignorance.

This is why role prompts work even though they seem like they should not. You are not fooling the model into having expertise it lacks. You are giving it permission to ignore capabilities it has. The mask does not hide the actor; it hides everything in the actor that is not the character.

The Economics of Attention

If specialization is subtraction, then agent design is an exercise in strategic elimination.

Most teams over-specify their agent prompts. They add more instructions, more capabilities, more edge case handling. Each addition feels like progress. But attention is zero-sum; this is not a metaphor but a mathematical fact. Every token you add to the system prompt dilutes the focus on the tokens that matter.

Day-Lewis does not prepare for Lincoln by also studying other presidents for comparison. He goes narrower, not broader. He eliminates even potentially relevant context if it might contaminate the singular focus.

Your architect agent does not need instructions about coding standards. Those tokens compete for attention with architectural reasoning. Your QA agent does not need to know about deployment pipelines. That knowledge, present in the weights, will activate if relevant and stay dim if not.

The best agent prompts are remarkably short. They specify what to attend to by implying everything else should recede. This is not laziness. It is the discipline of the sculptor who knows when to stop chipping.

For Those Who Build

If you are designing agent systems, consider these principles:

The same model can inhabit different roles because roles are constraints, not capabilities. Do not shop for specialized models when specialization lives in the prompting.

Shorter, sharper role definitions outperform comprehensive ones. Every instruction you add is attention you redirect. Make each token earn its spotlight.

When an agent underperforms, the instinct is to add more guidance. Often the fix is subtraction. Remove the noise so the signal can emerge.

Test your agents by checking what they ignore, not just what they produce. A well-focused agent should be notably worse at tasks outside its role. That is the feature, not the bug.

The Old Truth, Newly Computed

Daniel Day-Lewis and a transformer attention mechanism have more in common than their surface differences suggest. Both achieve specialization through the same fundamental operation: the strategic elimination of everything outside the target domain.

The sculptor knew this. The monk knew this. The method actor knows this still. Focus is not about trying harder. It is about attending to less.

When you build your next specialized agent, remember you are not adding capabilities. You are positioning a spotlight on a stage that already contains everything. The art is in choosing what to illuminate.

And more importantly: what to leave in darkness.

Michelangelo would understand perfectly.

Love the principle that expertise (in the context of AI Agents) is more about permissions then knowledge. Now i have to work on the length and the insane amount of details I put in my prompts 🙃

To view or add a comment, sign in

More articles by Herman Geldenhuys

  • Why Faster AI Doesn't Make Faster Companies

    AI got an order of magnitude faster. Your company got two or three times faster.

    4 Comments
  • The Great Inversion

    For 25 years we built interfaces for humans. Now we are building them for agents.

  • Building for the Wrong Model

    Most enterprise AI investments are compensating for limitations that are about to disappear. When they do, the…

  • Trust, But Verify

    AI is confidently wrong some percentage of the time. The skill is not checking every line.

    1 Comment
  • Say What You Mean

    Prompting did not die. It fragmented into four different skills.

  • Leeloo Dallas, Memorypass

    Leeloo learned 5,000 years of human history in ten seconds. Your AI can't remember last Tuesday.

  • The New Management

    AI agents are fast, moderately reliable, and have no sense of direction. The skill they demand most is not technical.

    4 Comments
  • Think in Pieces

    The skill is not solving the problem. It is slicing it.

    9 Comments
  • The Taste Gap

    The most productive users of AI are the ones who are best at rejecting its output. The most underrated skill in the age…

    6 Comments
  • The Spec Is the Product

    The cost of building just dropped by an order of magnitude. Everyone is talking about that.

Others also viewed

Explore content categories