Back to Research

Narrow Creative Skills for Coding Agents

Give an agent one narrow creative job and review the output against a clear checklist.

Editorial illustration for Narrow Creative Skills for Coding Agents. Give an agent one narrow creative job and review the output against a clear checklist.
Rogier MullerApril 13, 20265 min read

Most coding agents are asked to be generalists. That is usually the problem.

A general agent can write code, explain code, and sometimes fix code. But when the task needs taste, domain judgment, or a repeatable style, broad instructions tend to produce broad output. The result is often acceptable and forgettable. The pattern here is narrower: give the agent one concrete creative skill, then use it inside a bounded workflow.

The example was ad creative. That does not mean every team should start making ads. It means the agent was more useful once it had a specific output shape, a specific audience, and a specific bar for review. That pattern transfers to engineering work.

Why narrow skills work better

Agents do better when the task has a clear local objective. “Make this better” is too open. “Rewrite this landing page hero for a technical buyer” is better. “Turn this feature spec into three test cases that cover edge conditions” is better still.

A narrow skill helps in three ways.

First, it reduces search space. The agent is not trying to solve product strategy, copywriting, and implementation at once.

Second, it makes review easier. Humans can judge whether the output matches the skill. You do not need to inspect every line for hidden intent.

Third, it creates reusable prompts and checks. Once a skill works, you can keep the same constraints and swap in new inputs.

The value is not in the skill itself. The value is in the repeatable boundary around it.

What this looks like in practice

A useful pattern is to treat the agent like a specialist with a narrow brief.

For example:

  • “Write three ad angles for this feature, each aimed at a different buyer objection.”
  • “Generate a migration plan that assumes one engineer, one week, and no downtime.”
  • “Draft test cases for this endpoint using only user-visible behavior.”
  • “Rewrite this error message so a support engineer can act on it in under a minute.”

The agent is still doing creative work, but the creativity is boxed in. That makes the output more comparable and easier to reject when it drifts.

In the source example, Claude Code was used with an ad creative skill. The important part is not the vendor or the exact tool. It is the workflow: a coding agent plus a domain-specific creative constraint. That combination can be effective when the task is not purely mechanical.

A simple implementation pattern

If you want to try this in an engineering team, start with one task class.

Pick something where quality is visible and review is fast. Good candidates are:

  • release notes
  • migration plans
  • test case generation
  • error copy
  • onboarding snippets
  • feature descriptions for internal docs

Then define the skill in plain language. Keep it short. Include the audience, the format, and the failure modes.

A workable prompt structure is:

  1. State the role narrowly.
  2. State the audience.
  3. State the output format.
  4. State what not to do.
  5. Ask for a small number of options.

Example: “You are writing three concise release-note variants for backend engineers. Keep each under 60 words. Avoid marketing language. Focus on behavior changes and operational impact.”

That is enough to get started. You can refine later.

Where this breaks down

This pattern is not a replacement for judgment. It works best when the domain is bounded and the review criteria are clear.

It breaks down when:

  • the task needs deep product context the agent does not have
  • the output depends on subtle brand or legal constraints
  • the team cannot agree on what “good” means
  • the prompt becomes so specific that it hard-codes one example

There is also a risk of overfitting. A skill that works for one product surface may fail on another. Ad creative is especially prone to this because style can hide weak reasoning. In engineering work, that means you should separate style from correctness. A good-looking answer is not a good answer if the constraints are wrong.

How to review it

Use a small review loop. Do not ask whether the agent is “smart.” Ask whether the output is usable.

A practical review pass can check:

  • Did it follow the requested format?
  • Did it stay within the intended audience?
  • Did it introduce unsupported claims?
  • Did it miss an obvious edge case?
  • Would a teammate actually use this draft?

If the answer is no, tighten the brief before you change the model or the tool. That is usually the cheaper fix.

Methodology note

This is a good place to Review the output against a small, explicit checklist before you scale the pattern. If the checklist is vague, the agent will drift.

The practical takeaway

The useful move is not “let the agent be creative.” It is “give the agent one creative skill with a narrow job to do.”

For agentic coding teams, that means you can often get better results by defining a local specialty than by asking for more general intelligence. The best version of the workflow is boring in a good way: clear brief, bounded output, fast review, repeat.

That is what holds up.

Related research

Ready to start?

Transform how your team builds software today.

Get in touch