HomeWork
//
ContactContact
Try searching for

AI powered.
Human engineered.
Growth driven.

Amsterdam·—·Studio open

Explore

  • Work
  • Services
  • Insights
  • University
  • About
  • The Collective

Connect

  • Contact
  • LinkedIn

Learn

  • University
  • AI Snapshot
  • AI Calculator

Notes from the studio

Short, useful, once or twice a month. Strategy, AI, craft, things we are making.

© 2026 Studio Hyra. All rights reserved.

Not sure what we do? We can explain it differently.Privacy Policy
What a disciplined agent-assisted coding process actually looks like
Technology6 min read

What a disciplined agent-assisted coding process actually looks like

May 9, 2026

Most agency teams using AI coding agents have the same experience: things go well for twenty minutes, then the model rewrites a file it wasn't supposed to touch, invents a dependency, and you spend the next hour unpicking it. The problem isn't the agent. The problem is the absence of a process.

Agent-assisted development is not just autocomplete at scale. It's a different mode of working that requires a different kind of discipline. The studios that are getting real output from it aren't prompting harder. They're structuring the work differently, phase by phase, before a single line of code is generated.

Stacked gray and yellow cubes and spheres on a grid, casting soft shadows.

Start with a specification, not a prompt

The most common mistake is opening a chat window and describing the feature you want. That hands the agent an underspecified problem and gives it license to make architectural decisions you haven't thought through yet.

A working process starts with a written specification. Not a formal PRD, but a tight document: what the feature does, what it does not do, what the data shapes look like, and what the acceptance criteria are. Two pages is usually enough. The point is to make every major decision before the agent touches the codebase.

This feels slow. It is not. Every vague sentence in your spec is a debugging session waiting to happen. The spec is also the artifact you hand the agent as context at the start of every session, which means you stop repeating yourself and the agent stops drifting.

The agent will fill any gap you leave with a reasonable guess. Your job is to leave fewer gaps.

Max Pinas, Studio Hyra

Break the work into phases with explicit checkpoints

Once the spec exists, break the implementation into phases small enough that you can read and understand the output of each one before moving to the next. In practice that means phases of roughly one to three hours of agent work, producing a discrete, testable unit.

At the end of each phase, you review the diff. Not the whole file, the diff. You check that the agent stayed inside its scope. You check that it didn't introduce side effects in adjacent modules. If it did, you stop, revert, and tighten the next instruction before continuing.

This is the checkpoint. Skipping it is where context debt accumulates. Once the agent has built on top of a mistake across two or three phases, you're not fixing a bug anymore, you're doing archaeology.

A useful heuristic. if you wouldn't be comfortable merging the diff after a five-minute review, the phase was too big.

A single yellow and gray spiral geometric structure on a grid.

Control context, not just prompts

Agents degrade with distance. The further a model gets from its initial context, the more it fills gaps with patterns from training rather than patterns from your codebase. This is not a flaw you can prompt your way out of. It's a property of how these systems work.

The practical response is aggressive context management. Start a new session for each phase. Feed the agent the spec, the relevant files, and nothing else. Resist the urge to carry a long conversation forward because it feels efficient. Long conversations feel efficient. They aren't. You're feeding the model its own earlier assumptions alongside your new instructions.

Some teams keep a short context file alongside each module. a plain text note describing what the module does, what it doesn't do, and which files it is allowed to touch. The agent gets this file as part of its context every time. It sounds bureaucratic. It cuts scope creep significantly.

There's a related point about file boundaries. Agents will happily refactor across files if you give them scope to do so. You usually don't want that. Explicit instructions about which files are in scope for a given phase are not overly cautious. They're the job.

Testing is not optional and it is not the agent's idea

The agent will write tests if you ask it to. It will write tests that pass against the code it just wrote. That is not the same as tests that describe the behavior your spec requires.

The discipline here is to write the test cases yourself, in the spec phase, before any implementation happens. Describe the inputs, the expected outputs, and the edge cases. Then hand that to the agent as part of its instructions for each phase. Now the tests are grounded in the spec, not in the implementation.

This is a version of test-driven development, applied to agent workflows. It's not about methodology for its own sake. It's about keeping the agent's output falsifiable. If you can't tell whether the agent did what you wanted, you can't build on it.

Multiple gray and yellow geometric pyramids arranged diagonally on a grid.

Tests written after the fact confirm what the agent did. Tests written before tell you whether what the agent did was right.

Max Pinas, Studio Hyra

What this means for agency teams

For studios running three or four product workstreams at once, this kind of structure has a second benefit beyond output quality. It makes agent-assisted work reviewable by someone who wasn't in the session.

A spec, a phase log, and a diff history are auditable. A long chat thread is not. When a client asks why a component works the way it does, or when a developer joins a project midway, the structured process pays for itself.

The teams getting the most from agent-assisted development right now are not the ones using the most capable models or the cleverest prompts. They're the ones who treated the process design as seriously as the tooling. The agent is fast. Structure is what makes fast useful.

If your current workflow is a chat window and optimism, that's a reasonable starting point. It's just not where you want to stay.

Ready when you are

Momentum starts with a conversation.

No forms, no intake. Just a real conversation with the people who do the work.

Book a callBook a call

Keep reading.

All insightsAll insights
Technology6 min read

What Murati's testimony tells us about AI safety culture

Mira Murati testified under oath that Sam Altman misled her. That single fact tells us more about AI safety culture than a decade of blog posts.

May 9, 2026
Technology6 min read

When your vendor's former CTO won't trust the CEO under oath

Mira Murati testified she could not trust Sam Altman. For agencies building on OpenAI, that is not a drama to watch. It is a procurement question to answer.

May 9, 2026