HomeWork
//
ContactContact
Try searching for

AI powered.
Human engineered.
Growth driven.

Amsterdam·—·Studio open

Explore

  • Work
  • Services
  • Insights
  • University
  • About
  • The Collective

Connect

  • Contact
  • LinkedIn

Learn

  • University
  • AI Snapshot
  • AI Calculator

Notes from the studio

Short, useful, once or twice a month. Strategy, AI, craft, things we are making.

© 2026 Studio Hyra. All rights reserved.

Not sure what we do? We can explain it differently.Privacy Policy
What a disciplined agent-assisted coding process actually looks like
Technology6 min read

What a disciplined agent-assisted coding process actually looks like

May 9, 2026

Most conversations about AI-assisted coding stay at the level of vibes. Someone says the model "just writes the code now" and leaves you to figure out what that actually means on a Tuesday afternoon when the deadline is real and the codebase is not clean.

That gap between the demo and the discipline is where most agencies get into trouble. They pick up a coding agent, start a project, and discover about three days in that they have produced a pile of plausible-looking code with no coherent structure underneath it. The model was helpful. The process was not.

What follows is how we think about this at Studio Hyra. a phase-by-phase account of what a disciplined agent-assisted coding process actually looks like, drawn from working on real client products. Not a tutorial. More like field notes.

Interconnected glowing nodes forming a network across a grid, viewed from above.

Start with a spec the model can argue with

The single biggest mistake in agent-assisted coding is treating the model as a search engine for syntax. You type a vague intent, it returns plausible code, you paste it in, repeat. This works until it catastrophically doesn't.

The phase that most teams skip is specification. Before you open a chat window, you need a written description of what you are building that is specific enough for someone to disagree with. Not a PRD for a product manager. A technical brief: the data shape, the constraint, the edge case you already know about, the thing you will not do.

We write these in plain markdown. A few hundred words is fine. The goal is to give the model something it can push back on. When we share the brief and ask the model to flag gaps, it almost always finds one or two. That conversation, before a single line of code is written, is often the most valuable part of the whole session.

The model is a very fast junior who reads everything and forgets nothing within the session. Your job is not to prompt it. Your job is to direct it.

Max Pinas, founder, Studio Hyra

Work in small, reviewable increments

Once the spec exists, the instinct is to hand everything to the model and let it run. Resist this. The output will be long, locally coherent, and globally confused. You will spend more time understanding what it built than it took to build.

The discipline is granularity. Break the work into units small enough that you can read and understand the output of each one in under ten minutes. A function, a component, a migration, a test suite for one module. Ask the model to complete that unit, then stop.

Review each increment as a peer, not a user. Ask yourself. does this solve the right problem, or the problem as I described it? These are often different things. The model does exactly what you asked. If your ask was slightly wrong, the code will be slightly wrong, and the next increment will compound that error.

We keep a running notes file during sessions. Every increment gets a one-line note: what changed and why. This sounds tedious and takes about ninety seconds per increment. When something breaks three hours later, the notes file is the only thing that makes the debugging session short.

Abstract geometric shapes, some yellow, suspended and interacting within a grid space.

The session boundary problem

Context windows are long enough now that most people stop thinking about session boundaries. This is a mistake.

Every coding session has a shape. At the start, the model has your spec, your code, your conversation history, and a clear picture of intent. As the session grows, that picture dilutes. Old assumptions stay in the context. Revised ones layer on top. The model is not confused, exactly, but it is working from an increasingly complicated and partially contradictory picture.

We treat the end of a session as a forcing function. Before we close, we ask the model to write a summary of what was built, what decisions were made, and what remains open. This goes into a file called HANDOFF.md at the project root. The next session opens by reading that file.

The handoff note is also useful for the human. There is a specific kind of cognitive fatigue that comes from long agent sessions, where you feel productive because a lot happened, but you have lost track of why the early decisions were made. The handoff note is a record of your reasoning, not just the model's output.

A handoff note is a record of your reasoning, not just the model's output. That distinction matters more than most people think.

Max Pinas, founder, Studio Hyra

When the model gets confident and wrong

The hardest thing to calibrate with any coding model is the confidence signal. The model writes with the same tone whether it is producing a textbook solution or a hallucinated API that does not exist. There is no shake in the voice.

This is not a flaw you can engineer around. It is a property of the system. The mitigation is process, not prompting.

Two practices help. First, when the model proposes a library, a pattern, or a third-party integration you have not used yourself, verify it before you build on it. Not the documentation, the actual behaviour. Write a small isolated test. Five minutes of verification can save a day of unravelling.

Second, ask the model to argue against its own proposal before you accept it. Something like: "What are the main reasons this approach might be wrong for this context?" The answers are sometimes defensive and thin. But sometimes the model surfaces a genuine risk it had downweighted. That asymmetry is worth the extra prompt.

Agencies that move fast with these tools tend to build in a dedicated review beat at the end of each phase, where someone who did not write the code reads it cold. Not to find bugs, necessarily. To find the places where the code is technically correct and conceptually off. The model is good at the first. It needs help with the second.

A large, prominent yellow arrow pointing forward on a gridded, minimalist surface.

What this means for how you staff projects

Agent-assisted coding changes the role more than it changes the headcount, at least in the near term.

The person directing the model needs to be technically literate enough to review output at the logic level, not just the visual level. They do not have to be a senior engineer. But they need to understand what the code is supposed to do, and they need to notice when it does something adjacent instead.

What tends to go wrong at agencies is that the model gets handed to someone whose strength is somewhere else: a designer who codes a little, a PM who is technically curious. These are not bad people. They are in the wrong role. The model will produce fluent, broken code and they will not know until a client is staring at it.

The practical implication. keep one technically grounded person close to every agent-assisted build, even if they are not writing code themselves. Their job is not to type. Their job is to read what the model produced and decide if the direction was right.

That is, in the end, what the whole process comes down to. The model does the work quickly. The craft is in directing it precisely, reviewing it honestly, and knowing when to stop and re-spec rather than push forward. That part has not changed at all.

Ready when you are

Momentum starts with a conversation.

No forms, no intake. Just a real conversation with the people who do the work.

Book a callBook a call

Keep reading.

All insightsAll insights
Technology6 min read

What Murati's testimony tells us about AI safety culture

Mira Murati testified under oath that Sam Altman misled her. That single fact tells us more about AI safety culture than a decade of blog posts.

May 9, 2026
Technology6 min read

When your vendor's former CTO won't trust the CEO under oath

Mira Murati testified she could not trust Sam Altman. For agencies building on OpenAI, that is not a drama to watch. It is a procurement question to answer.

May 9, 2026