A working repo template.
The structure your agent reads from on every session. /spec, CLAUDE.md, AGENTS.md, the four files in /spec. Forkable.
agents — forTwo-weekend live cohort. Thirty engineers. By application. For people who've shipped a feature with Claude Code, Cursor, Devin or Aider, and noticed where it breaks.
You’ve shipped with Claude Code. You’ve also seen a confident hallucination get to staging. A refactor that wasn’t. A six-hour loop on a typo. Three weeks of subtle naming drift, found at merge.
The tools are real. The output, sometimes, is not. The next phase isn’t more AI. It’s method for AI.
Not the prompt. Not the chat. Not the PR. The spec.
Every failure above is a spec failure dressed up as something else. Hallucination is a spec that didn’t constrain the API surface. Refactor-that-wasn’t is a spec that didn’t define non-goals. Naming drift is a spec that didn’t include a glossary. Once you write the spec the agent should be reading, the failure modes become legible — and most of them stop happening.
The spec is a decision-grade document the agent reads before it reads your message. Problem, constraints, non-goals, success criteria. The prompt becomes a thin layer on top.
A four-step loop the agent runs against the spec, not against your last instruction. The loop is what makes the work survive a long session without drift.
Variables, decisions, manuals — versioned alongside the code. The artifacts that let the next session pick up where this one stopped without re-explaining the whole project.
By the end of weekend two, you walk away with four things. Not slides. Not a workbook. Four things you can take into Monday.
The structure your agent reads from on every session. /spec, CLAUDE.md, AGENTS.md, the four files in /spec. Forkable.
Real code, against a real codebase you bring with you, demoed live in front of the cohort on day four.
The four-step loop, the failure forensics, the spec format — written down in your words so you can take it back to your team without re-deriving it.
A Slack that stays open after the cohort ends. Engineers who’ve lived the same failure modes and committed to the same method.
Eighteen hours total. Ten hours live across four sessions. Eight hours async between weekends, paced by structured touchpoints. No prerecorded content; this is a workshop, not a course.
Diagnose. We catalogue the failure modes you’ve already lived through and name them precisely. No code yet. The diagnosis is the foundation.
Live spec writing against a real codebase you bring. Repo setup. The four files in /spec. By end of session you have a spec the agent can read from.
The four-step loop, run live in front of the room. How to intervene. How to scope. How to maintain the spec as the code grows. When each tool wins, when each fails.
Ship a feature against your spec. Demo it to the cohort. Take questions. The artifact you leave with is the demo.
There are three legitimate methodologies for working with AI coding agents — spec-first, eval-first, and vibe-then-verify. Each has a place. This program teaches one of them, and is honest about it.
| If you're considering | Our distinction |
|---|---|
| YouTube tutorials, Twitter threads | Live, structured, accountable. You demo in front of the room. |
| Generic prompt engineering courses | We don't teach prompts. We teach specs. The prompt is downstream. |
| Internal experimentation at your company | Twenty-nine peers across the Gulf and SL, structured failure forensics, named methodology. |
| Vibe-coding bootcamps | This is spec-first by design. If you want to feel it out, this isn't the cohort. |
| Hiring a consultant | This trains you to do the work. Cheaper, more durable, transferable to your team. |
The hard floor is two years of professional engineering and at least one feature shipped with an AI coding agent. Below that and the cohort doesn't work for you. Above that and the failure modes will be familiar.
I'd been using Claude Code for six months and getting better at it incrementally. The spec-first method was the first thing that changed how I worked, not just how fast I worked. The repo template alone saved my team a fortnight in the first month.[Engineer name] — [role], [company], [country]
More voices land as cohort 2 progresses. We don't publish stats below the threshold we'd want to see ourselves.
Applications open in the next few weeks. The form takes about ten minutes and includes one question that does most of the qualifying work: describe a build that went sideways with an agent, and what you learned. Anyone who can answer that question is in the audience.
Drop your email and we’ll send the form the day it opens — along with cohort 2 dates, fee, and the qualifying questions in advance so you can think about them.
// we email twice. once when applications open. once when cohort 2 dates are confirmed.
Two Saturdays and two Sundays, three and a half to four hours each, plus about three hours async between weekends. Eighteen hours total across two consecutive weekends.
SGT, 10:00 start. Confirmed at signup. Recordings are released within six hours, but live attendance is the default. If you can't make three of four sessions live, this isn't the right cohort.
Sent after application acceptance. We application-gate to make sure the cohort fits, before talking pricing. Pricing lands in a conversation with someone we've already said yes to. That's the right shape for this kind of programme.
No. We cover Claude Code, Cursor, Devin and Aider. You should have used at least one of them to ship a feature. Tool-specific tutorials live on YouTube; we teach when each tool wins and when each fails.
Three legitimate methodologies for working with AI coding agents. Spec-first front-loads the contract: the agent reads a decision-grade spec before the prompt. Eval-first front-loads the test: you write the eval, then let the agent satisfy it. Vibe-then-verify lets the agent run loose, then formalises post-hoc. We teach spec-first because it's the methodology that holds up best for B2B software, internal tools, and regulated work, and we say so on the page.
English.
Yes, available on request. It's not the point. The repo and the method are the point.
Yes. We provide invoices. We also run private cohorts for teams of ten and above; ask us when you apply.
Janaka Ediriweera and Tiran, co-founders of specshop.dev. Nine years of spec-first delivery across more than a hundred and twenty products in five countries.
The cohort Slack stays open. One-to-one advisory and a contractor pipeline are available for top performers. Mentioned, not detailed.
The methodology is the same; the curriculum isn't. Each track is taught for the role it's for.
The partners on the spec side. Writing briefs that compile, working with engineers driving agents, what design-system.md looks like in practice.
register interest →The verification and feedback side. Test specs, eval design, the customer-facing loop after the agent ships.
register interest →