Uncovering
better ways of
developing
...yes, again.
Jorge Castro & Joan Leon
...yes, again.
Jorge Castro & Joan Leon
The question is not whether to use the milling machine. It is how to use it with master's degree.
We are uncovering better ways of developing
software by doing it and helping others do it.
Through this work we have come to value:
Individuals and interactions over processes and tools
Working software over comprehensive documentation
Customer collaboration over contract negotiation
Responding to change over following a plan
That is, while there is value in the items on
the right, we value the items on the left more.
© 2001, the above authors
this declaration may be freely copied in any form,
but only in its entirety through this notice.
Twelve Principles of Agile Software
It does not say "we have uncovered". It says "we are uncovering". A process that never ends.
senior developers
years working on their own repos
real tasks from their own projects
February 2026 — Claude Opus 4.6 completes tasks that would take an expert human ~15 hours, 1 out of 2 times.
Without clear context, the AI guesses. And when it guesses, it hallucinates.
Every session starts from zero. It is like hiring someone new every day.
One single session to research, plan and implement. The context gets polluted.
Code goes to production without review. Nobody can review everything the AI generates.
AI amplifies what you give it. If you give it a disciplined process, it amplifies your productivity. If you give it chaos, it amplifies the chaos.
| Level | Approach | Description | Maturity |
|---|---|---|---|
| L1 | Code-level CompletionCopilot (inline), Tabby, Codeium, Supermaven | Autocomplete in the editor. The human writes, the AI suggests. | Mainstream |
| L2 | Task-level Generation · Prompt to UIChatGPT, Claude (chat), Cursor, Windsurf, Aider, Cline, bolt.new, Lovable, v0 | The human defines a task, the AI generates code/UI. The human approves each step. | Adopted |
| Now We are here | Task-level with Optional AutonomyClaude Code, Codex CLI, Cursor (Agent Mode), Copilot Edits, Windsurf (cascade) | L2 by default, configurable to chain autonomous actions (headless mode). | Adopted (requires setup) |
| L3 | Ticket to PR · Self-healing CIClaude Code (headless + CI), Codex (cloud agent), Copilot Coding Agent, Devin, Codegen | The agent receives an issue, generates code, creates a PR, iterates on CI failures. Escalates when out of scope. | Early Adoption |
| L4 | AI Software EngineerDevin (enterprise), Factory.ai, Genie | Full cycle: requirements → code → deploy → monitoring → rollback. | Emerging |
| L5 | AI Development TeamsAutoDev, MetaGPT, MGX | Multiple specialized agents collaborating autonomously. | Experimental |
At the start of each session, the AI knows nothing about your code. Zero.
And it does not know when it does not know enough.
It invents APIs, methods and dependencies that do not exist... and tells you with full confidence.
The same prompt can produce different results every time you run it.
We generate code faster than our cognitive load allows us to review.
Each phase needs a clean context window. Keep the context clean or you will have problems.
Split work into deployable vertical slices. With AI you can generate a lot in little time. If the scope is not cut, the chaos scales just as fast.
Deliver value end-to-end continuously. Cut the feature into layers, implement the thinnest slice.
Each step must be reversible. Just because the agent can generate a lot very fast does not mean we should push it all to production at once.
Mutation testing, Acceptance Testing, Architectural Testing. All of them expand the feedback loop and now there are no excuses.
More generated code = more risk if you do not have a solid feedback loop.
Technical guardrails
Pre-merge checklist
.claude/plans/pending/2026… tengo un par de dudasTools change.
Responsibility does not.
Your value is in understanding the problem, not in typing the solution.
AI does not improve your process. It amplifies it. If it is good, it goes faster. If it is bad, it fails faster.
An agent without configuration is a junior without onboarding. Give it context, rules and a way to verify its work.
Share what you learn — let us keep uncovering together.