
LLMs, Agents, and Orchestration - How Traditional Software Development Is Changing

The work is not disappearing. It is shifting from typing effort toward goal definition, review, architecture, quality, security, and accountability.
On November 30, 2022, OpenAI released ChatGPT for free to the public. Three more years have passed, and the world has changed at high speed. OpenAI is the most recognized with ChatGPT, but not without competition. Google and Anthropic in the US, DeepSeek and several open-source models from China. Every few months, leading models leapfrog each other again.
Software development is a core focus for many AI models. Since December 2025, with Codex 5.2 and Claude Code, a stronger consensus has emerged: models are finally good enough to rethink the traditional coding workflow. Andrej Karpathy, former OpenAI co-founder, wrote on X that within weeks his split shifted from “manual + autocomplete” to “agent coding + manual”. At the same time, he warned teams to closely supervise models on critical code, because errors are less often obvious syntax issues and more often subtle conceptual mistakes that are easy to miss. Source: https://x.com/karpathy/status/2015883857489522876
The work is not disappearing. It is shifting and becoming more demanding: away from typing effort and toward goal definition, review, architecture, quality, security, and accountability.
How chatbots become agents
A standard chatbot is reactive: question in, answer out. The bridge to reality still has to be built manually afterward, for example changing code, running tests, finding documentation, checking screens, or updating tickets.
With tool use, AI becomes operational. The model can execute concrete actions inside a development environment under explicit rules. In tools like Cursor, this becomes tangible: instead of “Here is a suggestion,” you get “Files were updated, checks were run, and changes were explained.”
At its core, an agent is a system that:
- understands a goal, meaning outcome instead of just an answer,
- collects the context it needs, including code, docs, and error messages,
- executes relevant actions, such as changes, tests, and iterations,
- verifies and improves results.
What matters is this: models are improving fast, but they are not infallible. Without guardrails, they can produce output that sounds convincing while still being only almost right. In real projects, that is a serious risk.
Agent orchestration: when one agent becomes a system
Once agents become useful, the next step is obvious: multiple agents sharing work. That is where agent orchestration begins.
With multiple agents, several challenges must be solved simultaneously:
- Communication: Who passes which information to whom, and in what format?
- Decisions: Which agent takes over when, and when do you escalate?
- Shared memory: What is treated as fact, what as hypothesis, and how do you avoid repeated work?
- Error handling: What happens on unexpected states, wrong assumptions, or tool failures?
- Verification: How do you ensure the result is actually correct?
- Permissions: What is each agent allowed to do, based on least privilege?
This is complex because agents do not run like a typical deterministic program. They are meant to discover their own path to an objective. The larger the project, the higher the chance of quiet, systemic errors. This is where experiments and professional delivery diverge: orchestration requires engineering discipline, including specification, tests, reviews, observability, and security boundaries, not just better prompting.
From “AI writes code” to “we build reliable systems”
As models become more capable, the challenge shifts: less “Can the model build X?” and more “How do we build a system that reaches outcomes reliably, securely, and maintainably?”. In agent workflows and orchestration, complexity rises quickly because agents do not execute deterministically like classical programs. They choose paths, use tools, and make assumptions.
That is why domain expertise and deep understanding become the key leverage point. Teams that truly understand product, UX, frontend, architecture, QA, and security can shape AI-assisted workflows so they:
- receive the right goals and constraints,
- are validated through explicit checks, tests, and reviews,
- surface false assumptions instead of merely sounding persuasive,
- and stay stable in real project delivery.
Agent systems and orchestration are still in a global exploration phase. The larger the project scope, the higher the risk of quiet conceptual errors that appear late and can only be detected, contained, and prevented with experience.
This field is evolving rapidly. We already use AI as a process tool and combine design and engineering expertise to reach better decisions faster and shorten iteration cycles, so our clients get the best product for their purpose sooner.
The direction is clear
More automation, more agents, more speed. The bottleneck remains the same: correct decisions and dependable execution. If you simply let AI run, you get output. If you guide it with expertise, you get convincing outcomes.
![[object Object]](/_next/image?url=https%3A%2F%2Fcdn.sanity.io%2Fimages%2Fqnbxd1h1%2Fproduction%2F2c1c79dc87706b3288d709eb4948ff5edb2bc0c2-1500x2001.jpg&w=3840&q=75)
Oskar Pokorski
Software Engineer


SEO + GEO in 2026: Practical Strategy for Growth Teams
Convinced that we are the right ones for your project? Let's work together!
Begin your inquiry now
Contact us via email, and we will get back to you promptly.