Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124

Until recently, applications based on large language models(LLMs) were treated as a curiosity – a demonstration of AI capabilities or a new conversational interface. Today, they are increasingly becoming the basis of entire business systems. This shift has concrete consequences: it changes not only the way we design software, but also who designs it and what competences are needed to do it well.
IT architects are therefore faced with the question: is the traditional model of building applications still sufficient?
In the classic application architecture, the logic of the system was based on predictable code. Input data was processed according to clearly defined rules and the result was mostly deterministic. In an approach with LLM on board, this logic is partly transferred to the model – which does not operate according to code, but according to prompt.
Prompt is not just a command. It is a new form of logical interface, containing instructions, constraints, contextual data and an expected form of response. As a result, some of what used to be written as code is now 'designed’ as prompt. Instead of precise conditional instructions, designers must now think in terms of intent, examples and expected behaviour.
In practice, this means a new model for building applications in which the LLM acts as one of the main layers of logic. A typical AI-centric application architecture might include:
Added to this are components such as response caching, token cost analysis, query classification or semantic error detection. As a result, the application architecture begins to resemble a system based on a machine learning pipeline, rather than a classic monolith or microservice.
The biggest change, however, is that LLMs introduce an element of … into the application. unpredictability. The models are probabilistic – they can give different answers for the same input, depending on parameters like temperature, context or prompt structure.
This in turn means:
One of the most important architectural decisions becomes the answer to the question: where to place the model? There are three basic scenarios to choose from:
The choice depends on many factors: from regulatory requirements to the cost of tokens to security and confidentiality needs.
From an architectural point of view, two approaches become particularly relevant:
In doing so, it is worth asking the question: should every application be LLM-based? The answer is: no. There are many cases where the classical approach is:
If the system requires deterministic results, real-time operation or tight control of business logic – it is better to stay with a traditional architecture. LLMs work best for tasks that require natural language understanding, flexibility and adaptation – not necessarily in accounting or payment processing.
With the entry of LLMs into the architecture, the structure of IT teams is also changing. New roles are emerging:
This also means a different approach to iteration – versioning prompts, testing their effectiveness and managing them as product artefacts.
Implementing large language models into applications is more than a change in technology – it is a change in the way we think about system logic, interface and architecture. The LLM stops being a tool and becomes a new layer of the system – equal to a database or a business rules engine.
IT architects today must not only know the limitations and capabilities of LLM, but also understand how to build a resilient, secure and cost-effective system around them. The new era of software is not just AI-first – it is also rethink-architecture-first.