For astir a decade, unreality architecture has been shaped by a elemental rule: Keep everything stateless. AWS Lambda and different serverless platforms encouraged teams to carve workloads into tiny, short-lived functions pinch nary persistent authorities and minimal section resources. This exemplary scaled well, was cost-efficient and became nan backbone of modern microservices.
By 2025, AI agents were pushed straight into accumulation crossed sectors specified arsenic finance, healthcare, energy, e-commerce and package development. As they began tackling existent business workflows — investigating incidents, analyzing documents, coordinating multistep tasks, moving tests aliases navigating soul systems — 1 shape became unavoidable: The Lambda-style attack could not support them. The assumptions that made serverless charismatic were precisely nan assumptions agents violated.
Two lessons defined 2025 arsenic enterprises tried to operationalize agents safely and predictably.
Lesson 1: Virtual backstage unreality (VPC) and on-premises became nan default for delicate supplier workloads.
In high-stakes environments, nan displacement toward backstage unreality and on-premises execution became particularly clear. Early shape experimentation pinch nationalist ample connection exemplary (LLM) APIs gave measurement to difficult questions from risk, compliance and information teams: Where is nan information going, who tin spot it, what is logged and tin we enforce audit requirements? For supplier workflows that touched customer information, soul APIs aliases delicate documents, nationalist endpoints were nary longer acceptable.
Snowflake’s continued 2025 description of Cortex illustrated really nan manufacture responded. Historically, Snowflake operated a centralized, cloud-hosted exemplary successful which compute and retention were intimately tied to Snowflake-managed infrastructure. Cortex marked a departure from that pattern. Instead of moving models successful Snowflake’s environment, Cortex executes straight wrong a customer’s existing Snowflake VPC. That intends embeddings, exemplary inference, logs and agent-driven instrumentality calls each enactment wrong nan enterprise’s ain web perimeter. This displacement was not conscionable a convenience. Snowflake framed it arsenic basal for industries pinch strict audit rules and zero tolerance for outer data movement, but agentic consequence guidance was implicit.
A akin modulation was visible successful different sectors. Sana’s study of business and financial deployments successful precocious 2025 emphasized that enterprises progressively demanded VPC aliases on-premises installations of their supplier platforms. These organizations needed their agents to authenticate utilizing soul personality providers, respect soul support structures and run wholly wrong pre-existing unreality information boundaries. Anything other introduced unacceptable operational and regulatory risk.
This inclination reflects a broader realization: Once an supplier gains nan expertise to interact pinch delicate information aliases return actions wrong accumulation systems, it becomes a privileged package component. And privileged systems cannot unrecorded extracurricular nan endeavor perimeter.
Why it matters: Enterprise AI programs must now presume that supplier workloads rubbing soul data will require VPC-native aliases on-premises deployment. This displacement shapes vendor selection, costs planning, web architecture, personality creation and really organizations building nan pipeline for supplier improvement and monitoring.
Lesson 2: Agent execution moved distant from stateless functions toward persistent unreality workstations.
The 2nd defining displacement of 2025 was architectural. AI agents do not run successful milliseconds. They activity crossed sequences of steps, referring to past context, creating intermediate files, moving validations, calling aggregate devices and returning to tasks complete extended periods. This workflow is fundamentally incompatible pinch nan serverless presumption that thing persists erstwhile an invocation ends.
Where accepted serverless presumes functions will tally briefly, ne'er clasp section authorities and reload tooling each time, supplier workflows require thing person to a long-lived workspace. They request unchangeable entree to their tools. They must clasp discourse crossed steps. They cannot spend to rebuild their situation for each action. And they must beryllium observable arsenic a continuous portion of work, not a bid of isolated events.
Frontier exemplary labs demonstrated this displacement passim 2025. Public demos from OpenAI and Anthropic showed agents tin of penning and executing code, navigating browser interfaces, searching done documents and coordinating crossed tools. These workflows only made consciousness wrong persistent execution environments that preserved authorities and tooling crossed steps. The underlying infrastructure looked much for illustration lightweight unreality workstations than ephemeral functions.
Research from Google DeepMind reflected nan aforesaid pattern. Its tool-using agents for debugging, codification execution and browser-based tasks each relied connected unchangeable environments wherever dependencies, caches and trial runners remained disposable crossed repeated attempts. Without continuity, nan workflows would neglect aliases go prohibitively slow.
Many enterprises discovered this firsthand. Attempting to unit agents into stateless architectures led to situation rebuilds, slow warm-up times, inconsistent behaviour and debugging challenges. As teams embraced persistent environments, they gained predictable performance, easier troubleshooting and nan expertise to observe supplier trajectories alternatively than isolated logs.
Why it matters: Moving to persistent execution environments changes nan measurement organizations deliberation astir orchestration, assets management, costs and security. The halfway portion of compute is nary longer a micro-invocation. It is simply a session. Enterprises request tooling that tin schedule, pause, resume, audit and discontinue full supplier sessions while preserving isolation and governance.
Where Cloud Architecture Is Heading
These 2 shifts constituent toward a caller unreality shape for AI agents:
- Deployment must align pinch information residency, auditability and regulatory requirements.
- Execution must support stateful, multistep workflows.
- Orchestration must dainty an agent’s afloat convention — not individual punctual calls — arsenic nan portion of reliability and control.
This represents a departure from 10 years of unreality creation built astir dividing activity into small, stateless pieces. Agents forced enterprises to judge a elemental reality: Meaningful automation requires continuity, discourse and control.
Looking ahead, nan organizations that accommodate to this caller architecture will tally agents that are reliable, governed and fresh for production. Those who effort to agelong serverless-era patterns astir supplier workloads will find that their infrastructure — not their models — is nan existent bottleneck.
YOUTUBE.COM/THENEWSTACK
Tech moves fast, don't miss an episode. Subscribe to our YouTube channel to watercourse each our podcasts, interviews, demos, and more.
Group Created pinch Sketch.
English (US) ·
Indonesian (ID) ·