The Next Leap of Edge AI: Towards an ‘Agent Operating System’

Author:Peng Zhao (Founder of Zhici Fang and Co-founder of Yunhe Capital)IoT Think Tank Original

This is my 376th column article.

We seem to be on the verge of a scene where, on a certain night, a drone is buzzing in the sky, its camera precisely locking onto the mechanical vibration anomalies of the main control pump room. Meanwhile, a quadruped robot on the ground receives an anomaly code, swiftly avoiding obstacles to reach the site. The two are not coordinated through cloud scheduling but self-organize locally through an “Edge Agent Operating System” to form task collaboration: the drone is responsible for visual recognition and path analysis, while the ground robot executes and provides feedback. The entire process requires no human intervention and does not connect to a remote cloud platform.

This is not science fiction, but a real evolution of edge AI from inference engines to collaborative agents.

In the past few years, the evolution path of edge AI has been clear—from the initial exploration of low-power AI inference with TinyML, to the practical application of edge inference frameworks, to the rise of platform-level AI deployment tools, and the recent popularity of vertical models, we have achieved the task of “getting models to run.”

However, the next step for edge AI is not to stack more models and parameters, but to answer a more fundamental question:Once AI models are running, can they collaborate?

This limitation is the “invisible ceiling” that edge AI faces in moving towards a higher form of intelligence.

True edge intelligence goes beyond making judgments; it must make decisions, form systems, and execute tasks.This is precisely the starting point for the evolution of edge AI from static inference to dynamic agents.

What we need is not a larger model, buta group of models that can collaborate. It can be said thatmodels allow devices to see the world, while agents enable devices to participate in the world.

In this article, we will explore how edge AI evolves from model deployment to an agent operating system based on the latest market data, technological advancements, and platform trends, and how this trend will reshape the interaction methods, system architecture, and commercial value of smart terminals.

From Model Deployment to System Autonomy, AI Agents are Landing on the Edge

In the past, the mainstream way for enterprises to deploy edge AI was still a combination paradigm of “model-driven + platform scheduling”: develop a model, deploy it to a terminal, and complete resource allocation and state visualization through an edge platform.

This method indeed solved the problem of whether models could run in the early stages, promoting the landing of a large number of AI capabilities on the edge, but it also exposed an increasingly prominent structural bottleneck: as deployment scales expand and scene complexity increases, this model fails to answer a more essential question:Can models collaborate? Does the system have autonomous capabilities?

This shift in focus has already become clear at the enterprise decision-making level.

The Next Leap of Edge AI: Towards an 'Agent Operating System'

According to a global CIO survey report released by ZEDEDA at the beginning of 2025, 97% of the surveyed CIOs stated that their enterprises have deployed or plan to deploy edge AI internally in the next two years; 54% of enterprises explicitly hope that edge AI will become part of system-level capabilities rather than isolated single-point functions; more notably, 48% of enterprises listed “reducing dependence on the cloud and enhancing local autonomous response capabilities” as a key goal for the next stage.

This set of data reflects that an industry-wide consensus is forming:The future of edge AI is no longer just about whether models can run, but whether systems can achieve self-organization, self-perception, and self-response capabilities.

The core carrier of this capability leap is“Edge AI Agents.”

Compared to traditional model deployment paradigms, edge agents are no longer passive inference engines but the smallest intelligent units with perception, decision-making, action, and collaboration capabilities. They can not only run models but also initiate actions, negotiate roles, and allocate resources locally based on environmental states, system rules, and task objectives, becoming proactive intelligent nodes in edge systems.

For example, in a smart manufacturing scenario, the value chain of edge agents can be intuitively understood: when a camera on the conveyor belt identifies a defect in the material, the visual inspection agent will immediately generate an event signal; this signal triggers the material handling agent to automatically schedule mobile robots for the transfer of problematic materials; subsequently, the quality inspection agent conducts a secondary review upon receiving the signal; finally, the MES system agent synchronously updates the production schedule and the next process plan.

The entire process from anomaly identification to task execution no longer relies on a centralized scheduling system but is completed through multiple edge agents collaborating autonomously on-site. This “perception—decision—collaboration—feedback” closed loop not only enhances response efficiency but also gives the system a high degree of resilience and adaptability.

The Next Leap of Edge AI: Towards an 'Agent Operating System'

If model deployment solves the question of “whether devices have thinking capabilities,” then agent deployment further answers the question of “whether devices have participation capabilities.” To truly achieve this participatory nature, edge agents must possess a complete capability system.

We can summarize it asPCE Model—Perception, Coordination, Economy three levels of capability stack.

First is the perception layer (Perception).

Agents must be able to understand their environment, read and interpret data from multimodal sensors such as images, sounds, temperature and humidity, vibrations, and make task judgments based on contextual information. ZEDEDA’s research shows that over 60% of enterprises have deployed multimodal AI models in edge devices, providing a rich foundation for environmental perception for agents.

Next is the coordination layer (Coordination).

An agent cannot complete all tasks alone; a truly intelligent system relies on efficient collaboration among multiple agents. This collaboration is not merely data exchange but an intelligent agent network based on state sharing, role negotiation, and task division. The coordination capability upgrades edge systems from device interconnection to intelligent mutual assistance.

Finally, the economic layer (Economy).

When edge agents begin to possess capabilities such as task acceptance, resource negotiation, and cost control, they naturally become participants in the machine economy. The foundation for this layer’s realization is device wallets, encrypted identities, and programmable contract mechanisms. According to my judgment in the article “AI Explosion at the Edge + Cryptocurrency Transformation, Device Wallets Open the Door to AI Agency Economy,” the future M2M transaction volume between AI devices is expected to exceed the total transaction volume between humans, and agents will become active nodes in the edge economic network. Economic capability not only enables agents to execute tasks but also gives them collaborative value.

The three layers of perception, coordination, and economy together constitute the “PCE capability stack” of edge agents. It not only defines what capability modules an agent should possess but also provides a reference framework for the future system design of edge AI platforms.

Why Do Edge Agents Need an AI Operating System?

The Next Leap of Edge AI: Towards an 'Agent Operating System'

Although edge AI has achieved a leap from model deployment to platform management in recent years, the current mainstream edge AI platforms still remain at the level of “model runtime environment.” However, as AI evolves from models to agents, this traditional platform paradigm becomes inadequate.

The reason is thatagents are not static inference services but dynamic services with state perception, task negotiation, and autonomous action capabilities. They require not only execution space but also a complete operating system.

We call it the“Edge AI Operating System.”

Compared to traditional AI platforms, the edge AI operating system needs to meet three core capability requirements from the underlying architecture.

First, it must have the capability toschedule heterogeneous computing resources. In edge devices, AI models may run on various computing units such as CPUs, GPUs, NPUs, or even ASICs. How to dynamically allocate and balance loads among these heterogeneous computing resources becomes a technical challenge at the operating system level.

Second, a true edge AI operating system should supportmulti-agent runtime management. This means the system must not only run models but also schedule agents: including state perception, task scheduling, permission control, and behavior coordination among agents.

This also leads to the third core capability of AI OS:the integrated control mechanism of tasks, resources, and states. In traditional platforms, tasks are usually statically configured, resource allocation is demand-driven, and state management relies on external monitoring. However, in agent systems, these three are dynamically coupled: whether an agent can execute a task depends on its current state, available resources, and feedback from the behavior of other agents in the system.

These trends point to a fact:The rise of edge agents is forcing a reconstruction of the operating system paradigm.

If traditional operating systems were born for programs, then the upcoming edge AI operating systems are born for agents; they must understand hardware, models, behavior, collaboration, and ecology.

Currently, CIOs are facing not the question of “whether to deploy AI,” but the challenge of “how to systematically plan AI.” The emergence of agents is gradually transforming AI from a “project expenditure” to a “systemic infrastructure expenditure.”

Research from ZEDEDA indicates that over 54% of enterprises have adopted a “cloud + edge” hybrid deployment model, and it is expected that more than 60% of new AI budgets in the next two years will be allocated to edge deployment, with nearly half explicitly targeting the construction of “autonomous AI capabilities.” This reflects a fundamental change in the structure of enterprise AI spending: from CAPEX-dominated “model procurement + deployment costs” to OPEX-dominated “intelligent services + agent subscriptions.”

Enterprises will no longer pay based on the “number of models” but will manage budgets based on the “lifecycle of agents.” Enterprises will no longer purchase a model outright but will subscribe to certain agent functionalities and be billed based on effectiveness. All of this means that the industrialization path of edge agent systems is about to accelerate.

From “Models Can Run” to “Agents Can Live”—Four Thresholds

Although the future of edge agents is becoming clearer and the technical path is gradually unfolding, the transition from “models can run” to “agents can live” is not a linear evolution but a systemic upgrade that crosses four thresholds.

First,the complexity of scheduling is one of the most realistic and challenging issues today.

Edge scenarios are inherently heterogeneous, with diverse device types, varying computing structures, intermittent network conditions, and different models, resources, and sensor interfaces that agents rely on, making unified scheduling strategies difficult to implement. More complex is that agents themselves have dynamic states, and their behaviors are dependent on the environment and subject to temporal fluctuations; the scheduling system must not only allocate resources but also understand the current intentions and feasibility of the agents.

Second,model diversity constitutes the second threshold.

In practical applications of edge AI, an increasing number of tasks require collaboration between general language models and vertical industry models. However, these two types of models differ significantly in their operating mechanisms, input structures, computing power requirements, and response times, making traditional model-centric scheduling inadequate for the collaborative operation of agents.

More challenging is the third threshold—data privacy and compliance.

The biggest feature of edge AI is localized intelligence, which also means that the data it relies on is highly privatized and sensitive, involving core assets such as enterprise operational metrics, user behavior trajectories, and production chain statuses. In traditional AI, data is uploaded to the cloud for unified training and inference, but in agent systems, data is often generated, processed, and decided locally; how the system can achieve collaboration and learning among agents without violating data privacy becomes a difficult problem.

Finally,the issue of agent governance is gradually surfacing.

Multiple agents collaborating in the same system inevitably lead to phenomena such as resource competition, task conflicts, strategy competition, and even information deception. The traditional task priority system becomes complex within the agent framework, especially when agents possess learning or self-updating capabilities, making their behavior paths unpredictable and increasing system risks.

Only by crossing these four thresholds can agents truly “come to life,” capable not only of running and collaborating but also of continuously evolving, self-repairing, and operating safely within complex systems.

Final Thoughts

The future of edge AI lies not in deploying more models but in activating more “understanding, acting, and collaborating” agents. So-called intelligence is no longer about stacking computing power and model inference in the cloud, but about machines possessing perception and purpose in the physical world, having the ability to react and judge in local environments. In this upcoming new phase, enterprises will no longer just deploy models but will schedule agents.

AI does not run at the edge; it begins to think from the edge.

For enterprises, this is no longer a question of “whether to adopt AI” but a strategic decision of “whether they can build their own agent ecosystem.”

The future of edge agents is not as tools but as partners. They will make decisions with humans, collaborate in execution, and coexist in the long term. We are not just training models; we are shaping new organizational boundaries, new system intelligence, and new industrial orders.

References:

1. Edge AI Matures: Widespread Adoption, Rising Budgets, and New Priorities Revealed in ZEDEDA’s CIO Survey, Source: ZEDEDA3. 16 Changes to AI in the Enterprise: 2025 Edition, Source: a16z.com4. Why is EDGE AI growing so fast, Source: imaginationtech.com

Leave a Comment