Less than three months after the release of ChatGPT 3.5 in early 2023, the paper on large language model (LLM)-powered autonomous agents was published, igniting immediate interest in AI agent technology. GPTs or customized versions of ChatGPT were launched at the OpenAI Developer Conference in November 2023. At present, agents have become a mainstream product of AI models.
Compared to general-purpose AI model products, agents, whether role-playing or task-specific, are technically easier to control. They offer more accurate outputs and are easier for users to understand and adjust. We will conduct an agent review to guide our follow-up research and work direction.
Current Technologies and Principles of Agents
From its initial structure-based definition to its current multi-modal model, the concept of an agent has gone through the phases of tools, social agents, workflows, multi-agent cooperation, and multiple modalities. Currently, an agent is defined as a virtual role based on the AI model with the ability to learn, remember, perceive the environment, recall past experiences, plan target tasks and execute them to influence the environment.
Each part of the agent, as shown in Fig. 1, is described as follows:
Agent Value Analysis
First, the current agent’s value mainly relies on LLMs, essentially a conditional probability generation model. Utilizing different prompts, such as text generation, task disassembly, logical inference, and scenario understanding, LLMs generate different types of outputs. Based on the output capability of LLMs, agents build humanized roles to serve within the production field.
Second, from a mainstream industry perspective, the value of agents is embodied by a composite expert team composed of both human and multiple virtual members. This setup massively improves the scope of work, enabling one person to do the work traditionally requiring multiple persons. The paradigm has shifted from using tools to orchestrating multiple agents, which then use tools to complete tasks. Compared with conventional tools, LLM-based agents can provide greater generality and flexibility in judgment and decision-making.
Experimental Results of ZTE
At present, ZTE has developed four types of agents based on its understanding of of LLMs and the communications industry. These include assurance assistant, intelligent Q&A, fault assistant, and network observation assistant.
The assurance assistant, utilized in major activity support scenarios, has a high degree of automation. It replicates real workflows to a virtual space, where key support experts, assistants, and troubleshooting experts work together to automatically complete workflows. They communicate with people through summarization, reporting and risk assessment. This is a complex type of job agent, developed with the aim of achieving L5 fully autonomous network.
The other three agents are technically task agents:
Agent Development Trends and Technology Breakdown
At present, the mainstream types of agents in the academic community are consistent with the experimental results of ZTE. They are as follows:
From the perspective of development trends, self-evolving agents are also important, as they can self-learn.
Mainstream agent products are categorized according to their technical level, as shown in Table 1.
We conduct further paper scanning and research on the technologies mentioned in Table 1, and find the following:
Insights About Agent Trends
Based on the above analysis, we can draw the following conclusions:
Leveraging AI models, a simple task agent can provide inspiring information for other agents. If a certain number of task agents can be reached, one of the two necessary conditions for the emergence of group intelligence can be met. Second, with the abstract summarization capability of AI models, an agent within a team can combine multiple highly correlated information fragments from different agents, fulfilling another necessary condition of group intelligence. Once these two necessary conditions are met, the phenomenon of group intelligence may start to emerge.
To sum up, through academic tracing, product experimentation, and technical decomposition of different types of agents, we have derived an insight: in the upcoming year, we expect a rapid increase in the number of ordinary agents, and the phenomenon of group intelligence may emerge before that of powerful agents.
Building on this insight, we need to further consider the following aspects: