Opportunities and Risks of Large Model Agents in the Digital World

Opportunities and Risks of Large Model Agents in the Digital World

Introduction

Language Agents, which are intelligent agents based on large language model technology, have great potential for general artificial intelligence (AGI) and large-scale automation of existing human labor if deployed responsibly. They may usher in a new era of scalable artificial intelligence and human collaboration. However, like all new technologies, we must also pay attention to and effectively mitigate the risks that come with them to avoid undesirable outcomes.
Research Areas: Language Agents, General Artificial Intelligence, Automated Society
Opportunities and Risks of Large Model Agents in the Digital World
Shunyu Yao | Author
Liu Peiyuan | Translator

This article is translated from https://princeton-nlp.github.io/language-agent-impact/

While 2022 was the year that language models like ChatGPT captured public attention, 2023 witnessed the rise of language agents (also known as agents). Papers like “ReAct” and “Toolformers”, along with frameworks such as “LangChain” and “ChatGPT Plugins”, demonstrated that language models can connect with web pages, software, tools, and APIs to enhance their capabilities through computational tools and customized, up-to-date information sources. This ability to act and influence the world allows language models to be applied in a broader range of fields, beyond traditional language processing. For instance, they can navigate websites to gather information, control software like Excel, or engage in interactive programming with execution feedback.

The “language agents” referred to in this article are intelligent agents based on large language model technology.

Referring to these machines simply as “language models” (whose optimization objective is to predict the next token) significantly underestimates their capabilities, as they are evolving into autonomous agents capable of using language as the primary medium to solve general digital tasks—in short, “language agents” in the digital world.

Despite the excitement surrounding demonstrations and papers about language agents, what does this mean for the future of artificial intelligence and society? This blog post aims to provide our insights on this question and spark discussions around the inherent opportunities and risks in the development of language agents. For a technical overview of these agents, please refer to the excellent blog post by Lilian Weng. Additionally, for more papers, blog posts, benchmarks, and other resources about language agents, please visit our repository.

https://github.com/ysymyth/awesome-language-agents

Language Agents in the Digital World:

A New Prospect for General Artificial Intelligence

The long-standing goal of artificial intelligence has been to create autonomous agents that can interact intelligently with their environment to achieve specific goals. Reinforcement learning (RL) is a powerful framework for addressing these challenges, with notable successes such as AlphaGo and OpenAI Five. However, RL has always struggled with the issues of lack of inductive bias and environmental constraints. Injecting human visual-motor or physical priors has been challenging, meaning that RL models often require millions of interactions to train from scratch. Thus, learning in physical, real-world environments has been fraught with challenges, as robot interaction speeds are slow and collection costs are high. This also explains why major reinforcement learning success cases have occurred in games—where simulations are fast and cheap, but also present closed and limited domains that are difficult to transfer to complex real-world intelligent tasks.

Opportunities and Risks of Large Model Agents in the Digital World

Agents interacting with physical or gaming environments face challenges for scalable learning or practical applications (top image), while agents interacting with the digital world enjoy both benefits (bottom image).

While physical environments and game worlds each have their limitations, the digital world (primarily using language) offers unique scalable environments and learning advantages. For example, WebShop is a shopping website environment with millions of products, where agents need to read web pages, input queries, and click buttons to shop, just like humans. Such digital tasks challenge multiple aspects of intelligence, including visual understanding, reading comprehension, and decision-making, and can be easily scaled. This also provides opportunities to guide agents to fine-tune using pre-trained prior knowledge—large language model prompts can be directly applied to WebShop or any ChatGPT plugin tasks, which is difficult to achieve in traditional reinforcement learning domains. As more APIs are integrated into the environment, an extremely diverse and highly open ecosystem of digital tools and tasks will emerge, giving rise to more general and capable autonomous language agents. This will pave new paths toward general artificial intelligence.

The Huge Potential of an Automated Society

While physical environments and game worlds each have their limitations, the digital world (primarily using language) offers unique scalable environments and learning advantages. For example, WebShop is a shopping website environment with millions of products, where agents need to read web pages, input queries, and click buttons to shop, just like humans. Such digital tasks challenge multiple aspects of intelligence, including visual understanding, reading comprehension, and decision-making, and can be easily scaled. This also provides opportunities to guide agents to fine-tune using pre-trained prior knowledge—large language model prompts can be directly applied to WebShop or any ChatGPT plugin tasks, which is difficult to achieve in traditional reinforcement learning domains. As more APIs are integrated into the environment, an extremely diverse and highly open ecosystem of digital tools and tasks will emerge, giving rise to more general and capable autonomous language agents. This will pave new paths toward general artificial intelligence.

An autonomous machine capable of acting has huge potential to alleviate human labor burdens across various fields. From robotic vacuum cleaners to self-driving cars, these machines are typically deployed in physical environments, equipped with task-specific algorithms and narrow application scopes. On the other hand, language agents like ChatGPT plugins and Microsoft 365 Copilot offer general solutions for automating a wide range of digital tasks, which is especially crucial in an era where much of human life and work is conducted in digital environments.

In a study involving 95 people, we glimpse the upcoming revolution—Github Copilot reduced average coding time by over 50%. However, Github Copilot only provides preliminary suggestive actions—an increasingly autonomous agent capable of repeatedly writing code, executing it, and utilizing automatic environmental feedback (such as error messages) to debug code is emerging.

Designers, accountants, lawyers, and any profession dealing with digital tools and data may experience similar situations. Furthermore, considering the Internet of Things connecting the physical world with the digital world, language agents could interact with physical environments far beyond simple functions like Alexa’s “turn on the light.” For instance, with cloud robotics lab services, language agents could participate in complex decision cycles for automated drug discovery: reading data, analyzing insights, setting parameters for the next experiment, reporting potential results, and so on.

Opportunities and Risks of Large Model Agents in the Digital World

The automation opportunities of language agents and their capability ladder.

Faced with endless possibilities, how should we classify them? There seems to be no single answer, just as human work can be classified or organized from multiple dimensions (salary level, work environment, knowledge level, general vs. specialized, etc.). Here, we propose a three-step progressive ladder based on agent capabilities.

• Step One: Enhancing the robustness of tedious digital labor: Tasks such as interacting with web pages and software to fill out various forms, repetitive Excel operations, or customer support tasks, or fixing code errors, involve multiple rounds of information retrieval and trial and error. These digital activities (excluding coding) can onboard newcomers with just a few hours of training, yet are repetitive and tedious for humans, and may also lead to errors due to fatigue. Similarly, automating these jobs does not seem to have fundamental barriers. Providing a few examples to GPT-4 can achieve reasonable performance on many such simple tasks. However, achieving human-level reliability and safety remains a challenge (see below). Once this is achieved, a significant portion of these jobs is expected to be automated, potentially marking the initial rise of automation driven by language agents.

• Step Two: Enhancing collaboration and communication skills for tasks requiring interaction with digital tools and humans: These tasks include sales while querying and recording information, acting as a project manager to take meeting notes and delegate tasks, or working collaboratively as a personal assistant across various digital platforms while recording user preferences. These tasks require not only the robustness to execute various digital routines but also human-like communication skills (such as pragmatics, theory of mind, personality understanding, etc.) to ensure successful and lasting collaboration with human (or agent) partners. Cultivating such skills and earning human trust is also a gradual process, akin to improving agent robustness for increasingly complex digital work.
• Step Three: Exploring innovative or knowledge domains: This includes accessing online literature and other information to draft reports; investigating research fields and proposing research ideas by navigating knowledge networks; and discovering mathematical knowledge through interaction with logical environments (such as Coq). These creative tasks resemble the work of scientists, artists, and writers, requiring not only strong digital and communication skills (how to search, how to communicate ideas and incorporate feedback, etc.) but also intrinsic motivation to define tasks for themselves and pursue long-term, scarce rewards of exploration.
Coq proof assistant https://coq.inria.fr/
This ladder also corresponds to different levels of task ambiguity and reward scarcity: from explicit instructions and clear task completion signals to contextual, implicit human intentions and actual human feedback inference, to self-defined tasks with intrinsic reward signals. The ability to study the latter does not have to wait for the former, but industrial deployment may proceed in this order from easy to difficult.

Balancing Progress and Safety

Opportunities and Risks of Large Model Agents in the Digital World

Robustness, malicious use, job insecurity, and the presence of risks. While history has provided insights into the first three issues, the presence of risks is less understood and more unknown.

All advancements in automation will inevitably raise some concerns, ranging from people losing jobs to existential crises. We have identified four potential issues that need to be addressed as language agents rise:
• Robustness of real-world applications: Compared to applications of large language models like text generation or question answering, the risks posed by agents taking autonomous actions are higher, as their actions directly impact the world—such as deleting files or executing transactions—and may scale rapidly. Any small error could have significant consequences and may go unnoticed until causing substantial damage.
• Malicious use: Language agents capable of completing complex tasks also imply a greater potential for malicious use, such as attacking websites, designing complex phishing schemes, or even releasing nuclear weapons—any malevolent hacking behavior that could exploit computers. This will require a comprehensive reform of current defenses, which are primarily deterministic and rely on simple tests like CAPTCHAs. Hackers could also inject malicious code into websites or other applications, causing benign agents running on them to malfunction in unexpected ways, such as leaking sensitive information like social security numbers or credit card details.
• Replacing human jobs: Like past technological advancements, the emergence of language agents will inevitably lead to the replacement of certain job positions while also creating new employment opportunities, just as the advent of cars transformed coachmen into drivers. Certain types of human jobs may disappear, evolving into more abstract forms, where humans supervise a team of agents to complete the same tasks more efficiently.
• AGI and existential risks: In extreme cases, autonomous agents also represent an important step toward AGI systems capable of performing complex tasks at human intelligence levels across a wide range of domains. This could pose existential risks to humanity, especially when agents are given control to change the world.

How to Address These Risks

Opportunities and Risks of Large Model Agents in the Digital World

Addressing safety issues of language agents (and AI in general) requires collaboration and multi-layered efforts from developers, researchers, educators, policymakers, and even AI systems.

The aforementioned issues are actively being discussed, and there are no definitive conclusions yet, but we can jointly assess them from historical perspectives and critical thinking.
1. Enhancing robustness through safeguards and calibration: Enhancing the robustness of language agents is a key step that requires the implementation of effective safeguards and calibration mechanisms. Currently, basic safety measures such as sandboxing or heuristic restrictions on the action space of agents (for example, OpenAI limiting ChatGPT plugins to perform GET requests on the web, or disabling the os function in CodeX) are adopted to prevent unsafe behaviors or error propagation. However, as language agents become increasingly autonomous and operate in more complex action spaces, ensuring their safety becomes more challenging. For this issue, we can explore several possible paths:
• Human involvement to enhance trust: Implementing gradual and cautious deployment strategies that include human supervision and alignment-oriented processes. This involves having human reviewers or supervisors participate in monitoring and guiding agent behavior during deployment. By incorporating human judgment and expertise, potential risks and unintended consequences can be identified and mitigated in a timely manner. This practice aligns with the research direction of “human-in-the-loop” systems.
• Providing formal guarantees against worst-case scenarios: Exploring the development of formal guarantees to ensure that language agents behave within acceptable bounds in specific action spaces. Drawing inspiration from adversarial reinforcement learning research, where techniques have been developed to defend RL agents against adversarial attacks, similar methods can be adapted to provide safety and robustness guarantees for language agents. By setting boundaries and limitations on agent actions, the impact of worst-case scenarios can be mitigated.
• Prompt-based behavioral guidance, such as Constitutional AI models: Adopting prompt-based behavioral guidelines inspired by legal frameworks (such as constitutions). By training language agents to follow specific prompt instructions that align with ethical principles and guidelines, their behavior can be guided to be consistent with social norms. This approach involves defining clear and specific rules for language agents to ensure their responsible and ethical behavior.

Constitutional AI https://arxiv.org/abs/2212.08073

2. Preventing malicious use through regulation: Responsible ownership, control, and oversight of large language models and their applications are crucial. In addition to technical solutions for robustness and safeguards, legal, regulatory, and policy frameworks are needed to govern their deployment. For instance, OpenAI has proposed a licensing system for large models, an idea that may soon be implemented in countries like China. Furthermore, strict data permission protocols and regulations can be established to prevent misuse and unauthorized access to sensitive information. At the same time, potential criminal behaviors should be considered, and punitive measures should be established based on the experiences of cryptocurrency crimes and their legal consequences.

OpenAI licensing system https://www.bloomberg.com/news/articles/2023-07-20/internal-policy-memo-shows-how-openai-is-willing-to-be-regulated

3. Employment impacts and educational policy needs: In the face of a (potential) employment crisis, implementing comprehensive education and policy initiatives is crucial. By equipping individuals with the skills and knowledge needed to adapt to changing environments, we can facilitate the smooth integration of language agents into various industries. This can be achieved through educational programs, vocational training, and reskilling initiatives to prepare the workforce for the demands of a technology-driven future.

4. Managing existential risks through understanding and research: Deepening our understanding of language agents and their impacts is crucial before taking further action. This involves a thorough understanding of how these models operate, their limitations, and potential risks. Additionally, establishing scalable oversight mechanisms to ensure responsible deployment and prevent potential misuse is also vital. One approach is to leverage language agents themselves to monitor and evaluate the behavior of other language agents, proactively identifying and mitigating any harmful consequences. Promoting further research in the field of language agents will help us gain a more comprehensive understanding of their safety implications and assist society in developing effective safeguards.

Final Thoughts

If deployed responsibly, language agents hold great potential for general artificial intelligence and large-scale automation of existing human labor, potentially ushering in a new era of scalable artificial intelligence and human collaboration. However, like all new technologies, there are still risks that must be immediately addressed and effectively mitigated to avoid undesirable outcomes. We believe this blog post is just the beginning, and we look forward to community discussions and collaborative efforts to advance the development of language agents safely.

“Post-ChatGPT” Reading Club

On November 30, 2022, a phenomenal application was born on the internet, which is ChatGPT developed by OpenAI. From Q&A to programming, from summarizing to essay writing, ChatGPT has demonstrated diverse general intelligence. Consequently, tech giants like Microsoft, Google, Baidu, Alibaba, iFlytek are gearing up to enter the arena… But please take a moment to calm down… Is it really appropriate to go all-in on large language models now? It is important to note that the foundation behind ChatGPT is deep learning + big data + large models, and these elements have been heating up since the era of AlphaGo five years ago. Those who missed the opportunity five years ago, why should they be able to catch the train of large language models now?
The Collective Intelligence Club is organizing a “Post-ChatGPT” reading club, initiated by Professor Zhang Jiang from Beijing Normal University and co-founded with several teachers including Xiao Da, Li Yanran, Cui Peng, Hou Yueyuan, Zhong Hanting, and Lu Yi, aiming to systematically sort out ChatGPT technology and discover its weaknesses and shortcomings.

Opportunities and Risks of Large Model Agents in the Digital World

For more details, see:
“Post-ChatGPT” Reading Club Launched: From General Artificial Intelligence to Conscious Machines

AGI Reading Club Launched

To delve into AGI-related topics, the Collective Intelligence Club, in collaboration with Yuetao Yue, Director of the Institute of Deep Perception Technology, MIT Ph.D. Shen Macheng, and Temple University Ph.D. student Xu Bowen, has jointly initiated the AGI reading club, covering topics including: definitions and measurements of intelligence, principles of intelligence, large language models and intelligent information worlds, perception and embodied intelligence, artificial intelligence from multiple perspectives, alignment techniques and AGI safety, and future societies in the AGI era. The reading club will begin on September 21, 2023, every Thursday evening from 19:00 to 21:00, expected to last for 7-10 weeks. Friends interested are welcome to sign up to participate!

Opportunities and Risks of Large Model Agents in the Digital World

For more details, see:

AGI Reading Club Launched: Interdisciplinary Pathways to General Artificial Intelligence

Recommended Reading

1. Wolfram: Will AI Take All Jobs and End Human History?
2. The Dawn of Artificial Intelligence: Viewing ChatGPT from the Perspective of Information Dynamics
3. Is General Artificial Intelligence (AGI) Already Here? A Deep Dive into the Mathematical and Physical Mechanisms of Intelligence Gained by ChatGPT
4. Zhang Jiang: The Foundation of Third-Generation Artificial Intelligence Technology—From Differentiable Programming to Causal Reasoning | New Course at Collective Intelligence Academy
5. Become a Collective Intelligence VIP to unlock all site courses/reading clubs
6. Join Collective Intelligence, let’s get complex!
Click “Read the original text” to sign up for the reading club

Leave a Comment