Part I. Building LLM-Based Agents
Conclusions
Part II. Enhancing LLM-Based Agents for Reasoning and Safety Alignment
Post-Training: Adapting LLMs for Agents
7.1 LLMs for Agents
7.2 Training and Adaptation of LLMs for Agents
7.3 Data Preparation and Generation
7.4 Reinforcement Learning (RL) and Post-Training for Agent LLMs
7.5 RL-Scaling Law and Emergence of Reasoning Capabilities of LLMs
7.6 General-Purpose Agent LLMs
Test-Time Compute: Leveraging LLMs in Agents
8.1 Introduction to LLM Test-Time Compute
8.2 Test-Time Compute with Search-Based Strategies
8.3 Test-Time Compute with Deliberation and Self-Refinement Strategies
8.4 Test-Time Scaling Law
Safety and Alignment of LLM-Based Agents
9.1 Multi-Dimensional Risks of LLM-Based Agents
9.2 RL-Based LLM Post-Training for Agent Alignment
9.3 Scaling LLM Inference-Time Compute for Agent Alignment
Agents is a form of artificial life—an autonomous agent capable of independently achieving predefined objectives—to assist in managing a wide range of tedious and complex tasks. The agents are designed to perform tasks that typically require human intelligence, incorporating the ability to learn, reason, and self-improve. In essence, the combination of perception, learning, and action constitutes an intelligent agent.