Agents

Since the release of ChatGPT by OpenAI in October 2022, and with the subsequent emergence of projects like AutoGPT and AgentGPT, LLM-related agents have gradually become a research hotspot and a practical application direction in AI in recent years. This article will introduce the basic concepts, core technologies, and recent application progress of agents.

LLM Agent

A Large Language Model Agent (LLM agent) utilizes an LLM as the system’s brain, combined with modules for planning, memory, and external tools, to achieve the automated execution of complex tasks.

  • User Request: The user inputs a task through a prompt to interact with the agent.
  • Agent: The system’s brain, composed of one or more LLMs, responsible for overall coordination and task execution.
  • Planning: Decomposes complex tasks into smaller sub-tasks and formulates an execution plan, continuously optimizing the results through reflection.
  • Memory: Includes short-term memory (capturing task information in real-time using in-context learning) and long-term memory (using external vector stores to save and retrieve key information, ensuring information continuity for long-term tasks).
  • Tools: Integrates external tools such as calculators, web search, and code interpreters to call external data, execute code, and obtain the latest information.
Fig. 1. The illustration of LLM Agent Framework. (Image source: DAIR.AI, 2024)

Fig. 1. The illustration of LLM Agent Framework. (Image source: DAIR.AI, 2024)

RL Agent

The goal of Reinforcement Learning (RL) is to train an agent to take a series of actions ($a_t$) in a given environment. During the interaction, the agent transitions from one state ($s_t$) to the next and receives a reward ($r_t$) from the environment after each action. This interaction generates a complete trajectory ($\tau$), typically represented as:

$$ \tau = \{(s_0, a_0, r_0), (s_1, a_1, r_1), \dots, (s_T, a_T, r_T)\}. $$

The agent’s objective is to learn a policy ($\pi$), which is a rule for choosing actions in each state, to maximize the expected cumulative reward, usually expressed as:

$$ \max_{\pi} \, \mathbb{E}\left[\sum_{t=0}^{T} \gamma^t r_t\right], $$

where $\gamma \in [0,1]$ is the discount factor, used to balance short-term and long-term rewards.

Fig. 2. The agent-environment interaction. (Image source: Sutton & Barto, 2018)

Fig. 2. The agent-environment interaction. (Image source: Sutton & Barto, 2018)

In the context of LLMs, the model can be viewed as an agent, and the “environment” can be understood as the user’s input and the desired response format:

  • State ($s_t$): Can be the current conversation context or the user’s question.
  • Action ($a_t$): The text output by the model (answer, generated content, etc.).
  • Reward ($r_t$): Feedback from the user or the system (e.g., user satisfaction, automatic scoring by a reward model).
  • Trajectory ($\tau$): The entire sequence of text interactions from the beginning to the end of a conversation, which can be used to evaluate the model’s overall performance.
  • Policy ($\pi$): The rule by which the LLM generates text in each state (conversation context), generally determined by the model’s parameters.

For LLMs, they are traditionally pre-trained on massive offline datasets. In the post-training reinforcement learning phase, the model is trained with feedback from humans or other models to produce high-quality text that better aligns with human preferences or task requirements.

Comparison

The table below shows the differences between the two:

Comparison DimensionLLM AgentRL Agent
Core PrincipleAutomates complex tasks through planning, memory, and tools.Continuously optimizes its policy to maximize long-term rewards through a trial-and-error feedback loop with the environment.
Optimization MethodDoes not directly update model parameters. Performance is improved mainly through context extension, external memory, and tools.Continuously and frequently updates the policy model’s parameters, relying on reward signals from the environment for optimization.
Interaction MethodInteracts with users or external systems using natural language, flexibly calling various tools to obtain external information.Interacts with a real or simulated environment, which provides rewards or penalties, forming a closed feedback loop.
ObjectiveDecomposes complex tasks and completes them with the help of external resources, focusing on the quality and accuracy of the task results.Maximizes long-term rewards, seeking an optimal balance between short-term and long-term returns.

As research deepens, the combination of LLM and RL agents presents more possibilities, such as:

  • Using reinforcement learning methods to train Reasoning LLMs (like o1/o3), making them more suitable as base models for LLM agents.
  • Simultaneously, recording the data and feedback from LLM agents executing tasks to provide rich training data for Reasoning LLMs, thereby enhancing model performance.

Planning: Task Decomposition

The core components of an LLM Agent include planning, memory, and tool use, which work together to enable the agent to autonomously execute complex tasks.

Fig. 3. Overview of a LLM-powered autonomous agent system. (Image source: Weng, 2017)

Fig. 3. Overview of a LLM-powered autonomous agent system. (Image source: Weng, 2017)

Planning is crucial for the successful execution of complex tasks. It can be approached in different ways depending on the complexity and the need for iterative refinement. In simple scenarios, the planning module can use an LLM to outline a detailed plan in advance, including all necessary sub-tasks. This step ensures that the agent systematically performs task decomposition and follows a clear logical flow from the outset.

Chain of Thought

Chain of Thought (CoT) (Wei et al. 2022) describes the reasoning process by generating a series of short sentences step-by-step, known as reasoning steps. Its purpose is to explicitly show the model’s reasoning path, helping it better handle complex reasoning tasks. The figure below shows the difference between few-shot prompting (left) and chain-of-thought prompting (right). The few-shot prompt yields an incorrect answer, whereas the CoT method guides the model to state its reasoning process step-by-step, more clearly reflecting the model’s logical process and thus improving the answer’s accuracy and interpretability.

Fig. 4. The comparison example of few-shot prompting and CoT prompting. (Image source: Wei et al. 2022)

Fig. 4. The comparison example of few-shot prompting and CoT prompting. (Image source: Wei et al. 2022)

Zero-Shot CoT (Kojima et al. 2022) is a follow-up to CoT that proposes an extremely simple zero-shot prompting method. They found that by simply adding the phrase Let's think step by step to the end of a question, an LLM can generate a chain of thought, leading to more accurate answers.

Fig. 5. The comparison example of few-shot prompting and CoT prompting. (Image source: Kojima et al. 2022)

Fig. 5. The comparison example of few-shot prompting and CoT prompting. (Image source: Kojima et al. 2022)

Multimodal Chain-of-Thought (MCoT) (Zhang et al. 2023) integrates text and visual information into the reasoning process, breaking the limitation of traditional CoT, which relies solely on the language modality. Its framework is divided into two stages:

  1. Rationale Generation: Generates an explanatory reasoning chain based on multimodal information (text + image).
  2. Answer Inference: Uses the generated reasoning chain as an aid to infer the final answer.
Fig. 6. Overview of Multimodal-CoT framework. (Image source: Zhang et al. 2023)

Fig. 6. Overview of Multimodal-CoT framework. (Image source: Zhang et al. 2023)

Self-Consistency Sampling

Self-consistency sampling (Wang et al. 2022a) is a method that samples multiple diverse answers for the same prompt with temperature > 0 and then selects the best one. The core idea is to improve the final answer’s accuracy and robustness by sampling multiple reasoning paths and then taking a majority vote. The criteria for selecting the best answer can vary across tasks, but majority voting is generally used as a universal solution. For tasks that are easy to verify, such as programming problems, the answers can be validated by running them through an interpreter and using unit tests. This is an optimization of CoT and, when used in combination, can significantly improve the model’s performance on complex reasoning tasks.

Fig. 7. Overview of the Self-Consistency Method for Chain-of-Thought Reasoning. (Image source: Wang et al. 2022a)

Fig. 7. Overview of the Self-Consistency Method for Chain-of-Thought Reasoning. (Image source: Wang et al. 2022a)

Here are some subsequent optimization works:

  • (Wang et al. 2022b) later used another ensemble learning method for optimization, increasing randomness by changing the order of examples or replacing human-written reasoning with model-generated ones, followed by a majority vote.
Fig. 8. An overview of different ways of composing rationale-augmented ensembles, depending on how the randomness of rationales is introduced. (Image source: Wang et al. 2022b)

Fig. 8. An overview of different ways of composing rationale-augmented ensembles, depending on how the randomness of rationales is introduced. (Image source: Wang et al. 2022b)

  • If the training samples only provide the correct answer without the reasoning, the STaR (Self-Taught Reasoner) (Zelikman et al. 2022) method can be used: (1) Have the LLM generate reasoning chains and keep only the reasoning for correct answers. (2) Fine-tune the model with the generated reasoning, iterating until convergence. Note that a high temperature can easily lead to results with correct answers but incorrect reasoning. If there are no ground-truth answers, the majority vote can be considered the “correct answer.”
Fig. 9. An overview of STaR and a STaR-generated rationale on CommonsenseQA. (Image source: Zelikman et al. 2022)

Fig. 9. An overview of STaR and a STaR-generated rationale on CommonsenseQA. (Image source: Zelikman et al. 2022)

  • (Fu et al. 2023) found that more complex examples (with more reasoning steps) can improve model performance. When separating reasoning steps, the newline character \n works better than step i, ., or ;. Additionally, using a complexity-based consistency strategy, which only performs a majority vote on the top $k$ most complex generated reasoning chains, can further optimize the model’s output. It was also shown that replacing Q: with Question: in the prompt provides an additional performance boost.
Fig. 10. Sensitivity analysis on step formatting. Complex prompts consistently lead to better performance with regard to different step formatting. (Image source: Fu et al. 2023)

Fig. 10. Sensitivity analysis on step formatting. Complex prompts consistently lead to better performance with regard to different step formatting. (Image source: Fu et al. 2023)

Tree of Thoughts

Tree of Thoughts (ToT) (Yao et al. 2023) expands on CoT by exploring multiple reasoning possibilities at each step. It first decomposes a problem into multiple thought steps and generates several different ideas at each step, forming a tree structure. The search process can use Breadth-First Search (BFS) or Depth-First Search (DFS), and each state is evaluated by a classifier (or an LLM can be used for scoring) or a majority vote. It involves three main steps:

  • Expand: Generate one or more candidate solutions.
  • Score: Measure the quality of the candidate solutions.
  • Prune: Keep the top $k$ best candidate solutions.

If no solution is found (or the quality of the candidates is not high enough), it backtracks to the expansion step.

Fig. 11. Schematic illustrating various approaches to problem solving with LLMs. (Image source: Yao et al. 2023)

Fig. 11. Schematic illustrating various approaches to problem solving with LLMs. (Image source: Yao et al. 2023)

Prompt Engineering

Prompt Engineering, also known as In-Context Prompting, is the skill of optimizing input prompts to guide an LLM to produce the desired output. Its core objective is to control the model’s behavior through effective communication without updating the model’s weights.

Zero-Shot Prompting

Zero-Shot Prompting directly provides the model with task instructions without any examples. This method relies entirely on the knowledge and instruction-following capabilities the model learned during its pre-training phase. For example, for sentiment analysis:

Fig. 12. Zero-Shot Prompting.

Fig. 12. Zero-Shot Prompting.

For models that have undergone instruction fine-tuning, such as GPT-5 or Claude 4, they can understand and execute these direct instructions very well.

Few-Shot Prompting

Few-Shot Prompting provides a set of high-quality examples in the prompt, each containing an input and the desired output. Through these examples, the model can better understand the user’s intent and the specific requirements of the task, thus achieving better performance than zero-shot prompting. However, the disadvantage of this method is that it consumes more of the context window. For example, providing a few examples for sentiment analysis:

Fig. 13. Few-Shot Prompting.

Fig. 13. Few-Shot Prompting.

Automatic Prompt Construction

Automatic Prompt Engineer (APE) (Zhou et al. 2022) is a method that searches within a pool of candidate instructions generated by the model. It filters the candidate set and ultimately selects the highest-scoring candidate instruction based on a chosen scoring function.

Fig. 14. Automatic Prompt Engineer (APE) workflow. (Image source: Zhou et al. 2022)

Fig. 14. Automatic Prompt Engineer (APE) workflow. (Image source: Zhou et al. 2022)

Automatic Chain-of-Thought (Auto-CoT) (Zhang et al. 2022) proposes an automated method for constructing chain-of-thought examples, aiming to solve the problem of manual prompt design being time-consuming and potentially suboptimal. The core idea is to sample questions through clustering techniques and then leverage the LLM’s own zero-shot reasoning capabilities to automatically generate reasoning chains, thereby constructing diverse, high-quality examples.

Fig. 15. Overview of the Auto-CoT method. (Image source: Zhang et al. 2022)

Fig. 15. Overview of the Auto-CoT method. (Image source: Zhang et al. 2022)

Auto-CoT consists of two main stages:

  1. Question Clustering: Embeds the questions in the dataset and runs a clustering algorithm like $k$-means. This step aims to group similar questions into the same cluster to ensure the diversity of the subsequently sampled questions.
  2. Example Selection and Rationale Generation: Selects one or more representative questions from each cluster (e.g., the question closest to the cluster centroid). Then, it uses a Zero-Shot CoT prompt to have the LLM generate reasoning chains for these selected questions. These automatically generated “question-rationale” pairs form the final few-shot prompt used to perform the task.

Knowledge-Augmented Prompting

When dealing with knowledge-intensive or commonsense reasoning tasks, relying solely on the LLM’s parametric knowledge is often insufficient and may lead to incorrect or outdated answers. To address this issue, researchers have proposed two types of approaches:

Generated Knowledge Prompting (Liu et al. 2022) is a method that has the model first generate relevant knowledge before making a prediction. The core idea is that when a task requires commonsense or external information, the model may make mistakes due to a lack of context; if the model is first guided to generate knowledge related to the input and then answers based on that knowledge, the accuracy of the reasoning can be improved.

Fig. 16. Overview of the Generated Knowledge Prompting. (Image source: Liu et al. 2022)

Fig. 16. Overview of the Generated Knowledge Prompting. (Image source: Liu et al. 2022)

  1. Knowledge Generation: Based on the input, the model first generates relevant factual knowledge.
  2. Knowledge Integration: The generated knowledge is combined with the original question to form a new prompt input.
  3. Answer Inference: The model answers based on the augmented input.

Retrieval Augmented Generation (RAG) (Lewis et al. 2021) is a method that combines information retrieval and text generation to address knowledge-intensive tasks. The core idea is that relying solely on an LLM’s parametric knowledge (which is static) can easily lead to factual errors, whereas introducing retrieval from external knowledge bases can improve the factual consistency and timeliness of the generated results.

Fig. 17. Overview of the Retrieval Augmented Generation. (Image source: Lewis et al. 2021)

Fig. 17. Overview of the Retrieval Augmented Generation. (Image source: Lewis et al. 2021)

  1. Retrieve: Retrieve relevant documents from an external knowledge source (e.g., Wikipedia or a private knowledge base).
  2. Augment: Concatenate the retrieved documents with the original input to serve as the prompt context.
  3. Generate: The generation model (the original paper used a pre-trained seq2seq model, but today LLMs are mainstream) outputs the answer based on the augmented prompt.

Active Prompt

Active Prompt (Diao et al. 2023) proposes an improvement to the traditional CoT method, which relies on fixed, manually annotated examples. The problem is that fixed examples are not necessarily the most suitable for all tasks, which can lead to poor generalization. Active Prompt introduces an active learning strategy to adaptively select and update the best task-relevant examples, thereby enhancing the model’s reasoning performance.

Fig. 18. Illustrations of active prompting framework. (Image source: Diao et al. 2023)

Fig. 18. Illustrations of active prompting framework. (Image source: Diao et al. 2023)

  1. Uncertainty Estimation: With or without a few manual CoT examples, the LLM generates k answers for training questions (in the paper, k=5), and an uncertainty metric is calculated based on the variance of these answers.
  2. Selection: Based on the uncertainty level, the most uncertain questions are selected.
  3. Human Annotation: The selected questions are manually annotated to supplement with new, high-quality CoT examples.
  4. Inference: The newly annotated examples are used for inference, thereby improving the model’s performance on the target task.

Planning: Self-Reflexion

Self-reflexion is a key factor that enables an agent to improve iteratively by refining past action decisions and correcting previous mistakes. It plays a crucial role in real-world tasks where trial and error are inevitable.

ReAct

The ReAct (Reason + Act) (Yao et al. 2023) framework achieves seamless integration of reasoning and acting in LLMs by combining task-specific discrete actions with the language space. This design not only allows the model to interact with the environment by calling external interfaces like the Wikipedia search API but also to generate detailed reasoning trajectories in natural language to solve complex problems.

The ReAct prompt template includes explicit thought steps, with the basic format as follows:

Thought: ...
Action: ...
Observation: ...
...(Repeated many times)
Fig. 19. Examples of reasoning trajectories for knowledge-intensive tasks (e.g. HotpotQA, FEVER) and decision-making tasks (e.g. AlfWorld Env, WebShop). (Image source: Yao et al. 2023)

Fig. 19. Examples of reasoning trajectories for knowledge-intensive tasks (e.g. HotpotQA, FEVER) and decision-making tasks (e.g. AlfWorld Env, WebShop). (Image source: Yao et al. 2023)

As seen in the figure below, ReAct’s performance in both knowledge-intensive and decision-making tasks is significantly better than the baseline method that relies only on Action, demonstrating its advantages in enhancing reasoning effectiveness and interaction performance.

Fig. 20. PaLM-540B prompting results on HotpotQA and Fever. (Image source: Yao et al. 2023)

Fig. 20. PaLM-540B prompting results on HotpotQA and Fever. (Image source: Yao et al. 2023)

Reflexion

Reflexion (Shinn et al. 2023) enables an LLM to iteratively optimize its decisions through self-feedback and dynamic memory.

This method essentially borrows ideas from reinforcement learning. In a traditional Actor-Critic model, the Actor selects an action $a_t$ based on the current state $s_t$, while the Critic provides an evaluation (e.g., a value function $V(s_t)$ or an action-value function $Q(s_t, a_t)$) and gives feedback to the Actor for policy optimization. Correspondingly, in Reflexion’s three main components:

  • Actor: Played by the LLM, it outputs text and corresponding actions based on the environment’s state (including context and historical information). This can be written as:

    $$ a_t = \pi_\theta(s_t), $$

    where $\pi_\theta$ represents the policy based on parameters $\theta$ (i.e., the LLM’s weights or prompt). The Actor interacts with the environment and produces a trajectory $\tau = {(s_1,a_1,r_1), \dots, (s_T,a_T,r_T)}$.

  • Evaluator: Similar to the Critic, the Evaluator receives the trajectory generated by the Actor and outputs a reward signal $r_t$. In the Reflexion framework, the Evaluator can analyze the trajectory using pre-designed heuristic rules or an additional LLM to generate rewards. For example:

    $$ r_t = R(\tau_t), $$

    where $R(\cdot)$ is the reward function based on the current trajectory $\tau_t$.

  • Self-Reflection: This module adds a self-regulating feedback mechanism on top of the Actor-Critic model. It integrates the current trajectory $\tau$, the reward signals ${r_t}$, and historical experiences from long-term memory, using its language generation capabilities to produce self-improvement suggestions for the next decision. This feedback is then written to external memory, providing richer context for the Actor’s subsequent decisions, thus achieving iterative optimization similar to updating policy parameters $\theta$ through dynamic adjustment of the prompt, without updating the LLM’s internal parameters.

Fig. 21. (a) Diagram of Reflexion. (b) Reflexion reinforcement algorithm. (Image source: Shinn et al. 2023)

Fig. 21. (a) Diagram of Reflexion. (b) Reflexion reinforcement algorithm. (Image source: Shinn et al. 2023)

The core loop and algorithm of Reflexion are described as follows:

  • Initialization

    • Instantiate the Actor, Evaluator, and Self-Reflection models simultaneously (all can be implemented by an LLM), denoted as $M_a, M_e, M_{sr}$.
    • Initialize the policy $\pi_\theta$ (including the Actor’s model parameters or prompt, and initial memory).
    • The Actor first generates an initial trajectory $\tau_0$ according to the current policy $\pi_\theta$. After evaluation by $M_e$, $M_{sr}$ generates the first self-reflection text and stores it in long-term memory.
  • Generate Trajectory

    • In each iteration, $M_a$ reads the current long-term memory and environmental observations, sequentially outputs actions ${a_1, a_2, \ldots}$, interacts with the environment, and receives corresponding feedback, forming a new trajectory $\tau_t$. $\tau_t$ can be considered the short-term memory for this task, used only in the current iteration.
  • Evaluate

    • $M_e$ outputs rewards or scores ${r_1, r_2, \ldots}$ based on the trajectory $\tau_t$ (i.e., the sequence of the Actor’s behaviors and environmental feedback). This step corresponds to internal feedback from $M_e$ or results given directly by the external environment.
  • Self-Reflect

    • The $M_{sr}$ module synthesizes the trajectory $\tau_t$ and reward signals ${r_t}$ to generate self-correction or improvement suggestions $\mathrm{sr}_t$ in natural language.
    • The reflection text can be seen as an analysis of errors or as providing new inspirational ideas, and it is stored in long-term memory. In practice, we can vectorize the feedback information and store it in a vector database.
  • Update and Repeat

    • After appending the latest self-reflection text $\mathrm{sr}_t$ to the long-term memory, the Actor can retrieve relevant historical information using RAG in the next iteration to adjust its policy.
    • The above steps are repeated until $M_e$ determines that the task is accomplished or the maximum number of rounds is reached. In this loop, Reflexion relies on the continuous accumulation of self-reflection + long-term memory to improve its decisions, rather than directly modifying the model’s parameters.

Below are examples of Reflexion applied to decision-making, programming, and reasoning tasks:

Fig. 22. Reflexion works on decision-making, programming, and reasoning tasks. (Image source: Shinn et al. 2023)

Fig. 22. Reflexion works on decision-making, programming, and reasoning tasks. (Image source: Shinn et al. 2023)

In an experiment with 100 HotPotQA questions, a comparison between the CoT method and a method with episodic memory showed that adding a self-reflection step at the end with the Reflexion method significantly improved search, information retrieval, and reasoning capabilities.

Fig. 23. Comparative Analysis of Chain-of-Thought (CoT) and ReAct on the HotPotQA Benchmark. (Image source: Shinn et al. 2023)

Fig. 23. Comparative Analysis of Chain-of-Thought (CoT) and ReAct on the HotPotQA Benchmark. (Image source: Shinn et al. 2023)

DeepSeek R1

DeepSeek-R1 (DeepSeek-AI, 2025) represents a major breakthrough in the open-source community’s efforts to replicate OpenAI’s o1 (OpenAI, 2024), successfully training an advanced reasoning model with deep reflection capabilities through reinforcement learning techniques.

For a detailed training process and technical implementation of DeepSeek R1, please refer to my previous blog post: Progress in Replicating OpenAI o1: DeepSeek-R1.

A key transformation during the training of DeepSeek-R1-Zero is that as training progresses, the model gradually emerges with a remarkable self-evolution capability. This capability is manifested in three core aspects:

  • Self-reflection: The model can look back and critically evaluate previous reasoning steps.
  • Proactive exploration: When it finds the current problem-solving path to be suboptimal, it can autonomously search for and try alternative solutions.
  • Dynamic thought adjustment: It adaptively adjusts the number of generated tokens based on the complexity of the problem, achieving a deeper thought process.

This dynamic and spontaneous reasoning behavior significantly enhances the model’s ability to solve complex problems, enabling it to tackle challenging tasks more efficiently and accurately.

Fig. 24. The average response length of DeepSeek-R1-Zero on the training set during the RL process. (Image source: DeepSeek-AI, 2025)

Fig. 24. The average response length of DeepSeek-R1-Zero on the training set during the RL process. (Image source: DeepSeek-AI, 2025)

A typical “aha moment” also emerged during the training of DeepSeek-R1-Zero. At this critical stage, the model suddenly realized that its previous line of thought was flawed, quickly adjusted its thinking direction, and ultimately successfully arrived at the correct answer. This phenomenon strongly demonstrates that the model has developed powerful self-correction and reflection capabilities during its reasoning process, similar to the “aha” experience in human thought.

Fig. 25. An interesting “aha moment” of an intermediate version of DeepSeek-R1-Zero. (Image source: DeepSeek-AI, 2025)

Fig. 25. An interesting “aha moment” of an intermediate version of DeepSeek-R1-Zero. (Image source: DeepSeek-AI, 2025)

Memory

Human Memory

Memory refers to the process of acquiring, storing, retaining, and retrieving information. Human memory is primarily divided into the following three categories:

Fig. 26. Categorization of human memory. (Image source: Weng, 2017)

Fig. 26. Categorization of human memory. (Image source: Weng, 2017)

  • Sensory Memory: Used to briefly retain sensory information after the original stimulus (visual, auditory, tactile, etc.) has disappeared, typically lasting for milliseconds or seconds. Sensory memory is further divided into:

    • Iconic Memory: The momentary image or visual impression retained by the visual channel, generally lasting 0.25–0.5 seconds, used to form visual continuity in video or animation scenes.
    • Echoic Memory: The brief storage of auditory information, which can last for several seconds, allowing a person to replay recently heard sentences or sound clips.
    • Haptic Memory: Used to retain brief tactile or force information, generally lasting from milliseconds to seconds, such as the brief finger sensation when typing on a keyboard or reading Braille.
  • Short-Term Memory: Stores the information we are currently conscious of.

    • Lasts for about 20–30 seconds, with a capacity of typically 7±2 items.
    • Handles the temporary processing and maintenance of information during complex cognitive tasks like learning and reasoning.
  • Long-Term Memory: Can store information for days to decades, with a virtually unlimited capacity. Long-term memory is divided into:

    • Explicit Memory: Can be consciously recalled, including episodic memory (personal experiences, event details) and semantic memory (facts and concepts).
    • Implicit Memory: Unconscious memory, primarily related to skills and habits, such as riding a bike or touch typing.

These three types of human memory are intertwined and together form our cognition and understanding of the world. When building LLM Agents, we can draw inspiration from this classification of human memory:

  • Sensory Memory corresponds to the LLM’s embedding representations of raw input data (such as text, images, and videos).
  • Short-Term Memory corresponds to the LLM’s in-context learning, limited by the model’s context window max_tokens. When the conversation length exceeds the window, earlier information is truncated.
  • Long-Term Memory corresponds to an external vector store or database, where the Agent can retrieve historical information on demand using RAG technology.

LLM Agent Memory

When an Agent interacts with a user over multiple turns or executes multi-step tasks, it can utilize different forms of memory and environmental information to complete its workflow.

Fig. 27. An overview of the sources, forms, and operations of the memory in LLM-based agents. (Image source: Zhang et al. 2024)

Fig. 27. An overview of the sources, forms, and operations of the memory in LLM-based agents. (Image source: Zhang et al. 2024)

  • Textual Memory

    • Full Interaction: Records all conversation and action trajectories, helping the Agent to trace back the context.
    • Recent Interaction: Retains only the conversation content highly relevant to the current task, reducing unnecessary context usage.
    • Retrieved Interaction: The Agent can retrieve historical conversations or records related to the current task from an external knowledge base and integrate them into the current context.
    • External Knowledge: When the Agent encounters a knowledge gap, it can retrieve and acquire additional information through APIs or external storage.
  • Parametric Memory

    • Fine-tuning: Injects new information or knowledge into the LLM to expand the model’s internal knowledge.
    • Knowledge Editing: Modifies or updates existing knowledge at the model level, achieving dynamic adjustment of the model’s internal parametric memory.
  • Environment

    • Represents the entities and context involved when the Agent interacts with users and external systems, such as the user Alice, or accessible tools or interfaces (e.g., a ticket booking system, a streaming platform).
  • Agent

    • The LLM Agent is responsible for read and write operations, i.e., reading information from the external environment or knowledge base and writing new actions or content.
    • It also includes a series of management functions, such as merging, reflecting, and forgetting, to dynamically maintain short-term and long-term memory.

Another example is an Agent completing two different but related tasks, requiring the use of both short-term and long-term memory:

  • Task A: Play a video: The Agent records the current plan, actions, and environmental state (e.g., searching, clicking, playing the video) in its short-term memory. This information is stored in memory and the LLM’s context window.
  • Task B: Download a game: The Agent uses its long-term memory of knowledge related to Arcane and League of Legends to quickly find a way to download the game. The figure shows the Agent searching on Google; we can consider Google’s knowledge base as an external knowledge source. All new search, click, and download actions are also updated in the short-term memory.
Fig. 28: Illustration of short-term memory and long-term memory in an LLM-brained GUI agent. (Image source: Zhang et al. 2024)

Fig. 28: Illustration of short-term memory and long-term memory in an LLM-brained GUI agent. (Image source: Zhang et al. 2024)

Common memory elements and their corresponding storage methods can be summarized in the following table:

Memory ElementMemory TypeDescriptionStorage Medium / Method
ActionsShort-Term MemoryTrajectory of historical actions (e.g., clicking buttons, entering text)Memory, LLM Context Window
PlanShort-Term MemoryThe plan for the next operation generated in the previous or current stepMemory, LLM Context Window
Execution ResultShort-Term MemoryThe result returned after an action, error messages, and environmental feedbackMemory, LLM Context Window
Environment StateShort-Term MemoryAvailable buttons, page titles, system status, etc., in the current UI environmentMemory, LLM Context Window
Self-ExperienceLong-Term MemoryHistorical task trajectories and execution stepsDatabase, Disk
Self-GuidanceLong-Term MemoryGuiding rules and best practices summarized from historical successful trajectoriesDatabase, Disk
External KnowledgeLong-Term MemoryExternal knowledge bases, documents, or other data sources to assist in task completionExternal Database, Vector Retrieval
Task Success MetricsLong-Term MemoryRecords of task success rates, failure rates, etc., for improvement and analysisDatabase, Disk

Additionally, researchers have proposed new training and storage methods to enhance the memory capabilities of LLMs:

LongMem (Language Models Augmented with Long-Term Memory) (Wang, et al. 2023) enables LLMs to remember long historical information. It uses a decoupled network structure, freezing the original LLM parameters as a memory encoder, while using an Adaptive Residual Side-Network (SideNet) as a memory retriever for memory checking and reading.

Fig. 29. Overview of the memory caching and retrieval flow of LongMem. (Image source: Wang, et al. 2023)

Fig. 29. Overview of the memory caching and retrieval flow of LongMem. (Image source: Wang, et al. 2023)

It mainly consists of three parts: a Frozen LLM, a Residual SideNet, and a Cached Memory Bank. Its workflow is as follows:

  • A long text sequence is first split into fixed-length segments. Each segment is encoded layer by layer in the Frozen LLM, and the attention’s $K, V \in \mathbb{R}^{H \times M \times d}$ vector pairs are extracted at the $m$-th layer and cached in the Cached Memory Bank.
  • When faced with a new input sequence, the model retrieves the top $k$ most relevant key-value pairs from the long-term memory bank based on the current input’s query-key. These are then integrated into the subsequent language generation process. Meanwhile, the memory bank removes the oldest content to ensure the availability of the latest contextual information.
  • During inference, the Residual SideNet fuses the hidden layer outputs of the frozen LLM with the retrieved historical key-values, effectively modeling and utilizing the context of ultra-long texts.

Through this decoupled design, LongMem can flexibly schedule massive amounts of historical information without expanding its native context window, balancing both speed and long-term memory capabilities.

Tool Use

Tool use is an important component of LLM Agents. By empowering LLMs with the ability to call external tools, their functionality is significantly expanded: they can not only generate natural language but also access real-time information, perform complex calculations, and interact with various systems (such as databases, APIs, etc.), effectively breaking through the limitations of their pre-trained knowledge and avoiding the inefficient process of reinventing the wheel.

Traditional LLMs primarily rely on pre-trained data for text generation, which makes them deficient in areas like mathematical operations, data retrieval, and real-time information updates. Through tool calling, models can:

  • Enhance computational capabilities: For example, by calling a specialized calculator tool like Wolfram, the model can perform more precise mathematical calculations, compensating for its own arithmetic shortcomings.

  • Access real-time information: Using search engines like Google, Bing, or database APIs, the model can access the latest information, ensuring the timeliness and accuracy of the generated content.

  • Improve information credibility: With the support of external tools, the model can cite real data sources, reducing the risk of fabricating information and increasing overall credibility.

  • Increase system transparency: Tracking API call records can help users understand the model’s decision-making process, providing a degree of interpretability.

Currently, various LLM applications based on tool calling have emerged, utilizing different strategies and architectures to cover everything from simple tasks to complex multi-step reasoning.

Toolformer

Toolformer (Schick, et al. 2023) is an LLM that can use external tools through simple APIs. It is trained by fine-tuning the GPT-J model, requiring only a few examples for each API. The tools Toolformer learns to call include a question-answering system, Wikipedia search, a calculator, a calendar, and a translation system:

Fig. 30. Examples of inputs and outputs for all APIs used. (Image source: Schick, et al. 2023)

Fig. 30. Examples of inputs and outputs for all APIs used. (Image source: Schick, et al. 2023)

HuggingGPT

HuggingGPT (Shen, et al. 2023) is a framework that uses ChatGPT as a task planner. It selects available models from HuggingFace by reading their descriptions to complete user tasks and summarizes the results based on their execution.

Fig. 31. Illustration of how HuggingGPT works. (Image source: Shen, et al. 2023)

Fig. 31. Illustration of how HuggingGPT works. (Image source: Shen, et al. 2023)

The system consists of the following four stages:

  • Task Planning: Parses the user’s request into multiple sub-tasks. Each task has four attributes: task type, ID, dependencies, and parameters. The paper uses few-shot prompting to guide the model in task decomposition and planning.
  • Model Selection: Assigns each sub-task to different expert models, using a multiple-choice format to determine the most suitable model. Due to the limited context length, models need to be initially filtered based on the task type.
  • Task Execution: The expert models execute their assigned specific tasks and record the results, which are then passed to the LLM for further processing.
  • Response Generation: Receives the execution results from each expert model and finally outputs a summary answer to the user.

LLM Agent Applications

Generative Agents

The Generative Agents (Park, et al. 2023) experiment simulates realistic human behavior in a sandbox environment with 25 virtual characters driven by large language models. Its core design integrates mechanisms for memory, retrieval, reflection, and planning/reaction, allowing agents to record and review their experiences and extract key information to guide future actions and interactions.

Fig. 32. The screenshot of generative agent sandbox. (Image source: Park, et al. 2023)

Fig. 32. The screenshot of generative agent sandbox. (Image source: Park, et al. 2023)

The entire system uses a long-term memory module to record all observed events, combines it with a retrieval model to extract information based on recency, importance, and relevance, and then generates high-level inferences through a reflection mechanism, ultimately translating these outcomes into concrete actions. This simulation demonstrates emergent behaviors such as information diffusion, relationship memory, and social event coordination, providing a realistic simulation of human behavior for interactive applications.

Fig. 25. The generative agent architecture. (Park, et al. 2023)

Fig. 25. The generative agent architecture. (Park, et al. 2023)

WebVoyager

WebVoyager (He et al. 2024) is an autonomous web interaction agent based on large multimodal models that can control the mouse and keyboard to browse the web. WebVoyager uses the classic ReAct loop. In each interaction step, it views a browser screenshot annotated with a method similar to SoM (Set-of-Marks) (Yang, et al. 2023), which provides interaction cues by placing numerical labels on web elements, and then decides on the next action. This combination of visual annotation and the ReAct loop allows users to interact with web pages using natural language. For a concrete example, you can refer to the WebVoyager code using the LangGraph framework.

Fig. 33. The overall workflow of WebVoyager. (Image source: He et al. 2024)

Fig. 33. The overall workflow of WebVoyager. (Image source: He et al. 2024)

OpenAI Operator

Operator (OpenAI, 2025) is an AI agent recently released by OpenAI, designed to autonomously execute web tasks. Operator can interact with web pages like a human user, completing specified tasks by typing, clicking, and scrolling. The core technology behind Operator is the Computer-Using Agent (CUA) (OpenAI, 2025). CUA combines the visual capabilities of GPT-4o with enhanced reasoning abilities gained through reinforcement learning, and it has been specially trained to interact with graphical user interfaces (GUIs), including the buttons, menus, and text boxes that users see on the screen.

Fig. 34. Overview of OpenAI CUA. (Image source: OpenAI, 2025)

Fig. 34. Overview of OpenAI CUA. (Image source: OpenAI, 2025)

CUA operates in an iterative loop consisting of three stages:

  • Perceive: CUA “observes” the web page content by capturing browser screenshots. This vision-based input allows it to understand the page’s layout and elements.

  • Reason: Using a chain-of-thought reasoning process, CUA evaluates the next action based on the current and previous screenshots and the actions already taken. This reasoning ability enables it to track task progress, review intermediate steps, and make adjustments as needed.

  • Act: CUA interacts with the browser by simulating mouse and keyboard operations (such as clicking, typing, and scrolling). This allows it to perform various web tasks without needing specific API integrations.

CUA differs from the pre-existing WebVoyager in that it is an agent specifically trained with reinforcement learning, rather than a fixed-flow workflow built by directly calling GPT-4o. Although CUA is still in its early stages and has certain limitations, it has achieved state-of-the-art results on the following benchmarks.

Fig. 35. OpenAI CUA Benchmark Results. (Image source: OpenAI, 2025)

Fig. 35. OpenAI CUA Benchmark Results. (Image source: OpenAI, 2025)

Deep Research

Deep Research is essentially a report generation system: given a user’s query, the system uses an LLM as its core agent to generate a structured and detailed report after multiple rounds of iterative information retrieval and analysis. Currently, the implementation logic of various Deep Research systems can be mainly divided into two approaches: Workflow Agent and RL Agent.

Workflow Agent vs. RL Agent

The Workflow Agent approach relies on developers to pre-design a workflow and manually craft prompts to organize the entire report generation process. Its main features include:

  • Task Decomposition and Flow Orchestration: The system breaks down the user’s query into several sub-tasks, such as generating an outline, information retrieval, and content summarization, and then executes them in a predetermined sequence.
  • Fixed Process: The calls and interactions between different stages are pre-set, similar to building a static flowchart or a directed acyclic graph (DAG), ensuring that each step has a clear responsibility.
  • Reliance on Manual Design: This method primarily depends on the experience of engineers, who improve the output quality through repeated prompt tuning. It is highly applicable but has limited flexibility.

The LangGraph framework can be used to build and orchestrate workflows in the form of a graph.

Fig. 36. A workflow of the LangGraph. (Image source: LangGraph, 2025)

Fig. 36. A workflow of the LangGraph. (Image source: LangGraph, 2025)

The following table compares 5 common workflow and agent patterns:

PatternCore MechanismAdvantagesLimitationsUse Cases
Prompt ChainingSequentially calls LLMs, passing results step-by-stepSuitable for phased reasoning, more accurate resultsFixed process, high latencyDocument generation (outline → content), translation polishing
ParallelizationSplits sub-tasks for parallel processing, or multi-model votingIncreases speed, more robust resultsSub-tasks must be independent, high resource consumptionParallel content moderation, multi-model code detection
RoutingFirst classifies, then assigns to different models/flowsHighly targeted, improves efficiencyEffectiveness depends on classification accuracyCustomer service query routing, dynamic model size selection
Evaluator-OptimizerGenerate → Evaluate → Optimize iterativelyImproves result quality, suitable for tasks with standardsHigh cost, multiple iterations increase latencyTranslation optimization, multi-round retrieval refinement
Orchestrator-WorkerCentral orchestration, dynamically decomposes and schedules sub-tasksFlexible, can handle complex tasksComplex architecture, high scheduling costMulti-file code modification, real-time research integration
AgentLLM makes autonomous decisions, calls tools based on environmental feedbackHighly flexible, adapts to dynamic environmentsHard to predict, cost and security need controlAutonomous research agents, interactive problem solving

Currently, there are several open-source projects on GitHub that have implemented workflow-based Deep Research Agents, such as GPT Researcher and open deep research.

Fig. 37. An overview of the open deep research. (Image source: LangChain, 2025)

Fig. 37. An overview of the open deep research. (Image source: LangChain, 2025)

The RL Agent is another implementation method that uses RL to train a reasoning model to optimize the agent’s multi-round search, analysis, and report writing process. Its main features include:

  • Autonomous Decision-Making: The system is trained through reinforcement learning, allowing the agent to autonomously judge, decide, and adjust its strategy when facing complex search and content integration tasks, thus generating reports more efficiently.
  • Continuous Optimization: Using a reward mechanism to score and provide feedback on the generation process, the agent can continuously iterate and optimize its own policy, improving the overall quality from task decomposition to the final report.
  • Reduced Manual Intervention: Compared to fixed processes that rely on manually crafted prompts, the reinforcement learning training approach reduces the dependence on manual design, making it more suitable for variable and complex real-world application scenarios.

The table below summarizes the main differences between these two approaches:

FeatureWorkflow AgentRL Agent
Process DesignPre-designed fixed workflow with clear task decomposition and flow orchestrationEnd-to-end learning, with the agent making autonomous decisions and dynamically adjusting the process
Autonomous Decision-MakingRelies on manually designed prompts; the decision process is fixed and immutableThrough reinforcement learning, the agent can autonomously judge, decide, and optimize its strategy
Manual InterventionRequires extensive manual design and tuning of prompts, with significant manual interventionReduces manual intervention, achieving automatic feedback and continuous optimization through a reward mechanism
Flexibility & AdaptabilityWeaker adaptability to complex or changing scenarios, with limited extensibilityBetter suited for variable and complex real-world scenarios, with high flexibility
Optimization MechanismOptimization primarily relies on the experience of engineers, lacking an end-to-end feedback mechanismUtilizes the reward feedback of reinforcement learning to achieve continuous, automated performance improvement
Implementation DifficultyRelatively straightforward to implement, but requires tedious process design and maintenanceRequires training data and computational resources, with a larger initial development investment but better long-term results
Requires TrainingNo additional training needed; relies solely on manually constructed flows and promptsRequires training the agent through reinforcement learning to achieve autonomous decision-making

OpenAI Deep Research

OpenAI Deep Research (OpenAI, 2025) is an intelligent agent officially released by OpenAI in February 2025. Designed for complex scenarios, it can automatically search, filter, analyze, and integrate multi-source information to ultimately generate high-quality comprehensive reports. The system is built on o3 as its core base model and incorporates reinforcement learning methods, significantly improving the accuracy and robustness of its multi-round iterative search and reasoning processes.

Compared to traditional ChatGPT plugin-based search or conventional RAG techniques, OpenAI Deep Research has the following outstanding advantages:

  1. Reinforcement Learning-Driven Iterative Reasoning Leveraging the o3 reasoning model and reinforcement learning training strategies, the agent can continuously optimize its reasoning path during multi-round search and summarization, effectively reducing the risk of distortion caused by error accumulation.

  2. Multi-Source Information Integration and Cross-Validation Breaking the limitations of a single search engine, it can simultaneously call upon various authoritative data sources such as specific databases and professional knowledge bases, forming more reliable research conclusions through cross-validation.

  3. High-Quality Report Generation The training phase introduces an LLM-as-a-judge scoring mechanism and strict evaluation criteria, enabling the system to self-evaluate when outputting reports, thereby generating more clearly structured and rigorously argued professional texts.

Training Process

The training process for OpenAI Deep Research utilized a browser interaction dataset specifically tailored for research scenarios. Through these datasets, the model mastered core browsing functions—including searching, clicking, scrolling, and parsing files—and also learned to use Python tools in a sandboxed environment for computation, data analysis, and visualization. Additionally, through reinforcement learning training on these browsing tasks, the model can efficiently perform information retrieval, integration, and reasoning across a vast number of websites, quickly locating key information or generating comprehensive research reports.

These training datasets include both objective tasks with ground-truth answers that can be automatically scored, as well as open-ended tasks equipped with detailed scoring rubrics. During training, the model’s responses are rigorously compared against the ground-truth answers or scoring criteria, and the model’s generated CoT thought process is used by an evaluation model to provide feedback.

Furthermore, the training process reused the safety datasets accumulated during the o1 model’s training phase and was supplemented with safety training data specifically for Deep Research scenarios, ensuring that the model strictly adheres to relevant compliance and safety requirements during automated search and browsing.

Performance

The model achieved state-of-the-art results on Humanity’s Last Exam (Phan, et al. 2025), a benchmark that evaluates AI’s ability to answer expert-level questions across various professional domains.

Fig. 38. Humanity’s Last Exam Benchmark Results. (Image source: OpenAI, 2025)

Fig. 38. Humanity’s Last Exam Benchmark Results. (Image source: OpenAI, 2025)

Future Directions

Agents show great promise, but to achieve reliable and widespread application, the following key challenges still need to be addressed:

  • Context Window Limitations: The limited context window of LLMs restricts the amount of information they can process, affecting long-term planning and memory capabilities and reducing task coherence. Current research is exploring external memory mechanisms and context compression techniques to enhance long-term memory and the ability to process complex information. Currently, OpenAI’s latest model, GPT-4.5 (OpenAI, 2025), has a maximum context window of 128k tokens.

  • Interface Standardization and Interoperability: The current natural language-based interaction with tools suffers from a lack of format uniformity. The Model Context Protocol (MCP) (Anthropic, 2024) aims to unify the interaction between LLMs and applications through an open standard, reducing development complexity and improving system stability and cross-platform compatibility.

  • Task Planning and Decomposition Capabilities: Agents struggle to formulate coherent plans for complex tasks, effectively decompose sub-tasks, and lack the ability to dynamically adjust in unexpected situations. More powerful planning algorithms, self-reflection mechanisms, and dynamic policy adjustment methods are needed to flexibly respond to uncertain environments.

  • Computational Resources and Economic Viability: Deploying large model agents is costly due to multiple API calls and intensive computation, limiting their use in some practical scenarios. Optimization directions include more efficient model architectures, quantization techniques, inference optimization, caching strategies, and intelligent scheduling mechanisms. With the development of specialized GPU hardware like the NVIDIA DGX B200 and distributed technologies, computational efficiency is expected to improve significantly.

  • Security and Privacy Protection: Agents face security risks such as prompt injection. Robust authentication, permission control, input validation, and sandboxed environments are needed. For multimodal inputs and external tools, data anonymization, the principle of least privilege, and audit logs must be strengthened to meet security and privacy compliance requirements.

  • Decision Transparency and Interpretability: The decision-making process of agents is difficult to explain, limiting their application in high-stakes domains. Enhancing interpretability requires the development of visualization tools, chain-of-thought tracking, and decision rationale generation mechanisms to improve decision transparency, build user trust, and meet regulatory requirements.

References

[1] DAIR.AI. “LLM Agents.” Prompt Engineering Guide, 2024.

[2] Sutton, Richard S., and Andrew G. Barto. “Reinforcement Learning: An Introduction.” MIT Press, 2018.

[3] Weng, Lilian. “LLM-powered Autonomous Agents.” Lil’Log, 2023.

[4] Wei, Jason, et al. “Chain-of-thought prompting elicits reasoning in large language models.” Advances in neural information processing systems 35 (2022): 24824-24837.

[5] Kojima, Takeshi, et al. “Large language models are zero-shot reasoners.” Advances in neural information processing systems 35 (2022): 22199-22213.

[6] Zhang, Zhuosheng, et al. “Multimodal chain-of-thought reasoning in language models.” arXiv preprint arXiv:2302.00923 (2023).

[7] Wang, Xuezhi, et al. “Self-consistency improves chain of thought reasoning in language models.” arXiv preprint arXiv:2203.11171 (2022).

[8] Wang, Xuezhi, et al. “Rationale-augmented ensembles in language models.” arXiv preprint arXiv:2207.00747 (2022).

[9] Zelikman, Eric, et al. “Star: Bootstrapping reasoning with reasoning.” Advances in Neural Information Processing Systems 35 (2022): 15476-15488.

[10] Fu, Yao, et al. “Complexity-based prompting for multi-step reasoning.” arXiv preprint arXiv:2210.00720 (2022).

[11] Yao, Shunyu, et al. “Tree of thoughts: Deliberate problem solving with large language models.” Advances in neural information processing systems 36 (2023): 11809-11822.

[12] Zhou, Yongchao, et al. “Large language models are human-level prompt engineers.” The eleventh international conference on learning representations. 2022.

[13] Zhang, Zhuosheng, et al. “Automatic chain of thought prompting in large language models.” arXiv preprint arXiv:2210.03493 (2022).

[14] Liu, Jiacheng, et al. “Generated knowledge prompting for commonsense reasoning.” arXiv preprint arXiv:2110.08387 (2021).

[15] Lewis, Patrick, et al. “Retrieval-augmented generation for knowledge-intensive nlp tasks.” Advances in neural information processing systems 33 (2020): 9459-9474.

[16] Diao, Shizhe, et al. “Active prompting with chain-of-thought for large language models.” arXiv preprint arXiv:2302.12246 (2023).

[17] Yao, Shunyu, et al. “React: Synergizing reasoning and acting in language models.” International Conference on Learning Representations (ICLR). 2023.

[18] Shinn, Noah, et al. “Reflexion: Language agents with verbal reinforcement learning.” Advances in Neural Information Processing Systems 36 (2023): 8634-8652.

[19] Guo, Daya, et al. “Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning.” arXiv preprint arXiv:2501.12948 (2025).

[20] OpenAI. “Introducing OpenAI o1” OpenAI, 2024.

[21] Zhang, Zeyu, et al. “A survey on the memory mechanism of large language model based agents.” arXiv preprint arXiv:2404.13501 (2024).

[22] Zhang, Chaoyun, et al. “Large language model-brained gui agents: A survey.” arXiv preprint arXiv:2411.18279 (2024).

[23] Wang, Weizhi, et al. “Augmenting language models with long-term memory.” Advances in Neural Information Processing Systems 36 (2023): 74530-74543.

[24] Schick, Timo, et al. “Toolformer: Language models can teach themselves to use tools.” Advances in Neural Information Processing Systems 36 (2023): 68539-68551.

[25] Shen, Yongliang, et al. “Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face.” Advances in Neural Information Processing Systems 36 (2023): 38154-38180.

[26] Park, Joon Sung, et al. “Generative agents: Interactive simulacra of human behavior.” Proceedings of the 36th annual acm symposium on user interface software and technology. 2023.

[27] He, Hongliang, et al. “WebVoyager: Building an end-to-end web agent with large multimodal models.” arXiv preprint arXiv:2401.13919 (2024).

[28] Yang, Jianwei, et al. “Set-of-mark prompting unleashes extraordinary visual grounding in gpt-4v.” arXiv preprint arXiv:2310.11441 (2023).

[29] OpenAI. “Introducing Operator.” OpenAI, 2025.

[30] OpenAI. “Computer-Using Agent.” OpenAI, 2025.

[31] OpenAI. “Introducing Deep Research.” OpenAI, 2025.

[32] Phan, Long, et al. “Humanity’s Last Exam.” arXiv preprint arXiv:2501.14249 (2025).

[33] OpenAI. “Introducing GPT-4.5.” OpenAI, 2025.

[34] Anthropic. “Introducing the Model Context Protocol.” Anthropic, 2024.

[35] LangGraph. “A workflow of the LangGraph.” LangGraph Tutorials, 2025.

[36] Assaf Elovic. “GPT Researcher” GitHub Repository, 2025.

[37] LangChain. “Open Deep Research” GitHub Repository, 2025.

Citation

Citation: When reprinting or citing the content of this article, please indicate the original author and source.

Cited as:

Yue Shui.(Mar 2025). Large Language Model Agents. https://syhya.github.io/posts/2025-03-27-llm-agent

Or

@article{syhya2025llm-agent,
  title   = "Large Language Model Agents",
  author  = "Yue Shui",
  journal = "syhya.github.io",
  year    = "2025",
  month   = "Mar",
  url     = "https://syhya.github.io/posts/2025-03-27-llm-agent"
}