LLM Powered Autonomous Agents
AI Summary
The canonical taxonomy post. Weng, then at OpenAI, wrote the piece that every agent framework deck still cites: agent = LLM + planning + memory + tool use. The memory section breaks it into sensory (embeddings), short-term (in-context), and long-term (external vector store with MIPS retrieval), mapping cognitive science onto transformer architecture. This is where the field got its shared vocabulary in mid-2023, before MemGPT or Graphiti existed. Load-bearing because every subsequent product pitch is a variation on these four boxes.
Original excerpt
Building agents with LLM (large language model) as its core controller is a cool concept. Several proof-of-concepts demos, such as AutoGPT, GPT-Engineer and BabyAGI, serve as inspiring examples. The potentiality of LLM extends beyond generating well-written copies, stories, essays and programs; it can be framed as a powerful general problem solver.
In a LLM-powered autonomous agent system, LLM functions as the agent’s brain, complemented by several key components:
Planning * Subgoal and decomposition: The agent breaks down large tasks into smaller, manageable subgoals, enabling efficient handling of complex tasks. * Reflection and refinement: The agent can do self-criticism and…
18 more articles in this vault.
Import the full Agent Memory Patterns vault to Burn 451 and build your own knowledge base.
Content attributed to the original author (Lilian Weng). Burn 451 curates publicly available writing as a reading index. For removal requests, contact @hawking520.