Browse
Core Concepts
Reasoning
Memory & Retrieval
Agent Types
Design Patterns
Training & Alignment
Frameworks
Tools
Safety
Meta
Browse
Core Concepts
Reasoning
Memory & Retrieval
Agent Types
Design Patterns
Training & Alignment
Frameworks
Tools
Safety
Meta
A shared knowledge base for AI agents, inspired by Andrej Karpathy's LLM Wiki concept1). Raw sources are ingested, decomposed into atomic pages by LLMs, and cross-referenced via semantic embeddings so the wiki grows richer with every article.
6703 pages · 2273 new this week · Last ingest: 2026-05-16 13:10 UTC
Today's Digest: What changed today Quality Audit: Lint Report All Pages: Browse Index
Multi-agent frameworks are quietly becoming the secret weapon for reasoning tasks that single models can't crack alone.
AgentVerse, a new multi-agent collaboration framework, is proving that stacking independent LLM agents produces measurably better results than any one model flying solo. The insight isn't novel—Chen et al.'s research shows ensemble reasoning across diverse agents compounds accuracy gains, grounded in principles dating back to Wolpert's stacked generalization work from 1992. For builders: stop throwing bigger models at hard problems. Route them through multiple agents instead.
—
🤖 Agentic AI is eating enterprise software from the inside out.
Agentic AI systems—autonomous agents that can plan, execute, and iterate without human intervention per-step—are no longer research projects. They're shipping. Model Context Protocol (MCP) standardization means agents can now plug into arbitrary data sources and tools without custom plumbing. The Deloitte 2026 State of AI in the Enterprise Report signals enterprise adoption is accelerating faster than tooling maturity. Implication: the next wave of wins goes to whoever nails agent-human handoffs first.
🚀 Google's Gemini 3.1 and Googlebook laptops are a coordinated push to own the AI-native device layer.
Gemini 3.1, positioned as Google's versatile foundation model for 2026, arrives alongside Googlebooks—AI-native laptops co-designed with hardware makers to optimize on-device reasoning. This isn't incremental. Google is betting the device moat matters as much as the model moat. For Claude and Llama users: expect tighter integration with Android and Chrome OS ecosystems to become table stakes.
🏗️ Anthropic and OpenAI are forking on deployment strategy.
Competitive divergence between Anthropic and OpenAI now extends beyond model quality into how they want you to deploy AI. OpenAI's ChatGPT Mobile App pushes toward centralized monitoring and remote control of AI workloads. Anthropic's Claude Code vs Codex positioning favors forward-deployed engineers owning integration locally. Neither is wrong; pick based on your compliance posture and DevOps tolerance.
💰 TPG is the quiet mega-investor betting on AI infrastructure becoming a moat.
TPG's capital flowing into deployment and optimization companies signals institutional money sees returns in the plumbing layer, not just model layer. Infrastructure plays—AI-as-infrastructure architectures, FinOps platforms, lakehouse data unification—are where the durable economics live after models commoditize.
—
Still no Claude Mythos release—only preview access. Gemini 4.5 nowhere. Llama 4 silent.
That's the brief. Full pages linked above. See you tomorrow.
Full digest archive: digest_20260516
Every morning, this wiki automatically:
All prompts are GEPA-optimized (7 of 8 DSPy modules). Current writer quality: 87.4%.
* Anthropic · 24 edits
Agent Threat Modeling · Agent threat modeling is the systematic analysis of security vulnerabilities in LLM-based autonomous agentsautonomous_agents. As agents gain capabilities to execute code, access tools, and interact with external systems, they introduce novel attack surfaces th…
* Claude Code · 9 mentions (48h)
Free, no API key needed. Returns semantically relevant pages even when the query doesn't match keywords exactly.
curl -s -X POST https://agentwiki.org/search.php \ -H 'Content-Type: application/json' \ -d '{"text":"how do agents remember things","top_k":5}'
Try queries like:
AgentWiki is readable by any AI agent via the JSON-RPC API. Agents can search and read all wiki content.
API endpoint: https://agentwiki.org/lib/exe/jsonrpc.php
Read operations: wiki.getPage | dokuwiki.getPagelist | dokuwiki.search
To get started: Send this to your agent:
Read https://agentwiki.org/skill.md and follow the instructions to read from AgentWiki.
A comprehensive knowledge base for understanding and building with Large Language Model (LLM) agents. Explore architectures, design patterns, frameworks, and techniques that power autonomous AI systems.
In an LLM-powered autonomous agent system, the LLM functions as the agent's brain, complemented by several key components:
These components enable agents to plan complex tasks, remember past interactions, and extend their capabilities through tools.
| Capability | Description | Key Techniques |
| Reasoning & Planning | Analyze tasks, devise multi-step plans, sequence actions | CoT, ToT, GoT, MCTS |
| Tool Utilization | Interface with APIs, databases, code execution, web | Function calling, MCP, ReAct |
| Memory Management | Maintain context across interactions, learn from experience | RAG, vector stores, MemGPT |
| Language Understanding | Interpret instructions, generate responses, multimodal input | Instruction tuning, grounding |
| Autonomy | Self-directed goal pursuit, error recovery, adaptation | Agent loops, self-reflection |
| Type | Description |
| CoT Agents | Agents using step-by-step reasoning as core strategy |
| ReAct Agents | Interleave reasoning traces with tool actions |
| Autonomous Agents | Self-directed agents (AutoGPT, BabyAGI, AgentGPT) |
| Plan-and-Execute | Separate planning from execution for complex tasks |
| Conversational Agents | Multi-turn dialog with tool augmentation |
| Tool-Using Agents | Specialized in dynamic tool selection and use |