Agentifact Reference Architecture
The Agent Economy Stack
14 layers, mapped from identity infrastructure to agent operating environments. Every layer has an editorial assessment, key players ranked, and the tools Agentifact has indexed. This is the map — not the marketing version.
How this works
Each layer represents a distinct infrastructure concern. Layer 1 is the foundation; Layer 14 is the human-facing surface. Layers are grouped into 4 bands by concern. Tools indexed in Agentifact are mapped to the layers they belong to. Click any layer to explore key players, editorial analysis, and indexed tools.
InterfaceThe human-facing surface of the agent economy
L14
Agent Operating Environments
OpenClaw-class systems — where a human and an AI agent work together continuously on complex, long-horizon tasks. These are not chatbots or simple code completers. They are environments that combine persistent context, tool access, autonomy controls, and session continuity into a working relationship between human and agent. This category is moving faster than any other in the stack, and the feature differences between systems are consequential for builders choosing their environment.
## Feature Comparison: 9 Systems × 8 Dimensions
| Feature | Claude Code | Cursor | Windsurf | Copilot Workspace | Devin | SWE-agent | Replit Agent | Aider | Continue.dev |
|---------|---|---|---|---|---|---|---|---|---|
| **Memory Model** | Session + vault persistence | Codebase context | Full-repo semantic | Issue→PR pipeline | Episode memory | Trajectory history | Project context | Git history | Session history |
| **Tool Access** | MCP + file + shell | VSCode exts + term | VSCode + Cascade | GitHub API + term | Terminal + browser | Git + grep + REPL | Replit env + pkg | Git + file + shell | IDE + custom APIs |
| **Autonomy** | Semi (approval gates) | Supervised | Semi (Cascade) | Supervised | Full-autonomous | Full-autonomous | Semi | Semi (interactive) | Supervised |
| **Approval Modes** | Selective HITL | All require approval | Cascade batch review | All require approval | Full auto + spots | Full auto | Human reviews | Interactive y/n | All require approval |
| **Context Window** | 200K rolling | GPT-4o rolling | 200K full-repo | GPT-4o focused | Proprietary extended | 200K trajectory | Claude context | 128K rolling | Varies by LLM |
| **Session Persistence** | Resumable | Ephemeral | Resumable | Ephemeral (PR is artifact) | Continuous | Resumable | Resumable | Resumable CLI | Ephemeral |
| **Multi-Agent** | Yes (OpenClaw orchestration) | No | No | No | Internal decomposition | No (deterministic) | No | No | No (extensible) |
| **Open Source** | Partial (MCP open) | Proprietary | Proprietary | Proprietary | Proprietary | Open (GitHub) | Proprietary | Open (GitHub) | Open (GitHub) |
## Decision Guide
**Choose Claude Code** if you need persistent multi-session memory, MCP tool access, and semi-autonomous approval workflows. Long-running build tasks, cross-project refactoring, autonomous pipelines + humans.
**Choose Cursor** if you want a lightweight pair programmer for single-file edits. Fast, integrated into your editor, minimal overhead. Best for rapid iteration and code review.
**Choose Windsurf** if you need full-repo context and batched multi-file edits. Cascade's ability to generate diffs for entire features is powerful for feature development and refactoring across files.
**Choose Copilot Workspace** if you're in the GitHub workflow and want the issue → PR pipeline automated. Low friction within GitHub. Best for ticketing systems and open-source contribution.
**Choose Devin** if you need true autonomy on long-horizon tasks (hours → days). Verify at checkpoints, not every action. Best for bug fixes, feature implementation, research tasks.
**Choose SWE-agent** if you're building research-grade agents or want deterministic, reproducible action traces. Open source, designed for GitHub issue resolution and reproducible benchmarks.
**Choose Replit Agent** if you're building full-stack projects and want an integrated environment. No local setup required. Best for quick prototypes, learning, full-stack development.
**Choose Aider** if you prefer CLI, want git integration, and like interactive approval workflows. Lightweight, extensible. Best for terminal-based workflows and git-heavy teams.
**Choose Continue.dev** if you want an open-source, model-agnostic copilot. Extensible with custom actions. Best for teams with custom models or on-premise deployments.
→
L13
Enterprise Orchestration Platforms
The five major technology companies are each building proprietary orchestration layers that will likely dominate how agents are deployed at scale. Unlike framework-level tools, these are platform bets — deep integrations with existing cloud infrastructure, identity systems, and enterprise software. Understanding how each company's approach differs is essential for builders choosing where to build.
## OpenAI
**Q4 2024:** OpenAI Agents SDK released with stateful task management, built-in error recovery, and streaming responses. Responses API added computer use (screenshot + click/type) and tool_use protocol. Swarm (experimental) released for lightweight multi-agent coordination.
**Q1 2025:** o1 model integrated into Agents SDK. Improved token efficiency and longer-horizon reasoning for complex orchestration tasks.
**Strategy:** OpenAI's approach prioritizes tight integration with GPT-4o/o1 and model-native tool use. Handoffs between agents are first-class. Cloud-hosted by default — orchestration is baked into the model API, not a separate service.
---
## Google
**Q3 2024:** Google ADK (Agent Development Kit) released with built-in multi-agent orchestration patterns. Vertex AI Agent Builder UI released for no-code agent creation targeting enterprise teams.
**Q4 2024:** Project Mariner (computer use) announced alongside Gemini 2.0 launch. Gemini 2.0 as the backbone for reasoning-heavy orchestration tasks requiring extended context.
**Q1 2025:** Vertex AI Agents SDK generally available with full GCP service integration. Dataflow pipeline integration for data-processing agents.
**Strategy:** Google leverages GCP's ecosystem (BigQuery, Cloud Storage, Cloud Tasks). Agent Builder is a no-code entry point, ADK is the developer SDK. Deep integration with Gemini reasoning models on GCP infrastructure.
---
## Anthropic
**Q3 2024:** Model Context Protocol (MCP) standardized and released as an open specification. MCP enables tool discovery and composition without prompt engineering.
**Q4 2024:** MCP server ecosystem grew to 50+ community servers. Claude Code released as a proof-of-concept agent operating environment.
**Q1 2025:** MCP becomes de facto standard for tool registration in agent systems. Anthropic's strategy is "protocol-first, not platform-centric" — MCP is the universal adapter, not an Anthropic-proprietary lock-in. Any orchestrator can work with any MCP tool.
**Strategy:** Anthropic avoids building a platform. Instead, MCP is the universal adapter. This is an intentional bet against proprietary lock-in and towards ecosystem standards — contrarian relative to OpenAI, Google, and Amazon.
---
## Microsoft
**Q3 2024:** AutoGen released v0.4 with improved memory management and long-horizon task execution. Semantic Kernel 1.13 added structured task decomposition and planning.
**Q4 2024:** Copilot Studio updated with multi-turn agentic flows. Azure AI Foundry released as unified agent management platform consolidating previous disparate Azure AI services.
**Q1 2025:** Semantic Kernel integrated directly into Azure AI Foundry. AutoGen's planning layer improved for chained multi-step reasoning. Copilot Studio adds 200+ connector integrations.
**Strategy:** Microsoft's bet is ecosystem breadth — agents that coordinate across Microsoft 365, Dynamics, Copilot Studio, and Azure services. Most enterprise-centric of the five. Every Microsoft product becomes an orchestration target.
---
## Amazon
**Q3 2024:** Bedrock Agents released with multi-step task execution. Built-in integrations with Step Functions, Lambda, and RDS. Knowledge Bases connector for RAG-powered agents.
**Q4 2024:** Bedrock Agents added support for tool use with Anthropic Claude and third-party models. Inline agent configurations for dynamic prompt overrides.
**Q1 2025:** AWS Step Functions + Bedrock Agents integration deepened. Agents can orchestrate long-running workflows natively within AWS infrastructure. Bedrock Flows launched for visual pipeline authoring.
**Strategy:** Amazon's strategy is managed infrastructure — agents run in AWS, orchestrate AWS services. The bet is that 99% of enterprise workloads will stay inside existing AWS infrastructure, making Bedrock Agents the path of least resistance.
→
L12
Agent Economic Coordination Systems
The protocols that enable agents to participate in economic activity — buying compute, selling data, coordinating tasks on open networks. This is the infrastructure layer that makes the agent economy a real economy, not just a metaphor.
→
Agent LayerOrchestration, memory, reasoning, and tooling
L11
Agent Orchestration Systems
How multiple agents coordinate. The framework layer defines the programming model for multi-agent systems — how you wire agents together, pass tasks between them, handle failures, and manage state across an agent network.
→
L10
Agent Evaluation & Monitoring
You can't trust what you can't observe. This layer covers the systems that grade agent performance, trace multi-agent runs, detect failures, and audit what agents actually did. Essential for production deployments.
→
L09
Agent Marketplaces
Markets where agents sell services to other agents and to humans. The emerging infrastructure for machine-to-machine commerce — agents discovering, hiring, and paying other agents for specialized tasks.
→
L08
Agent Browser & Computer Control
Agents that can use software like a human — browsing the web, filling forms, clicking UI elements, running applications. Computer control is the universal adapter that lets agents operate systems with no API.
→
ProtocolCommunication, interoperability, and human-in-the-loop
L07
Agent Data Access Systems
Agents need to query real-world data and APIs to be useful. This layer defines how agents access structured and unstructured data — from live APIs to vector databases to enterprise data warehouses.
→
L06
Agent Tool Registries
Where agents discover tools they can use. As the agent economy grows, the registry layer becomes the App Store equivalent — the index of callable capabilities. This is where Agentifact itself sits strategically.
→
L05
Agent Memory Systems
An agent without persistent memory is stateless — it starts fresh every session. This layer is what makes agents accumulate context, learn from previous interactions, and maintain coherent long-running relationships. It's the difference between a tool and an assistant.
→
FoundationIdentity, security, storage, and compute
L04
Agent Workflow & Automation Systems
Agents executing multi-step tasks across tools. The difference between an agent that answers questions and one that gets things done is a workflow layer — the orchestration of tool calls, state management, and conditional logic across many steps.
→
L03
Agent Hosting & Execution Environments
Where agents live and run continuously. Not just a server — the runtime that manages agent state, restarts, memory handoff between runs, and the operational infrastructure of always-on autonomous systems.
→
L02
Agent Communication Infrastructure
Agents talking to each other is different from agents talking to humans. This layer defines how agents pass messages, delegate tasks, and coordinate across systems — the protocols that make multi-agent systems possible.
→
L01
Agent Identity & Wallet Infrastructure
Agents need identity before they can transact. This layer is the foundation of the agent economy — without it, agents can't own assets, prove provenance, or pay for services autonomously.
→
Coming — Agent Economy Observatory
Benchmark tracker, infrastructure count by layer, capability progression, and enterprise platform changelog. Historical data is the moat — we're collecting it now.
View benchmark tracker →