Agentifact Best Guide
Best MCP Servers (2026)
The highest-scored mcp servers in the Agentifact index, ranked by composite trust score across 5 dimensions. Independent assessment — no paid placements.
Top 5 by composite score
All MCP Servers (358)
Amazon Bedrock AgentCore
Amazon Bedrock AgentCore is AWS's dedicated enterprise runtime for securely deploying and operating AI agents at any scale without infrastructure management. Provides complete session isolation, Amazon VPC connectivity, AWS PrivateLink support, managed memory, code execution, and observability. Complements Bedrock Agents with production-grade security controls. Usage-based pricing.
Amazon Bedrock Agents
Amazon Bedrock Agents is AWS's managed service for building and deploying enterprise AI agents on top of foundation models. Agents orchestrate multi-step tasks using APIs, enterprise data sources, and action groups with no infrastructure management. AgentCore adds enterprise-grade isolation, VPC connectivity, PrivateLink, and session management. Used by 100,000+ organizations. Usage-based pricing.
Docker MCP
Manage Docker containers and images via MCP. Start, stop, inspect, and build containers from agent workflows.
LiveKit
LiveKit is a real-time audio and video infrastructure platform and the foundation on which LiveKit Agents is built. It provides a global WebRTC media server network, server-side SDKs in Python, Go, Node.js, and more, plus browser and mobile client SDKs. For voice agent use cases, LiveKit handles multi-party audio rooms, direct phone integration via SIP/PSTN, and real-time media transport with <100ms latency. LiveKit Cloud offers managed hosting with usage-based pricing; the entire server stack is open source and self-hostable for teams that want full control over their media infrastructure.
Stytch
Developer-first identity platform providing authentication, authorization (RBAC, SCIM), MFA, SSO, and fraud detection for applications that need to authenticate both human users and machine/agent identities. Backend and client-side SDKs support fine-grained API access control. Free for the first 10,000 MAUs; $0.10/user/month after that.
Datadog MCP
Connects AI agents to Datadog for monitoring metrics, logs, and alerts through MCP.
Monday.com MCP
Monday.com work OS via MCP. Boards, items, columns, automations.
Stripe MCP
Official Stripe MCP server. Query customers, charges, subscriptions, and products. Full read/write access with careful permission scoping.
Intercom MCP
Customer messaging platform via MCP. Conversations, contacts, articles, events.
Cloudflare MCP Server
Cloudflare's managed MCP servers offer robust protocol support with strong enterprise backing, excellent sandboxing, and official docs, tempered by limited performance data and a vendor-related incident.
Tavily MCP
Search API built for AI agents. Returns structured, citation-ready results. Faster and more agent-friendly than general search APIs.
Linear MCP
Official Linear MCP integration. Create issues, manage projects, query team workflows. Clean API surface with good error handling.
Supabase MCP Server
Official Supabase MCP server excels in protocol support and security scoping but lacks performance data and full stability due to pre-1.0 versioning.
Temporal MCP
Durable execution for agents. Start workflows, signal running executions, query state. Best-in-class reliability for long-running agent tasks.
MCP GitHub Server
Official GitHub MCP server excels in MCP interop and trust but carries prompt injection risks requiring careful token scoping.
ThoughtSpot (Spotter AI Agents)
ThoughtSpot's Spotter is an enterprise agentic analytics platform with specialized AI agents: SpotterModel for governed semantic modeling, SpotterViz for automated dashboard generation, and SpotterCode for AI-assisted embed coding. Integrates with Snowflake, BigQuery, Redshift, and other warehouses. SpotCache enables unlimited AI analytics at fixed cloud cost. API-first with embed SDK.
Miro MCP
Miro whiteboard via MCP. Create sticky notes, shapes, connectors, and read board content.
PayPal MCP
PayPal payment operations via MCP. Create orders, capture payments, manage subscriptions.
Vertex AI Agent Builder
Vertex AI Agent Builder is Google Cloud's full-stack enterprise platform for building, deploying, and governing AI agents in production. Supports open frameworks like ADK and LangChain, with managed Agent Engine runtime, Sessions, Memory Bank, and Code Execution. Provides audit trails, granular IAM permissions, and secure-by-design compliance. Usage-based pricing per API call and compute.
PlanetScale MCP
Serverless MySQL via MCP. Branch databases, deploy schema changes, query with connection pooling.
Trigger.dev MCP
Background jobs and scheduled tasks via MCP. Reliable job queuing for agent-triggered async work.
Webflow MCP
Webflow CMS and site management via MCP. Query collections, create items, manage pages.
Tinybird MCP Server
Mature hosted MCP server with strong official docs, token-based security, and enterprise backing, ideal for real-time data agents but lacks detailed failure recovery and load testing evidence.
Prisma MCP
Prisma ORM operations via MCP. Introspect schemas, run queries, manage migrations from agent workflows.
Stripe Agents Toolkit
Stripe's official MCP-oriented toolkit for agent-triggered payments and billing operations with approval controls.
Grafana MCP
Enables AI agents to query dashboards, metrics, and visualizations from Grafana via MCP.
Plaid MCP
Financial data access via MCP. Account balances, transactions, identity verification. High security bar.
Microsoft Azure AI Foundry (Agent Service)
Azure AI Foundry Agent Service is a managed, pro-code enterprise runtime for deploying, orchestrating, and governing AI agents at scale. Connects models, tools, and frameworks into a single secure runtime with enterprise identity, VPC networking, content safety, and audit observability. Used by 80,000+ enterprises and 80% of Fortune 500 companies. Billed on Azure consumption.
PostHog MCP
Product analytics via MCP. Query events, feature flags, session recordings, experiments.
Trello MCP
Trello board management via MCP. Cards, lists, boards, attachments.
MongoDB Atlas MCP
Official MongoDB Atlas MCP server. CRUD operations, aggregation pipelines, full-text search from agent workflows.
LiveKit Agents
LiveKit Agents is an open-source framework and managed cloud platform for building real-time voice, video, and multimodal AI agents. The SDK handles the hardest parts of voice AI—streaming STT-LLM-TTS pipelines, turn detection, interruption handling, and telephony integration—with plug-in support for all major AI providers. Developers can self-host the entire stack on LiveKit Server (the most-used open-source WebRTC server) or deploy to LiveKit Cloud. The framework is free and open source; LiveKit Cloud charges usage-based rates for hosted inference and media routing with a free tier for development.
Neon MCP
Serverless Postgres via MCP. Branch databases, query with natural language, manage schemas. Good fit for agent-driven development workflows.
Shopify MCP
Official Shopify MCP server. Products, orders, customers, inventory management from agent workflows.
Atlassian MCP Server
Atlassian's official Remote MCP Server delivers robust MCP protocol support with enterprise-grade security and reliability, ideal for agentic AI integrations with Jira/Confluence, though beta status limits some maturity scores.
Amazon Comprehend
AWS NLP service with dedicated PII detection APIs (Detect PII, Contains PII) for identifying and redacting sensitive entities in text at scale. Supports real-time and batch processing. Used to prevent agent pipelines from logging or transmitting sensitive user data. Free tier: 50,000 units/month per API for 12 months. Then usage-based per 100 characters.
Linear MCP Server
Linear's official MCP server delivers solid MCP protocol support with strong trust signals from a mature company, though lacks performance data and full security scoring due to profile.
E2B
Secure cloud sandbox infrastructure for running AI-generated code in isolated microVM environments. Agents execute untrusted code without risk to host systems. Supports Python, JavaScript, and other runtimes. SDK integrates with any LLM. Usage-based billing per second of sandbox runtime (~$0.05/hr for 1 vCPU). Free Hobby plan includes $100 credit; Pro at $150/month.
Google Gemini Enterprise (formerly Agentspace)
Google Gemini Enterprise is an enterprise intranet search, AI assistant, and agentic workflow platform (formerly Google Agentspace). Enables knowledge workers to discover, create, share, and run AI agents grounded in organizational data. Integrates with Google Workspace, third-party SaaS, and internal data sources via APIs. Available through Google Cloud with custom enterprise pricing.
Confluence MCP
Atlassian Confluence knowledge base via MCP. Search pages, create content, manage spaces.
ClickUp MCP
ClickUp task management via MCP. Tasks, spaces, lists, goals, and time tracking.
HashiCorp Vault
Identity-based secrets management platform for securely storing, distributing, and governing access to API keys, tokens, and credentials used by AI agents. Fully API-driven with dynamic secrets, automatic rotation, and fine-grained access policies. Open-source self-hosted version is free; Vault Secrets SaaS free up to 25 secrets; Enterprise pricing custom.
Firecrawl
Firecrawl is a web data API that turns any website into clean, LLM-ready Markdown or structured JSON, powering AI agents and RAG pipelines with real-time web content. It handles JavaScript rendering, CAPTCHA solving, proxy rotation, and concurrent crawling behind a simple API. An AI Extract feature pulls structured data from pages using a plain-English prompt. A free tier offers 500 one-time credits; paid plans start at $16/month for 3,000 credits, scaling to $333/month for 500,000 credits.
AWS Lambda
AWS serverless compute platform for running agent logic, inference tasks, and event-driven AI workflows without provisioning servers. Supports Python, Node.js, and container images for packaging LLM inference with ONNX or lightweight models. Billed per request and per GB-second of execution. Free tier: 1M requests and 400K GB-seconds monthly. Popular for cost-efficient AI inference when GPU is not required.
Anchor Browser
Anchor is a cloud-hosted browser platform that lets AI agents interact with the web like a human, running up to 50,000 concurrent sessions in the cloud or on-premises. It is designed to be deterministic where possible — reverting to AI reasoning only when required at runtime — making it more reliable than fully LLM-driven approaches. Anchor has partnered with Cloudflare for verified browser agent support and is trusted by Groq, Unify, and Browser-use. Pricing is $0.01 per step with $5 in free monthly credits.
Perplexity MCP Server
MCP server connector for Perplexity API enabling web search. Allows agents to perform research without leaving MCP ecosystem.
Cohere North
Cohere North is a privacy-first enterprise AI agent platform deployable in private VPCs, on-premises, or via Cohere's secure Model Vault. Enables design and deployment of agent-powered workflows connected to enterprise tools via flexible APIs and built-in connectors. Features granular access control, continuous red-teaming, and compliance with GDPR, SOC-2, and ISO 27001. Subscription-based.
Playwright
Playwright is Microsoft's open-source browser automation framework supporting Chromium, Firefox, and WebKit with a single API across JavaScript/TypeScript, Python, Java, and .NET. It is the most widely adopted foundation for building browser-based AI agents, used directly by Stagehand, AgentQL, and dozens of agent frameworks. Playwright now includes native AI agent support via the MCP server and integrates with GitHub Copilot's Coding Agent. It is entirely free and open-source under the Apache 2.0 license.
Perplexity MCP
Real-time web search with citations via MCP. Research-grade search with source attribution.
Kubernetes MCP
Kubernetes cluster management via MCP. Deployments, pods, services, config maps. High capability, high risk — approval mode required.
Figma MCP
Allows AI agents to access Figma files, designs, and components through MCP protocol.
MCP Memory Server
Popular open-source MCP memory server with strong protocol support and performance but limited formal docs and enterprise trust signals.
Browserbase MCP
Cloud browser automation for agents. Managed Chrome instances with session replay, anti-detection, and proxy support.
Cerebras Code MCP
Enables AI agents to access Cerebras code generation and inference capabilities via MCP.
AssemblyAI
AssemblyAI provides production-ready speech-to-text and audio intelligence APIs used widely as the STT layer in voice agent stacks. Its Universal model supports both pre-recorded and real-time streaming transcription with speaker diarization, sentiment analysis, entity detection, and topic classification available as add-ons. The streaming STT API is purpose-built for low-latency agent pipelines with sub-500ms transcript delivery. Pricing starts at $0.15/hr (Universal) with $50 in free credits; streaming audio billed identically to batch, with audio intelligence features priced separately per hour of audio processed.
Pinecone MCP
Official Pinecone vector database MCP. Upsert and query embeddings directly from agent pipelines. Clean schema, reliable uptime.
Atlassian MCP
Connects AI agents to Jira, Confluence, and Bitbucket for project management via MCP.
MongoDB
MCP server for querying and analyzing MongoDB collections. Provides agent builders with database access tools for NoSQL data operations.
Figma MCP Server
Official Figma MCP server that feeds design tokens, component variants, text layers, and component properties directly to AI coding assistants including GitHub Copilot, Cursor, and Claude Code. Transforms Figma design files into machine-readable infrastructure. Supports two-way sync with Claude Code — publish designs from Claude to Figma AND pull designs into Claude. Part of Figma's broader Code Connect ecosystem bridging visual and implementation layers.
Firebase MCP
Provides AI agents access to Firebase services including Firestore, Auth, and Functions via MCP.
Jina Reader
Jina Reader is a free API that converts any URL to clean, LLM-friendly Markdown with a simple prefix: r.jina.ai/[URL]. It natively handles PDFs, images (via automatic captioning), and web pages across 29 languages, powered by the ReaderLM-v2 1.5B parameter model that supports documents up to 512K tokens. It is the fastest way to give an AI agent web reading capability without building scraping infrastructure. The Reader API is free with flexible rate limits; higher-volume usage is available via Jina's token-based billing.
Amplitude MCP
Behavioral analytics via MCP. Query user journeys, funnels, retention cohorts.
Airtable MCP
Airtable bases via MCP. Read records, create rows, update fields, manage views. Good for low-code agent workflows.
Make MCP
Make.com automation scenarios via MCP. Trigger scenarios, pass data, monitor runs.
Zapier MCP Server
Zapier MCP Server delivers robust MCP protocol support with massive app ecosystem and strong security, tempered by limited technical docs and recent minor incident.
Twilio Voice
Twilio Programmable Voice is the industry-standard telephony API for building voice AI agents that make and receive phone calls over PSTN and VoIP. It supports TwiML (XML-based call control), WebSockets for streaming audio to AI models in real time, IVR flows, conference calling, call recording, and ASR in 120 languages. Twilio is the underlying telephony layer used by many voice AI platforms including Retell AI and Vapi. PAYG pricing is $0.013/min outbound and $0.0085/min inbound (US); phone numbers start at $1.15/mo. Volume discounts and committed-use pricing are available for enterprise.
PagerDuty MCP
Incident management via MCP. Trigger alerts, acknowledge incidents, manage escalations.
Pipedream MCP
Pipedream MCP excels in MCP protocol support and massive integrations with strong security/compliance, though lacks performance data and detailed error docs.
BentoML
Unified inference platform for packaging, serving, and scaling AI models and multi-model pipelines in Python. Supports any model format and runtime, with built-in task queues, dynamic batching, multi-GPU orchestration, and distributed serving. BentoCloud provides managed compute for rapid production deployment. Used by agent builders to compose and serve LLMs, embeddings, and custom models as microservices.
n8n MCP
Trigger n8n workflows via MCP. Execute webhooks, pass data to automation pipelines, retrieve results.
Vercel MCP
Official Vercel MCP server. Deploy projects, manage environment variables, inspect deployments and build logs.
Pinecone MCP Server
Official Pinecone MCP server excels in protocol support and docs for dev tools but limited by early access status and lack of production readiness evidence.
GitHub Actions MCP
Trigger and monitor GitHub Actions workflows via MCP. List runs, check status, download artifacts.
LiteLLM
Open-source LLM proxy and SDK that routes agent calls to 100+ LLM providers through a single OpenAI-compatible API. Handles cost tracking, load balancing, rate limiting, guardrails, and spend controls across Anthropic, OpenAI, Bedrock, Azure, and more. Self-hostable for free; enterprise tier adds SSO, audit logs, and Prometheus metrics. Critical infrastructure for multi-provider agent cost management.
Modal
Serverless cloud platform that runs Python functions in isolated containers using gVisor sandboxing for secure execution of untrusted code. Agent builders use Modal to run AI-generated code without exposing host systems. Sub-second cold starts with per-second billing. Free tier includes $30/month compute credits; GPU compute from $0.59/hr for T4.
Vapi
Vapi is a developer-first voice AI infrastructure platform for building, testing, and deploying voice agents at scale. It provides an abstraction layer over STT, LLM, and TTS providers, letting developers swap in any combination of services via a single API. The platform handles real-time call orchestration, WebSocket streaming, tool calling, and structured output. Pricing is $0.05/min for Vapi's platform layer, with LLM, voice, and telephony billed separately from third-party providers; total production costs typically run $0.25–$0.33/min.
Bright Data
Bright Data is an enterprise-grade web data infrastructure platform providing proxies, web scraping APIs, and datasets for AI development. It offers 150M+ real-user IPs across 195 countries, pre-built scrapers for 120+ domains, a Scraping Browser for complex JavaScript-heavy pages, and a Web Unlocker for bypassing anti-bot systems. The platform powers AI model training, RAG pipelines, and agentic web access for 20,000+ companies. Pricing is usage-based starting at $0.001/record; datacenter proxies from ~$10.50/GB.
OpenRouter
Unified API gateway providing access to 300+ AI models from 60+ providers through a single OpenAI-compatible endpoint. Routes agent model calls with automatic provider fallback, latency-optimized routing, and cost comparison. Charges at provider rates with no markup and no subscription required — buy credits and pay as you go. Critical for agent systems requiring model diversity, redundancy, and cost optimization across providers.
Neon MCP Server
Strong MCP server leveraging Neon's stable infrastructure with excellent protocol support but limited performance and audit details.
Supabase MCP
Connects AI agents to Supabase for database queries, authentication, and edge functions through MCP.
Notion MCP Server
Reliable hosted MCP protocol server from established Notion with solid interop and docs, lacking detailed error/performance evidence.
Deepgram
Deepgram provides a high-accuracy, low-latency speech-to-text API built for production voice AI applications. Its Nova-3 model delivers real-time streaming transcription at $0.0077/min and batch transcription at $0.0043/min, with $150 in free credits to start. Beyond transcription, Deepgram offers text-to-speech, speaker diarization, sentiment analysis, and a Voice Agent API that bundles STT, LLM routing, and TTS into a single WebSocket session. The platform is widely used as the STT backbone inside Retell AI, Vapi, and Pipecat pipelines.
Asana MCP
Asana task management via MCP. Create tasks, manage projects, update assignees.
Sora
Sora is OpenAI's video generation model that produces cinematic-quality videos up to 25 seconds from text prompts. The API (Sora 2) is available through OpenAI's developer platform with usage-based pricing: $0.10/second for 720p standard and up to $0.50/second for 1024p Pro quality. A minimum $50 credit purchase is required for API access. ChatGPT Plus ($20/mo) and Pro ($200/mo) subscribers also get direct web access. Essential for AI builders who need OpenAI-ecosystem video generation with tight integration into existing GPT-based workflows.
ElevenLabs
ElevenLabs is a leading voice AI platform offering ultra-realistic text-to-speech, voice cloning, and a fully managed Conversational AI API. Developers can clone any voice from a short sample, stream audio with sub-200ms latency, and build end-to-end voice agents using the EVI (Empathic Voice Interface) SDK. The Conversational AI product handles STT, LLM turn management, and TTS in one hosted pipeline with support for tool calling and interruptions. Pricing ranges from a free tier (10,000 chars/mo) up to a $1,320/mo Business plan; API and enterprise pricing is usage-based by character.
Doppler
Centralized cloud secrets manager that syncs environment variables and API keys to agent infrastructure across environments. Provides a REST API for programmatic secret access, supports service accounts for machine identities, and integrates with CI/CD pipelines. Free plan for up to 3 users; Team plan at $21/user/month with no extra cost for service accounts.
Weaviate
Open-source AI-native vector database supporting hybrid dense and sparse search for agent memory, RAG, and knowledge retrieval. Deployable as self-hosted Docker/Kubernetes or via Weaviate Cloud (shared/dedicated). Integrates natively with LangChain, LlamaIndex, OpenAI, and Cohere. Serverless Cloud starts at $25 per 1M vector dimensions/month. Supports multi-tenancy for serving many agent instances from one cluster.
Pinecone
Managed vector database purpose-built for production AI and agent applications. Stores and retrieves high-dimensional embeddings for RAG, semantic search, and agent long-term memory with sub-millisecond query latency at billion-vector scale. Includes Pinecone Assistant for agent-based chat, and Pinecone Inference for managed embedding models. Free Starter tier; Standard plan from $50/month minimum.
Inngest MCP
Durable workflow triggers via MCP. Send events, cancel runs, query function history. Excellent for agent orchestration.
Replicate
Cloud API platform for running and deploying AI models without managing infrastructure. Hosts 100+ official models — LLMs, image generation, audio — always on with stable APIs, plus custom model deployment via Cog. Scales automatically from zero to high traffic. Agents call models via simple REST API. Billed per-second of compute only when models are running; no idle charges.
Cloudflare MCP
Official Cloudflare MCP server. Manage Workers, KV, R2, D1, and Durable Objects from agent workflows.
Brave Search MCP
Official MCP server for Brave Search API. Web and local search with privacy-first results. Reliable, fast, no rate limit surprises on paid plans.
Composio
Composio delivers robust agent tool integrations with strong interop and security but lacks latency data and explicit data portability.
Anthropic Computer Use
Anthropic's Computer Use is a capability that lets Claude control a desktop environment — moving cursors, clicking buttons, typing text, and taking screenshots — enabling fully autonomous computer-use agents. It is accessible via the Claude API as a tool alongside standard message APIs, with no additional setup beyond standard tool-use billing. Developers can build agents that automate any GUI task across any application or website. Pricing follows standard Claude API token rates (e.g., Sonnet 4.6 at $3/M input, $15/M output), with a small ~480-token system prompt overhead per call.
LiteLLM Proxy
Self-hosted OpenAI-compatible proxy server from LiteLLM that centralizes LLM access for all agents in a deployment. Manages virtual API keys per agent or team, enforces per-key spend budgets and rate limits, load-balances across multiple providers, and logs all requests for cost attribution. Prevents runaway agent costs and simplifies multi-provider LLM infrastructure. Free to self-host; enterprise from $250/month.
ZenRows
ZenRows is a web scraping toolkit that handles rotating proxies, anti-bot bypassing, fingerprinting, CAPTCHA solving, and JavaScript rendering behind a single API. It offers a Universal Scraper API, a Scraping Browser for Puppeteer/Playwright workflows with built-in anti-detection, and 55M+ residential proxies with geotargeting. A free trial covers 1,000 URLs with no credit card required. Paid plans start at $69/month for 250,000 credits; JavaScript rendering with premium proxies uses 25 credits per request.
Lakera Guard
Real-time API security layer for AI agents that detects prompt injections, jailbreaks, indirect prompt injection, PII leakage, toxic content, and malicious links. Integrates in under 5 minutes via a single API call wrapping any LLM call. Available as SaaS or self-hosted. Free tier available; enterprise and custom plans via demo request.
Qdrant MCP
Qdrant vector database MCP server. High-performance similarity search for agent memory and RAG systems.
Langfuse
Open-source LLM engineering platform providing traces, evaluations, prompt management, and metrics. Supports LLM-as-judge scoring, human feedback collection, manual labeling, and custom evaluation pipelines via API. Integrates with OpenAI, LangChain, LlamaIndex, and LiteLLM via OpenTelemetry. Cloud-hosted freemium with paid tiers from $29/month; self-hostable under MIT license.
Qdrant
High-performance open-source vector database and similarity search engine written in Rust, designed for billion-scale agent memory and semantic retrieval. Deployable via Docker, Kubernetes Helm chart, or Qdrant Cloud (managed). Supports ANN indexes, filtering, multi-vector payloads, and distributed horizontal scaling. Official client SDKs for Python, TypeScript, Go, Java, and Rust. Free Cloud tier available.
Milvus
Cloud-native open-source vector database built for massive-scale similarity search, maintained under LF AI & Data Foundation with Apache 2.0 license. Supports Lite (Python in-process), Standalone (single Docker), and Distributed (Kubernetes for billion-scale) deployment modes. Integrates with LangChain, LlamaIndex, OpenAI, and HuggingFace for RAG and agent memory pipelines. Managed via Zilliz Cloud.
Apify
Apify is a full-stack web scraping and automation platform with a marketplace of 10,000+ ready-made automation tools called Actors — serverless cloud programs that scrape data, automate web tasks, and run AI agents. It provides an MCP server so AI models like Claude can dynamically discover and use Actors as tools. Apify is trusted by Intercom, the European Commission, and thousands of developers, with SOC 2 Type II compliance. Free tier includes $5/month in credits; paid plans start at $29/month.
IBM watsonx Orchestrate
IBM watsonx Orchestrate is an enterprise multi-agent orchestration platform offering 100+ domain-specific AI agents and 400+ prebuilt tools. Supports both no-code drag-and-drop and pro-code agent construction with IBM Granite reasoning models. Integrates with enterprise HR, finance, procurement, and sales apps. Governed with centralized oversight, built-in guardrails, and automated policy enforcement.
Hyperbrowser
Hyperbrowser is cloud browser infrastructure for AI agents offering sub-second session launch times and support for 1,000+ concurrent sessions without latency degradation. It integrates with OpenAI's CUA and Claude Computer Use, includes built-in stealth features and automatic CAPTCHA solving, and provides Python and Node.js SDKs. Hyperbrowser is well-suited for large-scale, high-concurrency agentic workloads. Pricing is credit-based: one browser hour costs ~$0.10 and one scraped page costs ~$0.001.
Puppeteer MCP
Official browser automation MCP using Puppeteer. Navigate pages, fill forms, take screenshots, extract content.
Braintrust
AI observability and evaluation platform for production LLM systems. Runs experiments against datasets with automated scorers (LLM-as-judge, factuality, code-based), tracks regressions in CI/CD via GitHub Actions, and provides real-time production tracing. Multi-language SDKs (Python, TypeScript, Go, Ruby, C#). SOC 2 Type II, HIPAA, GDPR compliant. Free tier with paid plans from $249/month.
Nightfall AI
AI-native data loss prevention platform with a developer API for detecting PII, PHI, secrets, and credentials in text, files, and cloud data sources. Agents integrate via REST API to scan outputs before storage or transmission. Free plan covers 3 GB/month scanned; Enterprise plans offer custom volume limits and SLA support.
AlphaVantage MCP
Provides AI agents with financial market data, stock quotes, and technical indicators via MCP.
HeyGen
HeyGen is a leading AI avatar video generation platform enabling developers and marketers to create talking-head videos from text scripts using customizable AI avatars. Its developer API is billed by the minute (1 credit = 1 minute of avatar video), with plans starting from $5 for API credits. The Scale plan drops unit cost to $0.50/minute and unlocks translation and proofreading APIs. HeyGen supports video translation into 40+ languages and provides a full REST API for programmatic video generation at scale, making it the go-to choice for AI-driven video localization and personalization products.
Luma Dream Machine
Dream Machine by Luma AI is a text-to-video and image-to-video generation platform known for its cinematic motion quality and photorealistic output. It supports API access through a separate credit system from web subscriptions, with web plans from $9.99/month (Lite) to $94.99/month (Unlimited). A 5-second video costs 170 credits and a 10-second video costs 340 credits on web. Developers building AI creative tools can access Dream Machine programmatically; the Plus plan ($29.99/mo) is the minimum for commercial use.
Firecrawl MCP
Allows AI agents to crawl websites and extract structured data via MCP server.
Todoist
MCP server for natural language task management in Todoist. Enables agents to create, update, and query tasks programmatically.
Open Policy Agent (OPA)
CNCF-graduated open-source general-purpose policy engine for enforcing authorization rules across microservices, APIs, Kubernetes, and agent systems. Developers write policies in the Rego language and query OPA via REST API or Go library to make access control decisions for agent tool calls and resource access. Free and open source.
Cognigy
Cognigy.AI is an enterprise-grade conversational AI platform for building voice and chat agents deployed in contact center and customer service environments. It provides a low-code flow builder, a generative AI copilot layer, NLU, live agent handoff, omnichannel routing (voice, chat, messaging), and extensive CRM and CCaaS integrations. The platform runs on-prem or in the cloud and supports compliance requirements for regulated industries. Pricing starts at approximately $2,500/mo for enterprise subscriptions, with average annual contracts around $115,000; no self-serve or free tier is available.
Speechmatics
Speechmatics is a multilingual speech recognition API supporting 55+ languages and 69 translation pairs, designed for enterprise voice AI workloads requiring high accuracy across diverse accents and dialects. The API offers real-time streaming and batch transcription, speaker diarization, punctuation, and an enterprise real-time STT model with sub-second latency. It targets applications in contact centers, media, and voice agent post-call analytics. Pricing includes a free tier (8 hrs/mo, monthly reset), PAYG at approximately $0.0117/min, with automatic volume discounts above 500 hours; enterprise customers receive custom negotiated rates.
Jina AI MCP Tools
Offers AI agents access to Jina AI's search, embedding, and multimodal tools via MCP.
@modelcontextprotocol/server-filesystem
Enables AI agents to perform file system operations like reading, writing, and listing files through a secure MCP server.
Temporal
Durable execution platform for building reliable, fault-tolerant agent workflows and long-running AI pipelines. Automatically captures workflow state at every step — agents survive failures, retries, and infrastructure restarts without losing progress. MIT-licensed open source for self-hosting; Temporal Cloud managed service with $1,000 free credits on signup. Widely used for multi-step agent orchestration that must be resilient to chaos.
Redis MCP Server
MCP server providing access to Redis key-value stores through standardized tools. Enables fast caching and session storage for agents.
Turso MCP
Distributed SQLite via MCP. Create databases, run queries, manage schemas at the edge. Good for multi-tenant agent workflows.
Resend MCP
Developer-friendly email via MCP. Clean API, React email templates, excellent deliverability. Better DX than legacy email APIs.
Salesforce Agentforce
Agentforce is Salesforce's enterprise agentic AI platform that orchestrates autonomous agents across sales, service, and employee workflows 24/7. Developers build and deploy agents using Agentforce Builder with MuleSoft connectors, REST APIs, and the Atlas reasoning engine. Supports the full agent lifecycle—build, test, deploy, monitor. Pricing via Flex Credits, conversation-based billing, or per-user licensing.
Browserbase
Browserbase is a cloud browser infrastructure platform purpose-built for AI agents, providing managed headless browsers that are fast, reliable, and scalable. It integrates natively with Playwright, Puppeteer, Selenium, and its own Stagehand framework via the Chrome DevTools Protocol. Features include isolated browser sessions, CAPTCHA solving, stealth proxies, and SOC-2 Type 1 and HIPAA compliance. Plans start with a Free tier (1 concurrent browser, 1 hr/mo), Developer at $20/mo, and Startup at $99/mo.
Permit.io
Fine-grained authorization-as-a-service platform with native support for AI agent identity and access control. Assigns machine identities to agents, enforces permissions per tool and resource, and integrates with LangChain, PydanticAI, and MCP. Supports RBAC, ABAC, and ReBAC policies. Free tier for small teams; Pro and Enterprise tiers based on monthly active identities.
MCP Filesystem Server
The reference MCP implementation with excellent protocol support, dragged down by two 2025 CVEs and missing server-level changelog.
Beam
Serverless GPU infrastructure platform for AI inference and training with sub-10-second cold starts and instant autoscaling. Python-native API to define and deploy agent tasks, background jobs, and LLM inference pipelines across multi-cloud GPU fleets including H100s and 4090s. Open-source engine (Beta9) supports bring-your-own cloud. Scale-to-zero billing eliminates idle GPU costs for infrequent agent workloads.
Notte
Notte is a browser infrastructure platform for AI agents that provides remote browser sessions, serverless automation functions, authentication primitives, and a visual studio. It is model-agnostic, supporting OpenAI, Anthropic, Gemini, and others, and enables agents to navigate, extract data, and complete tasks at scale. Notte raised $2.5M pre-seed and is designed for production-grade agentic automation. A free tier includes 100 browser hours; paid plans start at $20/mo plus usage.
Firecracker
Open-source microVM monitor from AWS that provides KVM-based hardware-level isolation for untrusted workloads. Boots microVMs in 125ms with under 5MB memory overhead; supports 150 VMs/sec/host. Powers AWS Lambda and AWS Fargate. Developers use it to build agent sandboxing infrastructure where AI-generated code runs in fully isolated environments. Free, Apache 2.0 licensed.
Gladia
Gladia is an audio transcription and intelligence API built for real-time and async speech processing in agent pipelines. It supports multilingual transcription, speaker diarization, live streaming, and audio intelligence features like named entity recognition and summarization—all bundled into a single per-hour rate without add-on fees. The API handles pre-recorded and live audio with a unified interface, making it popular for meeting intelligence and voice agent post-call analytics. Pricing starts free (10 hrs/mo), with PAYG at $0.20/hr async and $0.25/hr real-time; enterprise plans include custom models and fine-tuning.
Label Studio
Open-source multi-type data labeling platform for building evaluation datasets and ground-truth sets for AI systems. Supports text, images, audio, video, and time-series annotation with ML-assisted labeling, configurable templates, and inter-annotator agreement scoring. Used by NVIDIA, Meta, and Cloudflare. Free open-source edition; enterprise plans via HumanSignal with cloud trial.
Sentry MCP
Query Sentry errors, performance data, and releases. Useful for agents involved in incident response.
Google Cloud Run
Fully managed serverless container platform on Google Cloud that natively supports hosting AI agents built with ADK, LangGraph, Dify, and other frameworks. Auto-scales from zero to handle traffic spikes with per-100ms billing. GPU support for serverless ML inference billed per-second. Official Google documentation covers agent deployment patterns. Free tier includes 180K vCPU-seconds and 2M requests/month.
Memory
Knowledge graph-based persistent memory system via MCP. Allows agents to store, retrieve, and query long-term memory across sessions.
D-ID
D-ID is an AI talking head video platform that animates still photos and generates realistic lip-synced videos using AI avatars driven by text or audio scripts. Its developer API offers per-second pricing at $0.40/second (standard) and $0.15/second (fast model), with a free 14-day trial including 3 minutes. The API Build plan starts at $18/month and deducts usage from the same balance as web studio usage. D-ID is widely used in conversational AI apps, AI tutoring platforms, and personalized marketing automation pipelines.
Scrapfly
Scrapfly is a web scraping API that combines anti-bot bypass (Cloudflare, DataDome, PerimeterX, Akamai, and 20+ systems), CAPTCHA solving, and cloud browser rendering into a single endpoint. It offers 130M+ rotating proxies from 120+ countries, AI-powered data extraction with LLMs, and screenshot capture. Scrapfly is designed for developers building data pipelines and AI agents that need reliable, high-volume web access. Pricing is credit-based with prepaid monthly plans plus pay-as-you-go overflow; a free tier is available.
Zendesk AI (Resolution Platform)
Zendesk's Resolution Platform provides enterprise-grade AI agents for customer and employee service, capable of resolving 80% of support issues autonomously via voice, chat, and email. Integrates with external knowledge sources (Confluence, SharePoint, Google Drive) via Knowledge Connectors. Exposes REST APIs for custom agent configuration and workflow integration. Used by ~20,000 enterprises.
Calendly MCP
Scheduling automation via MCP. Create booking links, check availability, manage event types.
Together AI (Video API)
Together AI is an AI infrastructure cloud that recently expanded into video generation, offering API access to models including MiniMax Hailuo, Google Veo 3.0, and OpenAI Sora 2 via OpenAI-compatible endpoints. Pricing uses per-model usage-based billing with transparent rates listed on the model pages. Together AI also hosts open-source models including Wan and Stable Diffusion variants. Existing Together AI customers can access video generation without changing their API integration pattern, making it a convenient choice for teams already using the platform for LLM or image generation workloads.
Browser Use
Browser Use is an open-source Python library that gives AI agents full control of a web browser, letting LLMs autonomously navigate, click, type, and extract data without pre-written scripts. It supports vision models (screenshot-based) and DOM extraction, and works with OpenAI, Anthropic, Google, and open-source models. The library has crossed 50,000 GitHub stars and is one of the fastest-growing AI open-source projects. The core library is free (MIT); Browser Use Cloud offers token-based pricing with $10 in free credits for new users.
Amazon Bedrock Guardrails
AWS managed service for adding configurable content safety controls to AI agents, including content filters, denied topic classification, PII redaction, prompt attack detection, and hallucination checking. Works with any foundation model including non-AWS models. Blocks up to 88% of harmful content per AWS benchmarks. Priced at $0.15 per 1,000 text units for content filters.
Portkey AI
Production AI gateway and observability platform that routes agent LLM calls across 1,600+ models with load balancing, fallbacks, retries, guardrails, and cost governance. Integrates natively with LangChain, LangGraph, CrewAI, and OpenAI Agents SDK so all model calls inherit routing and spend controls automatically. Logs and traces every request. Open-source gateway; managed cloud with usage-based pricing per recorded request.
Agora
Agora provides real-time voice, video, and AI engagement SDKs used by developers to embed low-latency communications into applications. Its Conversational AI Engine API enables developers to add AI voice agents to live voice channels, processing turn detection and AI responses in real time within Agora's global media network. The RTC SDK handles voice calling at $0.99 per 1,000 participant minutes; the Conversational AI Engine is priced from $0.0099/min of interaction. A free tier covers the first 10,000 minutes of video calling monthly, with additional free minutes for real-time speech-to-text during development.
Pega Infinity (Agent Experience)
Pega Infinity is an enterprise agentic transformation platform that orchestrates AI agents within deterministic workflow guardrails using Pega Agentic Process Fabric. Supports MCP and A2A protocols for third-party agent integration, with full audit trails via Agent Tracer. Agents built through the AI Agent Rule framework execute governed, predictable outcomes across regulated industries. Cloud-based subscription.
RunwayML
Runway is a professional AI video generation and editing platform offering Gen-3 Alpha and Gen-4 models for text-to-video, image-to-video, and video-to-video transformation. It exposes a developer API billed at $0.01 per credit, with Gen-3 Alpha costing 10 credits/second and the faster Turbo variant at 5 credits/second. Subscription plans start at $12/month with monthly credit allotments. Runway is widely used by AI product builders who need a stable, high-quality video generation API with predictable per-second billing.
Stagehand
Stagehand is an open-source AI browser automation SDK built by Browserbase that adds natural language control to Playwright-based browser automation. It combines the precision of code with the flexibility of LLM-driven instructions, making browser agents reliable and maintainable. Stagehand v3 is significantly faster on iframe and shadow-root interactions than previous versions. The SDK itself is free and open-source; running it at scale requires Browserbase cloud infrastructure, which starts at $20/month.
Vonage Voice API
Vonage (now part of Ericsson) offers a Programmable Voice API for building AI voice agents with WebSocket streaming, ASR in 120 languages, TTS, IVR, call recording, and real-time third-party AI integration. Its AI Studio provides a drag-and-drop builder for deploying voice AI virtual agents using Vonage's proprietary NLU engine without deep coding. Vonage recently partnered with AWS to support conversational agents via Amazon Nova Sonic and introduced MCP integration for AI-meets-programmable-communications workflows. Pricing is per-second usage-based starting at $0.014/min for developer voice features; free credits available on signup.
MLflow (Eval Component)
The world's most-downloaded open-source AI platform (30M+ monthly downloads) with a built-in evaluation module for LLMs, agents, and RAG systems. Provides 50+ metrics and LLM judges, dataset versioning for test cases, automated regression detection, and production monitoring. Framework and cloud agnostic under Apache 2.0; integrates with Databricks, AWS, Azure, and GCP.
Oxylabs
Oxylabs is an enterprise web data collection platform offering 177M+ IP proxies, a Web Scraper API with domain-specific endpoints, and AI-powered OxyCopilot that auto-generates scraping code from natural language prompts. It supports AI agent stacks with structured data extraction, JavaScript rendering, and geotargeted data collection. Oxylabs charges only for successful requests (2xx/4xx status codes), giving predictable billing. Web Scraper API plans start at $49/month; proxies are usage-based with volume discounts.
EdgeOne Pages MCP
Allows AI agents to deploy and manage static sites on EdgeOne Pages via MCP protocol.
LangSmith
LLM and AI agent observability platform by LangChain with built-in evaluation capabilities. Runs offline evals against curated datasets to catch regressions and online LLM-as-judge evals against live production traffic. Framework-agnostic Python, TypeScript, Go, and Java SDKs; supports self-hosted, BYOC, and managed cloud deployments. Free tier available with paid plans scaling by trace volume.
Jira MCP
Atlassian Jira via MCP. Create issues, manage sprints, query projects. Strong feature coverage.
Canva MCP
Enables AI agents to create, edit, and manage Canva designs programmatically via MCP.
Git
MCP server offering tools to read, search, and manipulate Git repositories. Enables code agents to interact with version control systems programmatically.
Microsoft 365 Copilot
Microsoft 365 Copilot is an enterprise AI productivity platform that embeds AI agents across Word, Excel, Teams, and Outlook, powered by the Work IQ intelligence layer. Enterprise agents are buildable via Copilot Studio or the M365 Agent SDK. Supports multi-channel agent deployment with org-wide knowledge grounding, semantic search, and governance controls.
AgentMail
Agent-native email identity and inbox infrastructure designed for autonomous workflows and machine-operated communication channels.
MiniMax MCP
Connects AI agents to MiniMax AI services for generation and processing tasks via MCP.
KServe
Kubernetes-native model serving platform providing standardized inference APIs for LLMs and predictive models across TensorFlow, PyTorch, ONNX, XGBoost, and more. Handles autoscaling, canary rollouts, A/B testing, health checking, and serverless inference on Kubernetes. CNCF sandbox project with vLLM and Hugging Face TGI backends. Production standard for model serving in cloud-native agent infrastructure.
Weaviate MCP
Weaviate vector search via MCP. Semantic search, hybrid search, and object management for RAG pipelines.
Invariant Guardrails
Python-native contextual security layer for MCP-powered and LLM-based agents. Deployed as an MCP or LLM proxy, it applies rule-based policies to intercept tool calls and detect PII, secrets, prompt injection, copyright infringement, and harmful content without invasive code changes. Open-source on GitHub. Free to self-host.
Bland AI
Bland AI is an enterprise-grade platform for automating high-volume AI phone calls with human-like conversational agents. It offers inbound and outbound calling, a visual flow builder, custom voices, real-time transfer, and deep CRM integrations. The API supports dynamic variables, live call injection, and webhook callbacks for every call event. Pricing follows a subscription-plus-usage model: a free Start tier at $0.14/min, Build at $299/mo ($0.12/min), and Scale at $499/mo ($0.11/min), with phone numbers at $15/mo each.
QuickBooks MCP
Intuit QuickBooks accounting via MCP. Invoices, expenses, customers, reports.
Hugging Face TGI
Text Generation Inference is Hugging Face's production-grade serving toolkit built in Rust and Python. Powers Hugging Chat and HF Inference Endpoints in production. Features tensor parallelism for multi-GPU serving, continuous dynamic batching, OpenTelemetry tracing, and Prometheus metrics. Ideal for teams deploying open-source LLMs — Llama, Falcon, Mistral — as backends for autonomous agent pipelines.
HubSpot MCP
CRM operations via MCP. Read contacts, companies, deals, and update pipeline stages. Requires OAuth setup.
Retell AI
Retell AI is a full-stack voice agent platform for building and deploying AI phone call agents. It provides an end-to-end API covering telephony, STT, LLM orchestration, and TTS in a single integration, with 60 free minutes and 20 concurrent calls on the base tier. Developers get real-time call control, phone number provisioning, custom LLM hooks, and knowledge base support. Pricing starts at $0.07/min for the voice engine with PAYG telephony costs on top, making total real-world cost roughly $0.13–$0.31/min depending on model and voice choice.
Redis MCP
Official MCP server for Redis. Get, set, list, and hash operations. Useful for agent state management and caching workflows.
Slack MCP Server
Slack's official hosted MCP server delivers enterprise-grade protocol compliance and security for AI agent Slack integrations, backed by comprehensive docs and Salesforce stability.
Cloudflare Workers AI
Serverless GPU inference platform running 50+ open-source AI models across Cloudflare's global edge network in 200+ cities. Zero infrastructure management — deploy LLM inference, embeddings, image classification, and speech-to-text as serverless functions. OpenAI-compatible API. Usage-based pricing with no idle costs; free tier included. Purpose-built for low-latency agent inference at the global edge.
Twilio MCP
SMS, voice, and WhatsApp via MCP. Full Twilio API surface for agent-driven communications. Approval mode recommended for outbound messages.
Synthesia
Synthesia is an enterprise AI avatar video creation platform with 230+ stock avatars and support for 140+ languages, designed for corporate training, marketing, and communications. The Creator plan at $89/month (or $64/mo annually) includes API access for programmatic video generation, 5 personal avatars, and branded video pages. Enterprise plans offer custom pricing with unlimited video minutes, SSO, and advanced collaboration tools. The API enables developers to automate personalized video creation at scale, and Synthesia is frequently used in L&D platforms and internal communications tooling.
SQLite MCP
Official SQLite MCP server. Local database operations with schema introspection. Ideal for local agent state.
Intercom Fin AI Agent
Intercom Fin is a purpose-built enterprise AI customer service agent averaging 66% conversation resolution and increasing 1% monthly. Fin 3 supports complex multi-step queries with procedure-based training, custom business logic, and deployment across any helpdesk or channel. Enterprise features include SOC 2, ISO 42001, HIPAA compliance, 99.97% uptime, and the patented Fin AI Engine.
Render
Developer-focused cloud platform for deploying web services, background workers, cron jobs, and databases with zero infrastructure management. Supports Docker containers and native runtimes, with automatic SSL, DDoS protection, pull request previews, and horizontal scaling. Free tier with 750 hours/month; Starter plans at $7/month. Ideal for hosting stateless agent microservices and API endpoints in production.
Notion MCP
Official Notion MCP server maintained by Notion. Search pages, create content, manage databases from agent workflows.
GitHub
MCP server for GitHub repository management and API access. Provides code collaboration tools for development agents.
NVIDIA NeMo Guardrails
Open-source Python toolkit for adding programmable guardrails to LLM conversational applications. Developers define rails in Colang to block off-topic queries, prevent unsafe outputs, enforce dialog flows, and apply content moderation. Includes Llama 3.1 NemoGuard 8B for content safety. Free to use; integrates with any LLM provider.
vLLM
High-throughput, memory-efficient inference engine for LLMs built at UC Berkeley. Uses PagedAttention and continuous batching to deliver up to 24x higher throughput than standard HuggingFace Transformers. Exposes an OpenAI-compatible API, supports multi-GPU and multi-node serving, and integrates with Docker, Kubernetes, and KServe for production agent deployments at scale. Fully open source, free to use.
Adobe Firefly (Video)
Adobe Firefly includes the Firefly Video Model, the first commercially safe AI video generation system for extending clips, generating B-roll, and producing video from text or image prompts, integrated into Premiere Pro and the Firefly web app. Subscription plans start at $9.99/month (Standard, ~20 five-second videos) and $29.99/month (Pro, ~70 videos). Adobe also offers Firefly Services for enterprise developers via a REST API. In December 2025, Adobe partnered with Runway to integrate Runway's generation models into Firefly workflows.
Chroma
Open-source AI-native vector database and search engine designed for fast prototyping through production deployment. Supports dense and sparse vector search, metadata filtering, and multi-modal retrieval across text and images. Embeds OpenAI, Google, Cohere, and HuggingFace models directly. Runs in-process for development or as a persistent server. Apache 2.0 licensed; Chroma Cloud for managed hosting.
XcodeBuildMCP
Allows AI agents to scaffold, build, run, and test iOS/macOS apps with UI automation via MCP.
LlamaIndex MCP Server
Creates MCP-compatible context servers from structured/unstructured data sources. Supports RAG pipelines with modular data loaders for agents.
Playwright MCP
Playwright MCP is an official Microsoft Model Context Protocol server that lets AI agents control web browsers using Playwright's automation capabilities. Instead of screenshots, it uses accessibility snapshots (roles, labels, states) so agents interact with pages semantically without requiring vision models. It is built into GitHub Copilot's Coding Agent and works with Claude Desktop, Cursor, VS Code, and other MCP-compatible tools. Playwright MCP is fully open-source and free under the Apache 2.0 license.
Microsoft Copilot Studio
Microsoft Copilot Studio is a low-code/pro-code enterprise agent builder for creating, governing, and scaling AI agents across Microsoft 365 and external systems. Agents integrate with Power Platform connectors, REST APIs, and Azure services. Supports natural language authoring as well as the Microsoft 365 Agent SDK for developers. Available as an add-on to Microsoft 365 subscriptions.
OpenAI Whisper
OpenAI's Whisper is a highly accurate, multilingual speech-to-text API available via the OpenAI platform, supporting 50+ languages at the same flat rate. The managed API handles audio files up to 25MB in mp3, mp4, wav, webm, and other formats, making it straightforward to add transcription to voice agent pipelines. GPT-4o Transcribe and GPT-4o Mini Transcribe are newer variants offering improved accuracy and cost options. Pricing is $0.006/min for Whisper and GPT-4o Transcribe, and $0.003/min for GPT-4o Mini Transcribe, with no volume tiers—ideal for moderate-volume use cases requiring broad language coverage.
Redis
In-memory data store widely used as a task queue backend, session store, and short-term agent memory layer. Supports pub/sub, sorted sets, streams, and list-based queues for dispatching agent jobs to worker processes. Redis for AI documentation covers agent infrastructure patterns including caching LLM responses and storing agent state. Free tier on Redis Cloud (30MB); paid plans from $5/month.
Everything
Reference MCP server providing prompts, resources, and tools for testing protocol capabilities. Enables agent builders to experiment with core MCP features like tool invocation and resource access.
Pulsetic MCP Server
Provides AI agents access to monitoring data, uptime status, and incident management via MCP.
Railway
All-in-one intelligent cloud provider for deploying web services, agents, databases, and workers from GitHub repos with zero config. Provides automatic scaling, PR preview environments, one-click rollbacks, built-in monitoring, and secrets management. Supports any Docker container or language runtime. Free tier available; paid plans scale usage-based with per-second billing making it cost-effective for async agent pipelines.
Mendix AI (Agent Builder)
Mendix AI is an enterprise low-code platform (Siemens subsidiary) for building AI-augmented applications and agentic systems via Agent Builder in Studio Pro. The Maia AI assistant and Mendix Agents Kit enable low-code composition of multi-agent workflows with flexible AI model deployment options. Provides governance controls over cost, security, and AI effectiveness. Free trial available.
Sentry
MCP server integrating Sentry.io for error tracking and performance monitoring. Enables agents to monitor application health and debug issues.
Inspect (AISI)
Open-source LLM evaluation framework by the UK AI Security Institute. Includes 100+ pre-built evals covering coding, agentic tasks, reasoning, knowledge, behavior, and multi-modal understanding. Features chainable solver components, MCP tool support, sandboxed execution, multi-agent primitives, and a VS Code extension plus web-based log viewer for monitoring runs. Free under open-source license.
Fly.io
Global application deployment platform that runs containerized agent workloads close to users across 30+ regions. Provides fast VM-based deployments with per-second billing, built-in Postgres and Redis, private networking, and auto-scaling. Free tier includes 3 shared-CPU VMs, 160GB transfer, and 3GB storage. Favored by agent developers for low-latency globally distributed deployments without managing Kubernetes.
Notte (Python SDK)
The Notte Python SDK is the programmatic interface to the Notte browser platform, enabling developers to write agent workflows in Python that spin up managed browser sessions, navigate pages, extract structured data, and complete web tasks. It is model-agnostic and integrates with any LLM provider. The SDK is open-source and available on PyPI; usage is billed through the Notte Cloud platform. A free tier includes 100 browser hours per month, with paid plans starting at $20/month plus usage.
Fetch MCP
Official HTTP fetch MCP server. Make GET/POST requests, convert HTML to markdown. Core utility for web-connected agents.
Steel
Steel is an open-source browser API built specifically for AI agents and apps, making it easy to run browser automations in the cloud — from large-scale scraping to fully autonomous web agents. Sessions start in under one second, can run up to 24 hours, and include built-in CAPTCHA solving. Steel supports Puppeteer, Playwright, and Selenium and can be self-hosted or used via Steel Cloud. The free tier includes 100 browser hours/month; paid plans scale up to 50 concurrent sessions at $499/mo.
ScrapingBee
ScrapingBee is a web scraping API that manages headless browsers, rotating proxies, and JavaScript rendering so developers can extract data without infrastructure complexity. It includes an AI extraction feature that retrieves structured data using plain-English descriptions instead of CSS selectors, making it suitable for LLM pipelines. A 1,000-request free trial is available with no credit card. Paid plans start at $49/month (Freelance) with higher tiers unlocking JS rendering, stealth proxies, and geotargeting.
Phoenix (Arize AI)
Open-source LLM observability and evaluation platform built on OpenTelemetry. Instruments AI applications across LangChain, LlamaIndex, OpenAI Agents SDK, LangGraph, and CrewAI to capture traces, then scores them with LLM-based evaluators, code checks, or human labels. Measures relevance, toxicity, retrieval quality, and custom metrics. 8.5k+ GitHub stars; self-hostable with no vendor lock-in.
Oracle AI Agent Studio (Fusion Applications)
Oracle AI Agent Studio for Fusion Applications is a no-code and low-code platform for designing and deploying AI agents across front and back office enterprise operations including Oracle ERP, HCM, and CX. Agents integrate with Fusion data and workflows via REST APIs. Part of Oracle's broader agentic AI platform announced for banking and enterprise sectors in 2026. Custom subscription pricing.
Filesystem
Secure MCP server for file operations with configurable access controls. Provides agent builders with safe local file read/write capabilities via MCP protocol.
Infisical
Open-source end-to-end secrets management platform for storing and syncing API keys, environment variables, and credentials across agent infrastructure. Supports dynamic secrets, automatic rotation, PKI, and fine-grained RBAC. Self-hostable under MIT license. Cloud-hosted free tier available; advanced features (rotation, dynamic secrets, HA) in paid tiers.
Mailchimp MCP
Email marketing via MCP. Manage audiences, create campaigns, query stats.
NVIDIA Triton Inference Server
Open-source production inference server supporting TensorRT, PyTorch, ONNX, OpenVINO, and Python backends. Runs on NVIDIA GPUs, x86/ARM CPUs, and AWS Inferentia. Exposes HTTP/REST and gRPC APIs with a dedicated model management API, Kubernetes health endpoints, and dynamic batching. Optimizes throughput for GPU-accelerated agent inference workloads across cloud, data center, and edge.
Pangea
Security-as-a-service platform offering a suite of API-based security services for AI applications including prompt injection guardrails, PII redaction, secrets management, audit logging, authentication, and IP/domain intelligence. All services are pay-as-you-go based on API call volume. Startup program provides up to $5,000 free usage; general pricing per pangea.cloud/pricing.
Fal.ai
Fal.ai is a high-performance AI inference platform specializing in fast video, image, and 3D model generation via a unified REST API. It hosts models including Kling, MiniMax Hailuo, Veo 3 Fast ($0.25/sec), Tripo3D, and many others with pay-per-use billing. A free tier with initial credits is available for new users. Fal.ai is known for its low-latency cold start times compared to other inference platforms, making it well-suited for real-time or interactive AI video generation features in developer products.
Hume AI
Hume AI builds the Empathic Voice Interface (EVI), a conversational AI API that understands and responds to human emotional cues in real time. EVI combines speech recognition, emotion detection from vocal prosody, and expressive TTS into a single streaming API, enabling agents that adapt their tone based on the caller's emotional state. The platform has received a major licensing deal with Google DeepMind, validating its research-grade emotion modeling. SDKs are available for React, TypeScript, Python, .NET, Swift, and more. Pricing ranges from free (10K chars/mo) to $70/mo Pro (1,200 EVI mins), with Scale at $200/mo and Business at $500/mo.
OpenAI Moderation API
Free API from OpenAI that classifies text and images for harmful content including hate speech, harassment, violence, self-harm, and sexual content. Powered by the omni-moderation model (GPT-4o based). Available free to all OpenAI API users with no usage limits counted against monthly quotas. Integrates via a single API call for agent output filtering.
Skyvern
Skyvern is an LLM and computer vision-powered browser automation platform that replaces fragile XPath and CSS selector scripts with AI that understands pages visually, the same way humans do. It works on websites it has never seen before without any per-site customization, supports 2FA/TOTP, CAPTCHA solving, and proxy networks. Skyvern achieved 85.85% on WebVoyager and excels at form-filling tasks. An open-source version is free; the cloud service is usage-based at $0.05 per step.
Elasticsearch MCP
Elasticsearch search and analytics via MCP. Index documents, run queries, manage indices.
Snorkel AI
Enterprise platform for programmatic data labeling and dataset curation for LLM fine-tuning and evaluation. Uses labeling functions and weak supervision to encode expert knowledge at scale, accelerating ground-truth dataset creation 2x faster than manual annotation. Supports RAG pipeline optimization and model evaluation workflows. Enterprise-only pricing; typical contracts start at $50k+ annually.
GitHub MCP
Official Anthropic-maintained MCP server for GitHub. Read repos, create issues, manage PRs, search code — full GitHub API surface via clean MCP tools.
PromptLayer
Prompt management and evaluation workbench for AI engineering teams. Enables visual prompt editing, version control, and regression testing against historical data via 20+ evaluation column types including LLM assertions and custom webhooks. Tracks cost, latency, and usage across OpenAI, Anthropic, Gemini, and Mistral. Free to start with paid tiers for team collaboration.
Hailuo AI (MiniMax)
Hailuo AI is MiniMax's video generation platform offering state-of-the-art text-to-video and image-to-video generation via both web app and Open Platform API. The latest Hailuo 2.3 model produces high-fidelity cinematic videos with stable motion and realistic expressions. API pricing is usage-based at approximately $0.045/second for 768p video, and the consumer Standard plan is $9.99/month for 1,000 credits. MiniMax also offers bulk and batch pricing at reduced rates, making it appealing for high-volume AI video production pipelines.
Ray Serve
Scalable model serving library built on Ray for deploying LLMs and agent pipelines as independent autoscaling microservices. Supports response streaming, dynamic request batching, prefix caching for multi-turn agent conversations, and fractional GPU allocation for cost-efficient multi-model hosting. Framework-agnostic and Python-native. Managed via Anyscale for production at enterprise scale.
pgvector
Open-source PostgreSQL extension that adds vector similarity search to existing Postgres databases. Store embeddings alongside relational data and query with exact or approximate nearest neighbor search using HNSW and IVFFlat indexes. Supports Postgres 13+, ACID compliance, replication, and point-in-time recovery. Eliminates the need for a separate vector database for agents already using Postgres as their data store.
Replicate (Video Models)
Replicate is a cloud platform hosting hundreds of open-source AI models including Wan2.1, Kling, CogVideoX, and many other video generation and 3D models, accessible via a simple REST API. Pricing is usage-based, billed per second of GPU compute time or per output unit depending on the model, with most video models in the $0.01–$0.10/second range. A free tier is available for initial experimentation. Replicate is the fastest way for developers to experiment with and productionize cutting-edge open-source video and 3D models without managing GPU infrastructure.
Jira Service Management (Atlassian Rovo)
Atlassian's Jira Service Management with Rovo AI agents enables enterprises to assign work to AI agents within ITSM workflows, automate incident triage, resolve support tickets autonomously, and draft knowledge base articles. Supports MCP integrations and third-party agent ecosystems. AI agents operate inline in Jira issues and comments. Subscription-based with per-agent pricing.
Activepieces MCP
Open-source automation platform via MCP. Trigger flows, pass structured data, get execution results.
SendGrid MCP
Send transactional emails via MCP. Template management, contact list operations, delivery tracking.
Crawl4AI
Crawl4AI is an open-source, async web crawler and scraper optimized for LLMs and AI pipelines. It generates clean Markdown for RAG systems, supports structured extraction via CSS, XPath, or LLM-based extraction, and offers advanced browser control with hooks and proxies. The project is the #1 trending open-source web crawler on GitHub and supports parallel crawling for high throughput. It is completely free and open-source under the Apache 2.0 license, with no API keys or paywalls required.
Microsoft Power Automate
Microsoft Power Automate is an enterprise workflow automation platform combining RPA, DPA, and AI flows. AI capabilities include Copilot-driven flow authoring, intelligent document processing, and 1,400+ prebuilt connectors to Dynamics 365, SAP, and Salesforce. Developers extend via custom connectors and APIs. AI-powered agents can trigger and execute across cloud and desktop workflows.
Segment MCP
Customer data platform via MCP. Track events, query profiles, manage audiences.
Context7 MCP Server
Popular, well-maintained MCP server from funded startup Upstash excels in protocol support and integrations but lacks detailed API reference and performance documentation.
Wan (Alibaba Tongyi)
Wan is Alibaba's open-source video generation model series (Wan2.1 and Wan2.2), released on Hugging Face and GitHub in early 2025. It includes text-to-video models in 1.3B and 14B parameter sizes, supporting 480P and 720P output, and is the top-ranked model on the VBench leaderboard. The 1.3B model requires only 8GB VRAM and can run on a consumer GPU. Wan2.2 uses a Mixture-of-Experts architecture for improved efficiency. Free to download and self-host under an open license, making it ideal for developers who want full control over video generation infrastructure.
Ultravox
Ultravox is a real-time speech-native voice AI API that provides an end-to-end voice agent infrastructure layer optimized for ultra-low latency. Unlike STT→LLM→TTS chains, Ultravox processes speech natively to reduce round-trip latency to levels comparable with GPT-4o Realtime but at a third of the cost. The API supports concurrent calls, tool calling, voice activity detection, and streaming responses via WebSockets. Pricing starts at $0.05/min with the first 30 minutes free; paid plans remove concurrency caps for production deployments.
Opik by Comet
Open-source LLM evaluation and observability platform by CometML. Traces agentic and RAG workflows, evaluates outputs with LLM-as-judge metrics including hallucination detection, answer relevance, and context precision, and integrates into CI/CD via pytest. Self-hostable via Docker or Kubernetes; handles 40M+ traces daily. Cloud-hosted free tier available with enterprise plans.
Diffbot
Diffbot is an AI-powered web data extraction platform that uses machine learning and computer vision to automatically structure data from any website without CSS selectors or custom parsers. Its Knowledge Graph connects 246M+ organizations and 1.6B+ articles as a queryable entity graph, ideal for building RAG pipelines and AI knowledge bases. The Crawlbot enables site-wide crawling, and natural language processing infers entities, relationships, and sentiment. Plans start at $299/month (Startup) up to $899/month (Plus) with enterprise custom.
LMNT
LMNT is a fast, affordable text-to-speech API optimized for real-time conversational applications such as voice agents, games, and interactive companions. It offers ultra-low latency streaming audio with sub-300ms time-to-first-audio, unlimited voice clones on all paid plans, and SDKs for Python and Node.js. LMNT is integrated directly into Vapi as a supported voice provider. Pricing tiers start at free (15K chars/mo), Indie at $10/mo (200K chars, $0.05/1K extra), Pro at $49/mo (1.25M chars), and Premium at $199/mo (5.7M chars), all with a commercial license.
Azure AI Content Safety
Microsoft Azure API for detecting harmful content in text and images generated by or fed into AI agents. Classifies content across hate, violence, sexual, and self-harm categories with severity scores. Integrates via REST API or SDK into any pipeline. Usage-based pricing at approximately $0.15 per 1,000 text units; free tier (F0) available for evaluation.
Workday AI Agents (Illuminate)
Workday Illuminate AI Agents are enterprise workforce and finance agents embedded in Workday's HCM and financial management platform. Agents automate recruiting, talent mobility, payroll, financial audits, and revenue contracts grounded in Workday's unified HR and finance data. The Agent System of Record tracks ROI and compliance. Generally available in 2026.
MCP Filesystem
Official reference implementation for local filesystem access. Read, write, list, and search files. The canonical MCP server for file operations.
@modelcontextprotocol/server-github
Provides AI agents with tools to access GitHub repositories, retrieve commits, issues, and perform repository operations via MCP protocol.
Promptfoo
Open-source CLI and library for LLM red-teaming, penetration testing, and vulnerability scanning of AI agents, RAGs, and prompts. Tests for 50+ vulnerability types including prompt injection, jailbreaks, PII leakage, and harmful outputs via declarative YAML configs. Integrates with CI/CD. Community plan free (10k probes/month); paid team and enterprise tiers available.
Tableau Pulse
Tableau Pulse is an AI-powered analytics experience embedded in Tableau Cloud that delivers personalized, proactive insights directly into Slack, Teams, email, and mobile workflows. AI detects drivers, trends, and outliers with natural language explanations using Salesforce's Agentforce Trust Layer. Included with all Tableau Cloud editions; premium Q&A capabilities in Tableau+ tier.
Chroma MCP
Open-source embedding database for agent memory. Simple API, local-first. Good for prototyping RAG systems.
MCP Git Server
Solid open-source MCP implementation for Git ops with strong protocol support but limited maturity in trust, security, and operational reliability.
Runway (Act-One)
Act-One is Runway's facial and performance capture feature within the RunwayML platform, enabling developers to transfer facial expressions and head movements from a reference video onto a generated AI character. It is accessible via the standard Runway API and subscription tiers (from $12/month). Act-One is used for creating expressive AI avatar animations for digital humans, interactive storytelling, and game cutscenes without requiring traditional mocap hardware. It integrates seamlessly into existing Runway Gen-3/Gen-4 video workflows.
Oracle OCI AI Agent Platform
Oracle OCI AI Agent Platform is a fully managed, cloud-native enterprise solution for building, deploying, and managing LLM-powered AI agents. Supports RAG against enterprise knowledge bases and can embed agents into Fusion Applications via AI Agent Studio. Developer APIs available through OCI documentation. Includes free tier with $300 in trial credits; pay-as-you-go pricing.
Celery
Mature, battle-tested distributed task queue for Python that dispatches agent tasks across worker fleets using Redis or RabbitMQ as message brokers. Supports real-time task execution, periodic scheduled tasks, retry policies, task chaining, and priority queues. Widely used in Django and FastAPI agent backends to handle async LLM calls, data processing, and background automation at scale. BSD-licensed and free.
Zhipu Web Search MCP
Provides AI agents with web search capabilities through Zhipu search service via MCP.
Cartesia
Cartesia is a real-time voice AI platform built specifically for low-latency agent applications, offering TTS, STT, and a voice agent platform (Line) under one API. Its Sonic-3 TTS model achieves 40–90ms time-to-first-audio and supports laughter, emotion, and 40+ languages with instant voice cloning from 3 seconds of audio. The unified credit system covers all three products—Sonic (TTS), Ink (STT), and Line (voice agent)—with plans scaling from a free hobby tier to custom enterprise. Usage-based pricing starts at $0.03/min for TTS, making it highly competitive for real-time voice agent builds.
Google Workspace Studio
Google Workspace Studio is a no-code AI agent builder embedded in Google Workspace that lets enterprise users create agents to automate workflows across Gmail, Drive, Docs, and Sheets using natural language. Agents handle tasks like intelligent prioritization, support triage, smart approvals, and content generation. Available across eligible Workspace enterprise editions.
Puppeteer
Puppeteer is Google's official Node.js library for controlling Chrome and Firefox via the DevTools Protocol or WebDriver BiDi. It is the foundational tool for headless browser automation in the JavaScript ecosystem and is used under the hood by many AI agent SDKs including Steel. Puppeteer supports screenshots, PDF generation, form automation, network interception, and full page interaction. It is free, open-source, and ships with a bundled version of Chrome for zero-setup automation.
Azure Container Apps
Fully managed serverless container platform on Azure with first-class support for agentic applications via Microsoft Foundry. Scales containers from zero based on HTTP traffic or events, with per-second GPU billing and scale-to-zero for AI workloads. Serverless GPU option includes NC T4 v3 and other GPU SKUs. Free tier: 180K vCPU-seconds, 360K GiB-seconds, and 2M requests per month.
OpenAI AgentKit (Agent Platform)
OpenAI's Agent Platform provides enterprises with Agent Builder (visual canvas for multi-agent workflows), Connector Registry (admin-governed data source management across ChatGPT and API), and ChatKit (embeddable chat interfaces). Built on the Responses API and Agents SDK. Supports GPT model orchestration, tool use, and handoffs. Usage-based pricing per token and API call.
Time
MCP server providing time and timezone conversion capabilities. Enables agents to handle temporal data and scheduling across different regions.
Windmill MCP
Open-source workflow platform via MCP. Run scripts, trigger flows, manage resources.
Google Drive MCP
Official Google Drive MCP server. Search, read, and manage files. Requires Google OAuth setup.
Midscene.js
Midscene.js is an open-source, vision-driven UI automation framework that controls web, Android, iOS, and custom GUIs using a unified JavaScript SDK backed by vision language models. Instead of CSS selectors, it localizes and interacts with UI elements using screenshots only, making it resilient to DOM changes. Midscene integrates with Playwright and Puppeteer for web automation and is recommended with Qwen-2.5-VL-72B for production (30–50% lower token use than GPT-4o). It is MIT-licensed and fully free to use.
MCP Fetch Server
Official MCP reference server strong in protocol interop/docs but held back by missing perf data and local access security warning.
Serper MCP Server
Enables AI agents to perform web searches and retrieve search results through MCP protocol.
MCP Sequential Thinking Server
Official reference MCP server with excellent protocol support and docs, limited by lack of production metrics and advanced features.
PixVerse
PixVerse is an AI video generation platform offering text-to-video and image-to-video creation with support for stylized, realistic, and anime aesthetics. It provides an API for developer integration and a web-based studio for consumer creation. The free tier allows limited daily generations, with paid plans offering more credits and higher quality output. PixVerse is popular among content creators and developers building AI video features who need diverse visual style support beyond photorealism, and it is available via Fal.ai for serverless inference.
DeepEval
Open-source LLM evaluation framework by Confident AI that runs like Pytest for AI systems. Provides 50+ research-backed metrics including G-Eval, hallucination detection, answer relevancy, task completion, and DAG-based agentic evaluation. Supports LLM-as-a-judge locally or via API, with CI/CD integration and a cloud platform for experiment tracking.
W&B Weave
Evaluation and tracing toolkit by Weights & Biases for GenAI applications. Automatically captures LLM call inputs, outputs, costs, and latency via the @weave.op decorator, then evaluates against datasets using custom or pre-built scorers measuring accuracy, latency, and cost. Supports side-by-side experiment comparison and CI/CD integration. Free tier with paid usage-based scaling.
Perspective API
Google API using machine learning to score the perceived toxicity, threat level, and conversational quality of text comments. Helps agent builders filter user inputs and outputs for harmful language. Free with a default quota of 1 QPS. Note: the service is scheduled to sunset at end of 2026; actively used projects should plan migration.
Howtocook MCP
Provides AI agents with recipe data and cooking instructions through MCP protocol.
ServiceNow AI Agents
ServiceNow AI Agents enable enterprises to deploy autonomous AI specialists that diagnose, plan, and execute multi-step IT and business workflows end-to-end. The Autonomous Workforce orchestrates role-specific agents such as L1 Service Desk Specialist and Security Operations Analyst. Integrates with Now Platform APIs, enterprise knowledge bases, and CMDB. Custom enterprise subscription pricing.
Krisp
Krisp is an AI-powered noise cancellation and meeting intelligence layer that removes background noise, echo, and crosstalk from voice streams in real time, making it a critical preprocessing component for voice agent call quality. Beyond noise cancellation, Krisp offers an AI Note Taker with transcription, summaries, and CRM sync in 16 languages, plus an accent conversion feature for reducing miscommunication. The SDK can be integrated into voice pipelines and call center software via API. Pricing: Free (60 min/day noise cancellation), Core at $8/mo (annual, unlimited), Advanced at $15/mo (annual, adds accent conversion and Salesforce integration).
Frontend Review MCP
MCP server that visually reviews AI agent design edits using screenshot comparison. Call reviewEdit with before/after screenshots and a description of the requested UI change — responds 'yes' if the edit visually satisfies the request, or 'no' with detailed explanation for further refinement. Uses Qwen2-VL-72B-Instruct from Hyperbolic (configurable). Works with browser-tools-mcp for screenshot capture. The missing verification layer for agentic UI builds.
Murf AI
Murf AI is an AI voice generator offering 200+ voices across 30+ languages, with a developer API (Murf Falcon) designed for low-latency voice synthesis in agent pipelines. The Falcon API delivers 55ms latency and is priced at $0.01/min or $0.03 per 1,000 characters PAYG, making it one of the most cost-efficient TTS options for high-volume agent deployments. The platform also includes a no-code voice studio for recording, editing, and syncing voiceovers to video. API access requires the Enterprise plan; individual creator and business plans ($19–$66/mo) cover the web studio with character quotas.
OpenAI Evals
OpenAI's open-source framework and benchmark registry for evaluating LLMs and LLM-based systems. Provides programmatic evaluation infrastructure, a growing library of community-contributed benchmarks, and direct integration with the OpenAI Dashboard for running evals via API. Model outputs are scored with custom or built-in graders. Actively used by OpenAI to guide model improvements.
BMC Helix
BMC Helix is an enterprise AI-powered ITSM, ITOM, and AIOps platform integrating HelixGPT for natural language interaction and agentic AI workflows. AI agents handle incident triage, knowledge drafting, change management, and proactive remediation across hybrid IT environments. Provides CMDB, asset management, and SLA monitoring with comprehensive REST APIs. Custom enterprise pricing.
Salesforce MCP
Salesforce CRM via MCP. SOQL queries, object CRUD, workflow triggers. Enterprise-grade but complex setup.
Mintlify
AI-native documentation infrastructure with machine-readable docs surfaces and MCP-oriented integration patterns.
mcp-server-calculator
Offers mathematical computation tools for AI agents to perform calculations through MCP server.
OutSystems Agent Workbench
OutSystems Agent Workbench is an enterprise low-code platform for the complete AI agent development lifecycle—creation, delivery, and monitoring—with built-in security and governance. Integrates any AI model, data source, or system with low-code abstraction. Developer Cloud supports AI Agent Builder for IT leaders to standardize and govern GenAI across applications. Custom pricing.
Airtable
Database platform with automation supporting human review/approval workflows. Agent builders use as backend for HITL task management and routing.
Zendesk MCP
Customer support ticket management via MCP. Create tickets, update status, search conversations.
Pipecat
Pipecat is an open-source Python framework by Daily for building real-time voice and multimodal conversational AI agents. It provides a pipeline architecture that chains streaming STT, LLM, and TTS services into a unified event loop with interruption support, turn detection, and multi-turn context management. Pipecat ships with 40+ service plugins (OpenAI, Anthropic, Deepgram, ElevenLabs, Cartesia, and more) and SDKs for Python, JavaScript, React, iOS, Android, and C++. The framework itself is fully free and MIT-licensed; compute costs come from the underlying AI service providers you connect.
Qdrant MCP Server
Official Qdrant MCP server excels in protocol support and documentation but lacks detailed failure semantics, performance benchmarks, and data portability features.
Hyper3D (Rodin)
Hyper3D (Rodin) is an AI 3D generation platform specializing in high-quality character and object mesh generation from text prompts and images, with a focus on game-ready topology and PBR texture output. It offers a developer API with credit-based pricing and a web studio. Rodin is particularly strong for generating humanoid characters and stylized assets suitable for real-time rendering in Unreal Engine or Unity. Plans include a free trial tier and paid subscriptions for production use, making it a competitive alternative to Meshy and Tripo for character-focused 3D pipelines.
TruLens
Open-source evaluation and tracing framework by Snowflake (acquired TruEra) for AI agents and RAG systems. Uses OpenTelemetry-based tracing combined with feedback functions to measure context relevance, groundedness, answer relevance, and safety metrics including bias and harmful language. Integrates via Python SDK or by ingesting existing OpenTelemetry traces.
Helicone
Open-source LLM observability platform and AI gateway that monitors, logs, and analyzes every agent LLM request through a single proxy integration. Tracks latency, cost, token usage, errors, and custom metadata across providers. Supports caching to reduce agent inference costs and rate limiting. Self-hostable for free; managed cloud includes 10,000 free requests/month with Pro and Enterprise tiers for production scale.
Bright Data MCP Server
Production-ready MCP server enabling AI agents with robust web access tools, backed by established company, strong interop but modest docs/versioning.
Zoom MCP
Schedule and manage Zoom meetings via MCP. Create meetings, retrieve recordings, manage webinars.
@modelcontextprotocol/server-sqlite
Allows AI agents to execute SQL queries and manage SQLite databases directly via MCP protocol.
PostgreSQL MCP
Official MCP server for PostgreSQL. Query, insert, and manage your database from agent workflows. Read-only mode available.
Ollama
Open-source tool for running and managing LLMs locally on developer hardware. Runs a local REST API server on port 11434 with an OpenAI-compatible interface, enabling agents to call models without cloud dependencies. Supports a growing library of open models including Llama, Mistral, Gemma, and DeepSeek. Designed for low-latency, privacy-first agent inference in development and edge environments.
Enkrypt AI
Comprehensive AI security platform covering detection, protection, monitoring, and compliance for LLMs and autonomous agents. Provides real-time guardrails for prompt injection, jailbreaks, PII exposure, toxicity, NSFW content, hallucination, and topic drift. API-based with support for multi-modal agent pipelines. Free trial available; enterprise pricing via consultation.
AgentQL
AgentQL is an AI-powered query language and SDK for connecting AI agents to the web, offering a natural-language alternative to fragile XPath and CSS selectors. It provides a REST API for public-facing data extraction without needing a browser, plus Playwright integrations and Python/JavaScript SDKs for full browser automation. Queries are self-healing and cross-site compatible, making agents resilient to page layout changes. Pricing tiers include Free ($0/mo), Professional ($99/mo), and Enterprise custom.
MySQL
MCP server integrating MySQL databases with configurable access controls and schema inspection. Enables SQL database interactions for data-driven agents.
Microsoft Presidio
Open-source PII detection and anonymization framework from Microsoft. Identifies, redacts, masks, and replaces sensitive entities (names, SSNs, credit cards, emails, etc.) across text, images, and structured data using NLP, regex, and rule-based recognizers in multiple languages. Integrates as a Python library, Docker container, or Kubernetes deployment. Fully free and self-hosted.
Xero MCP
Accounting and invoicing via MCP. Create invoices, query accounts, manage contacts.
Dolby.io
Dolby.io provides media processing and real-time communications APIs with broadcast-grade audio quality for voice agents, virtual events, and interactive applications. The Media Enhance API uses Dolby's audio science to denoise, normalize, and master audio programmatically, useful for improving voice recordings before STT processing. The Communications API powers WebRTC-based real-time voice and video with global CDN delivery. Pricing includes a free $50 credit on signup; Communications APIs start at $0.0025/min; Media processing APIs provide 200 free minutes/mo then $0.05/min; volume discounts apply as usage scales.
CalypsoAI
Enterprise AI security platform for red-teaming, runtime defense, and observability of LLM agents. Intercepts prompts and outputs via API to block prompt injection, data leakage, and malicious content in real time. Covers 80% of the OWASP Top 10 for LLMs. Model-agnostic and integrates with SIEM/SOAR workflows. Custom enterprise pricing only.
Guardrails AI
Open-source Python framework for adding input and output validation to LLM applications. Developers define validators from the Guardrails Hub that run as guards intercepting prompts and responses. Detects prompt injection, PII, toxic content, and off-topic outputs. Self-hosted version is free (Apache 2.0); managed Guardrails Pro uses usage-based pricing per validation operation.
Nuance (Microsoft)
Nuance, now part of Microsoft, provides enterprise voice AI solutions including Dragon speech recognition, the Nuance Healthcare Developer Platform, and the Azure-hosted Voice Live API for low-latency speech-to-speech voice agents. The Voice Live API (released mid-2025) unifies speech recognition, generative AI, and TTS into a single real-time interface with tiered Lite, Basic, and Pro tiers based on the underlying generative AI model. Dragon Medical One and related products target clinical speech documentation. Pricing varies by product: Azure Speech Services start at competitive per-minute rates; enterprise and healthcare products require custom contracts.
Gmail MCP
Read and send Gmail via MCP. Thread management, label operations, search. Sensitive — approval mode strongly recommended.
SoundHound AI
SoundHound AI is an independent voice AI company offering the Houndify developer platform and enterprise agentic voice solutions for automotive, restaurants, IoT, and customer service verticals. Its proprietary Speech-to-Meaning technology combines speech recognition and NLU in a single step for lower latency than sequential STT→NLU pipelines. Developers access voice AI capabilities, hundreds of pre-built content domains, and custom command creation via the Houndify API with an SDK for web, iOS, Android, and embedded devices. The developer tier is free for testing; enterprise solutions use custom usage-based pricing scaled to interaction volume.
Cypress
Cypress is a modern front-end testing and automation framework that runs directly in the browser alongside the application under test, providing real-time reloads, time-travel debugging, and automatic waiting. Its AI-powered features include test summaries, coverage analysis across every page and component, and CI test generation. Cypress is widely used to validate web automation scripts built by AI agents. A free-forever plan is available; paid Cloud plans with AI features and parallelization start at $67/month.
HubSpot Breeze AI Agents
HubSpot Breeze AI Agents is an enterprise CRM agent suite covering customer support, prospecting, and content creation. Agents integrate directly with HubSpot's CRM data, workflows, and contact records. Includes Breeze Studio for managing and customizing agent behavior with guardrails, approval workflows, and performance metrics. Available in Professional and Enterprise editions.
Seldon Core
MLOps and LLMOps platform for deploying, managing, and monitoring ML models on Kubernetes at enterprise scale. Packages models from any framework into production-ready microservices with real-time observability, drift detection, explainability, and canary deployments. MLServer is Apache 2.0 open source; Core 2 is source-available under BUSL; enterprise commercial tiers add production support and compliance features.
DeepMotion
DeepMotion is an AI motion capture platform that converts regular video footage into 3D character animations using markerless body and face tracking. The Animate 3D product supports real-time body, hand, and face tracking from monocular video, outputting BVH, FBX, or GLB animation files. Pricing starts at a free plan (60 seconds/month), with paid plans from $9/month (Starter) to $83/month (Studio, billed annually). A developer API is available for high-volume production use. DeepMotion is widely used in game development, VR avatar animation, and fitness app development pipelines.
Storybook MCP
MCP addon that exposes structured component metadata — lists, prop types, defaults, story-based examples, and documentation — to AI coding agents. Transforms Storybook from a documentation tool into machine-readable design system infrastructure. The Component Manifest reduces token consumption vs. loading entire codebases (50K-100K tokens per task) and enables autonomous correction loops where agents run component tests, see failures, and fix their own bugs.
Microsoft UFO
UFO (UI-Focused Agent) is an open-source Windows OS agent from Microsoft Research that uses GPT-Vision to understand and interact with native Windows application GUIs. It employs a dual-agent framework — one agent for high-level planning across apps and one for in-app execution — enabling seamless multi-application task completion. UFO³ Galaxy extends this to multi-device orchestration across heterogeneous platforms. The project is MIT-licensed, completely free, and integrates with any LLM provider via an API key.
Writer AI (AI HQ)
Writer is an enterprise AI platform for agentic work, trusted by Global 2000 companies. AI HQ provides a centralized hub for IT and business teams to build, activate, and supervise agents using a low-code Agent Builder, with 100+ prebuilt agents for finance, healthcare, retail, and HR. Powered by Palmyra LLMs and a Knowledge Graph for grounded enterprise answers. Custom pricing.
Scale AI (Evaluation)
Enterprise AI data and evaluation platform powering frontier LLM development with RLHF, human preference annotation, model red-teaming, safety testing, and capability benchmarking. SEAL lab runs private evaluations for AI labs including Meta and OpenAI. Offers self-serve pay-as-you-go data labeling alongside fully managed enterprise contracts for large-scale model evaluation pipelines.
Memory MCP
Official knowledge graph memory server. Store and retrieve agent memories across sessions.
Tripo AI
Tripo AI is a generative 3D platform that converts text prompts or single/multi-view images into high-quality 3D meshes via web app, API, and plugin integrations. The platform offers 300 free credits on signup, with paid plans unlocking commercial use and faster processing. The developer API is documented at platform.tripo3d.ai and supports image-to-3D, text-to-3D, and stylization tasks. Tripo3D v2.5 is also available via Fal.ai for fast serverless inference. It is frequently used by game studios and XR developers who need scalable 3D asset generation from image references.
Evidently AI
Open-source evaluation and observability platform with 100+ built-in metrics for LLM output quality, hallucination detection, PII leakage, RAG retrieval accuracy, toxicity, and sentiment. Generates evaluation reports, adversarial test datasets, and production monitoring dashboards. Supports custom LLM-as-judge metrics. Cloud platform with free tier; enterprise offers private cloud deployment.
Postman
MCP server for interacting with Postman API collections and environments. Allows agents to manage and execute API tests programmatically.
Vidnoz
Vidnoz is a free AI video generator offering over 1,900 AI avatars, 2,000+ voices in 140+ languages, and 2,800+ templates for fast avatar-based video creation. The free plan provides 3 minutes of generated video per day at 720p, with no credit card required. Paid plans start at $29/month (Pro) and $56.99/month (Business) for higher resolution, more daily minutes, and commercial rights. Vidnoz is well-suited for developers or startups looking for a low-cost entry point into AI avatar video generation with a usable free tier.
Freshservice AI (Freddy AI Agents)
Freshservice's Freddy AI Agents provide enterprise ITSM automation with human-like conversational support across Slack, Microsoft Teams, and service portals in 40+ languages. Features include AI-powered incident management, intelligent ticket routing, proactive root cause analysis, and visual analytics. Integrates via REST APIs with ITAM and ITOM systems. Subscription-based per agent/month.
Meshy AI
Meshy is a leading AI 3D model generation platform supporting text-to-3D, image-to-3D, and AI texturing workflows via both a web app and a developer API. The API is available on Pro plans ($20/month) and above, using a credits-based pay-before-you-go system. Generation tasks cost 10–20 credits each, with texturing priced at 10 credits. API documentation is available at docs.meshy.ai. Meshy supports export to OBJ, FBX, GLB, and STL formats, and is widely used by game developers, AR/VR creators, and AI product teams automating 3D asset pipelines.
Selenium
Selenium is the foundational open-source browser automation framework supporting Chrome, Firefox, Edge, and Safari across Java, Python, JavaScript, C#, Ruby, and Kotlin. It includes Selenium WebDriver for scripted browser control, Selenium Grid for distributed multi-browser test execution, and Selenium IDE for record-and-playback. The framework is widely used as the integration target for AI-enhanced testing tools and is the most battle-tested browser automation solution available. Selenium is completely free under the Apache 2.0 license.
LLM Guard
Security toolkit from ProtectAI with 2.5M+ downloads for sanitizing LLM prompts and responses. Input scanners detect prompt injection, PII, ban topics, and secrets; output scanners catch bias, malicious URLs, and deanonymization. Provides real-time safety and compliance filtering via Python library. Open-source core is free; enterprise support available from ProtectAI.
Gretel AI
Synthetic data platform (acquired by NVIDIA) purpose-built for generating agentic AI training and evaluation datasets. Produces high-quality tabular, text, and structured synthetic data with differential privacy guarantees via Gretel Navigator compound AI system. API-first with Python SDK, cloud playground, and integrations with BigQuery and Databricks. Usage-based cloud pricing with free sandbox.
Slack MCP
Official Slack MCP server. Send messages, read channels, manage threads. Requires workspace OAuth setup.
PyRIT
Microsoft's open-source Python Risk Identification Tool for automated red-teaming of generative AI systems. Security engineers use it to probe LLMs for harmful outputs, unsafe behaviors, and policy violations through single and multi-turn attack simulations. Supports Azure OpenAI, Hugging Face, and other model providers. Free, MIT licensed.
PromptBench
Microsoft's unified evaluation framework for testing LLM robustness against adversarial prompts. Generates adversarial inputs at character, word, sentence, and semantic levels to assess how vulnerable agent prompts are to attack. Covers 8 tasks and 13 datasets with 567,000+ test samples. Integrates via Python library. Free and open source.
Ragas
Open-source evaluation framework purpose-built for Retrieval-Augmented Generation pipelines. Measures context relevance, faithfulness, answer relevancy, and precision/recall without reference labels. Integrates with LangChain, LlamaIndex, and any LLM via Python SDK. Offers automated metrics and experiment tracking to iteratively benchmark RAG quality in CI workflows.
Spider
Spider is a high-performance web crawler and scraping API built in Rust, designed as the web data layer for AI agents and LLMs. It supports HTTP, Chrome CDP, and WebDriver rendering modes, and includes built-in stealth profiles that automatically handle Cloudflare, Akamai, and PerimeterX. Spider outputs clean Markdown for direct LLM consumption and offers pay-as-you-go pricing with no subscriptions. At roughly $0.48–$0.65 per 1,000 pages with no credit multipliers, it is one of the most cost-effective scraping APIs available.
SAP Joule
SAP Joule is an enterprise AI assistant and agent platform embedded across SAP applications for finance, HR, procurement, and supply chain. Joule Studio (GA Q1 2026) lets developers build custom Joule skills and agents with a low-code builder. Features 2,400+ prebuilt skills, 350 AI capabilities, Microsoft 365 Copilot integration, and GDPR-compliant data isolation within SAP tenancies.
Kling AI
Kling AI is Kuaishou's next-generation AI video generation platform capable of producing up to 1080p, 30fps videos from text or image inputs with realistic motion and physics. It offers an official developer API via the Kling AI developer portal for text-to-video, image-to-video, lip sync, and effects tasks with pre-paid resource packages. Consumer subscriptions range from a free tier (66 daily credits) to $10/month Standard through $92/month Premier. A strong choice for developers needing high-fidelity video generation at competitive API rates.
Appian AI Agents (Agent Studio)
Appian AI Agents lets enterprises build, monitor, and deploy AI agents through a guided low-code Agent Studio embedded in Appian's process orchestration platform. Agents connect to unified data fabric, enterprise APIs, and automation tools, executing multi-step tasks with full audit trails and compliance controls. Recognized as a Gartner Magic Quadrant Leader for low-code platforms.
OpenAI Operator (CUA)
OpenAI's Operator is an AI agent powered by the Computer-Using Agent (CUA) model, combining GPT-4o vision with reinforcement-learning reasoning to autonomously navigate browsers and complete web tasks like form filling, shopping, and research. It achieved 87% on WebVoyager and 58% on WebArena. For end users, Operator is available in ChatGPT Pro ($200/mo) and is now integrated as a core ChatGPT agent. The CUA model is also available as a research preview in the Responses API for developers in usage tiers 3–5 at $3/M input and $12/M output tokens.
Rube
MCP server connecting AI tools to 500+ apps including Gmail, Slack, GitHub, Notion. Provides unified access to multiple SaaS platforms for agents.
Descript
Descript is an AI-powered video and podcast editor that lets users edit media by editing the transcript, with features like filler word removal, AI voice cloning, studio sound enhancement, and AI-generated B-roll. As of September 2025, pricing is based on Media Minutes and AI Credits: the free plan includes 60 media minutes plus 100 AI credits, while the Creator plan is $35/month with 1,800 media minutes and 800 AI credits. Descript is highly relevant for AI builders automating content production pipelines or building on top of its overdub voice cloning API.
Limina AI (Private AI)
Enterprise PII, PHI, and PCI detection and de-identification API supporting 50+ entity types across 52 languages and formats including text, PDFs, images, audio, and structured data. Deploys as a container within the customer's own VPC with no external data transfer. Designed for HIPAA, GDPR, and PCI-DSS compliance in agent pipelines. Custom enterprise pricing.
Garak
Open-source LLM vulnerability scanner from NVIDIA for automated red-teaming of AI agents and models. Probes for jailbreaks, prompt injection, hallucination, data leakage, misinformation, and toxicity across 18+ attack methods. Supports OpenAI, Hugging Face, LiteLLM, REST APIs, and local GGUF models. Free to use; outputs JSONL vulnerability reports.
Luma Genie
Luma Genie is Luma AI's text-to-3D generation tool that creates 3D models from text prompts, currently in public beta and free to use with no generation limits. Each prompt returns four distinct 3D model variations for selection. Genie is accessible via the Luma web app, iOS app (which also captures real-world 3D via device camera), and Discord community. The tool is designed for rapid prototyping of 3D assets for games, AR/VR, and e-commerce visualization, and integrates with the broader Luma AI ecosystem including NeRF-based scene capture and Dream Machine video generation.
HarmBench
Standardized evaluation framework for automated red-teaming of LLMs developed by the Center for AI Safety. Contains 510 curated harmful behavior test cases across four categories and 18 adversarial attack modules. Used to benchmark LLM refusal robustness and compare defenses. Open-source on GitHub under the Center for AI Safety. Free research tool.
WordPress MCP
WordPress REST API via MCP. Posts, pages, media, users, taxonomies.
Pika
Pika is an AI video generation platform that converts text prompts, images, and existing video clips into short-form video content. It offers a browser-based studio and a public API accessible via third-party providers like Fal.ai and Wavespeed. Pricing starts with a free plan (80 credits, watermarked), with paid plans from $10/month (Standard) to $95/month (Fancy) for commercial use and faster generation. Ideal for developers prototyping AI video features or content automation pipelines.
BullMQ
Fast, robust Redis-based job queue and message queue library for Node.js, Python, Elixir, and PHP. Provides exactly-once queue semantics, horizontal worker scaling, priority queues, rate limiting, delayed jobs, and cron scheduling. Designed for high-throughput async agent task dispatch — offloading LLM calls, tool executions, and background processing from synchronous agent request handlers. MIT licensed and free.
Resemble AI
Resemble AI is a voice cloning and synthesis platform providing a developer API for real-time TTS, voice localization, and deepfake detection. It offers two voice clone types—Rapid Clones (from a few seconds of audio) and Professional Clones (higher fidelity, more training data)—suitable for production voice agents and interactive apps. The API exposes streaming audio with sub-500ms latency and supports custom pronunciation, SSML, and emotional tone control. Pricing starts at $19/mo (Creator, 15,000 seconds) up to $699/mo (Business, 360,000 seconds) with full API access; a free tier offers 150 seconds to test.
ZenML MCP Server
New but solid MCP server implementation from established MLOps company ZenML, strong protocol support and docs but lacks production maturity evidence like load testing and detailed failure modes. [GitHub Repo](https://github.com/zenml-io/mcp-zenml)
Vigil
Open-source Python library and REST API for scanning LLM prompts and responses against multiple security detectors. Uses vector database similarity search against known attack patterns, YARA heuristic rules, transformer-based classifiers, and canary token checks. Suitable for research and experimental agent hardening. Fully free and self-hosted.
PromptArmor
Prompt injection detection service that uses carefully designed LLM prompting strategies to identify and remove injected instructions from agent inputs. Achieves sub-1% false positive and false negative rates on the AgentDojo benchmark using GPT-4o. Publishes security research on real-world indirect prompt injection vulnerabilities in AI tools like Slack AI and Google Antigravity.
Daily Bots
Daily Bots is a hosted voice AI bot platform built on top of Pipecat and Daily's real-time media infrastructure, enabling developers to ship conversational voice agents with any LLM. It provides a managed cloud for running Pipecat pipelines with automatic scaling, global infrastructure, built-in telephony, and WebRTC transport. Partner integrations with Anthropic, Cartesia, Deepgram, ElevenLabs, and Together AI are pre-wired, or developers can supply their own API keys. Pricing is $0.02/min for Daily's media layer, with STT (Deepgram) at $0.0059/min billed at-cost on top.
Topaz Video AI
Topaz Video AI is a desktop application for professional AI-powered video enhancement, specializing in upscaling footage to 4K or 8K, frame interpolation to high frame rates, denoising, and stabilization. It uses trained diffusion models to hallucinate missing detail during upscaling. Pricing moved to subscription-only in late 2025, starting at $25/month or ~$299/year. While primarily a desktop tool rather than a cloud API, it is widely used by developers and studios preprocessing video for AI training datasets or production-quality output.
Typeform MCP
Form and survey management via MCP. Create forms, retrieve responses, manage response data.
MultiOn
MultiOn is an AI agent platform that autonomously completes web tasks on behalf of users — booking flights, ordering food, filling forms, and navigating complex multi-step workflows. It exposes an Agent API allowing developers to embed web-task automation into their own applications, with agents that plan and execute actions end-to-end. MultiOn is accessible via a Chrome extension, mobile app, and API. The platform operates on a freemium model with a free tier and paid plans for higher usage limits and faster processing.
Rokoko
Rokoko is a motion capture and character animation ecosystem combining hardware suits, AI-based video motion capture (Vision), and Rokoko Studio software for recording, editing, and exporting animation data. The Studio software supports a full-featured API and real-time streaming to game engines and DCC tools (Maya, Blender, Unreal, Unity) via plugins. Rokoko Studio Pro starts at $24/month (annual billing). The Vision AI mocap feature processes video input without wearable hardware, making it accessible for smaller teams. Rokoko is the industry standard for affordable, production-ready motion capture pipelines.
Argilla
Open-source collaboration platform (now part of Hugging Face) for building high-quality datasets for LLM fine-tuning, RLHF, and evaluation. Combines human expert annotation with AI-assisted suggestions and active learning to curate training and ground-truth evaluation sets. Integrates with LangChain and the Hugging Face ecosystem. Open-source with cloud-hosted option.
Move AI
Move AI is a markerless motion capture platform that converts standard multi-camera video into high-quality 3D human motion data for animation, sports analysis, and avatar rigging. It offers pay-per-use API pricing starting at $0.012/second for the s1 model and $0.024/second for the m1 model, with a $0.10 minimum per task. Subscription plans start at $50/month for individuals, with Move Pro offering custom multi-camera setups for studios. A free exploration tier is available. Developers use Move AI to build body tracking features in fitness apps, game animation pipelines, and metaverse avatar systems.
OpenAdapt
OpenAdapt is an open-source generative process automation framework that learns from human GUI demonstrations and replicates them using large multimodal models (LMMs). It records user interactions, trains on them, and then executes the same workflows autonomously across desktop and web applications. Built-in PII/PHI redaction ensures privacy compliance, and it supports OpenAI, Anthropic, Google, Ollama, and vLLM. OpenAdapt is MIT-licensed and completely free; users only pay for the LLM APIs they connect.
YData
Data-centric AI platform for synthetic data generation, profiling, and quality improvement. Generates statistically accurate synthetic datasets for tabular and time-series data to augment training sets, enable GDPR-compliant data sharing, and accelerate model testing. Ranked #1 in accuracy across 2023-2025 benchmarks. Available via Azure/AWS Marketplace; open-source ydata-synthetic SDK on PyPI.
Autodesk Flow Studio (Wonder Dynamics)
Autodesk Flow Studio (formerly Wonder Studio / Wonder Dynamics) is a cloud-based AI VFX platform that transforms live-action footage into fully animated, composited CG character scenes with automatic motion capture, lighting matching, and clean-plate generation. Pricing uses a credit system: a production plan at ~$1,000/year includes 2,000 credits (covering ~150 seconds of live-action/animation output), 4K export, and custom character uploads. Enterprise custom pricing is available. Developers and VFX teams use it to dramatically reduce the time and cost of integrating digital characters into live footage.
Rime AI
Rime AI provides natural, conversational text-to-speech models engineered for voice agent deployments where humanness and authenticity matter. Its Arcana v3 model captures natural speech patterns including breath, pacing, and emphasis, with time-to-first-byte around 175ms for standard tiers and sub-100ms for enterprise. The API supports English, Spanish, French, and German with 40+ voices spanning multiple regional accents, all accessible via a REST and streaming WebSocket API. Rime is popular in IVR, customer service, and outbound calling stacks. Pricing is tiered (Starter, Growth, Enterprise) with custom enterprise rates available on request.
Rebuff
Open-source prompt injection detector from ProtectAI with a four-layer defense: heuristics to filter suspicious inputs, an LLM-based classifier, a vector database of known attack embeddings, and canary tokens to detect prompt leakage. Integrates via Python SDK. Currently a prototype suitable for research and early-stage agent hardening. Free and self-hosted.
Hamming AI
Hamming AI is a QA and monitoring platform built specifically for testing voice AI agents before and after deployment. It auto-generates test scenarios from production call logs, replays call transcripts, and scores agents against 50+ customizable quality metrics covering accuracy, tone, safety, and task completion. Teams can catch regressions across prompt changes and model updates without manual call review. Hamming has tested over 4 million calls and integrates directly with platforms like Retell AI and Vapi. Pricing is custom and contact-based, with a free tier offering 100 test calls to get started.
MCP Advisor
Provides AI agents with advisory tools and recommendations through MCP server.
Fetch
MCP server for fetching web content and converting it for efficient LLM processing. Allows agents to access and parse external web data through standardized MCP tools.
Play.ht
Play.ht is a voice AI platform offering text-to-speech, voice cloning, and a conversational AI API supporting 142 languages. Voice cloning requires only 30 seconds of audio, and the PlayHT 3.0 model produces ultra-realistic voices optimized for interactive applications. The API supports SSML, streaming audio output, and batch processing with export in MP3, WAV, and OGG. A Conversational AI product enables developers to deploy low-latency voice agents. Pricing starts free (12.5K chars/mo), with the Creator plan at $31.20/mo (600K chars) and an Unlimited plan at $99/mo for commercial use.
MOSTLY AI
Privacy-safe synthetic data platform for generating high-fidelity tabular and text datasets for AI testing, training, and evaluation. Uses the TabularARGN model with built-in differential privacy. Provides an open-source Python SDK (Apache 2.0) for self-hosted generation and a cloud platform with collaboration tools. Free tier with 2 credits/day; paid plans for teams and enterprise Kubernetes deployment.
Sequential Thinking
MCP server for dynamic reflective problem-solving through structured thought sequences. Helps agents break down complex tasks into manageable reasoning steps.
AgentBench
Comprehensive benchmark from Tsinghua University (ICLR 2024) for evaluating LLMs as autonomous agents across 8 environments: operating systems, databases, knowledge graphs, web shopping, web browsing, card games, household tasks, and puzzles. Tests multi-turn reasoning and decision-making via containerized task workers configured through YAML. Free and open-source under Apache 2.0.
CapCut
CapCut is ByteDance's AI video editor used by over 200 million creators, offering AI-powered auto-captions, background removal, camera tracking, vocal isolation, and text-to-video generation features. The free plan supports 1080p export; the Pro plan at $9.99/month unlocks 4K export, expanded AI tools, and cloud sync. While CapCut does not publish an official public API, it is widely used for consumer-grade video automation and integrates deeply with TikTok's ecosystem, making it relevant for developer teams building social video tooling.
Loom MCP
Video messaging via MCP. Search recordings, get transcripts, create sharing links.
CogVideoX
CogVideoX is an open-source text-to-video and image-to-video diffusion model from Zhipu AI (formerly THUDM), generating 10-second videos at 768x1360 resolution and 16fps. The CogVideoX1.5-5B series supports higher-resolution and flexible input sizes. The project is available on GitHub and Hugging Face, with LoRA fine-tuning support and the CogKit framework for training and inference. Free to self-host, CogVideoX is a strong choice for developers building custom video generation workflows who need an open-weight model with active research support.
GAIA Benchmark
Benchmark from Meta AI and Hugging Face evaluating general-purpose AI assistant capabilities on 932 real-world tasks requiring reasoning, web browsing, multi-modal handling, and tool use across three difficulty levels. Human baseline is 92% vs ~15% for GPT-4 with plugins, making it a rigorous measure of agent capability gaps. Free and open-access with public leaderboard.
Shap-E (OpenAI)
Shap-E is OpenAI's open-source text-to-3D and image-to-3D generation model, released under the MIT license. It uses a diffusion process over implicit neural representations to generate 3D shapes and textures in seconds from text descriptions. The GitHub repository includes Jupyter notebook examples for text-to-3D and image-to-3D workflows. Shap-E is free to run locally or on cloud GPUs and integrates with PyTorch-based ML pipelines. It is particularly relevant for researchers and developers building 3D generation features who want an open-weight model they can fine-tune or deploy without licensing restrictions.
WebArena
Self-hostable benchmark environment for evaluating autonomous web agents on realistic multi-step tasks. Simulates functional e-commerce, content management, GitLab, and map services. Agents are scored on functional correctness of task completion using programmatic validators. Covers planning, reasoning, and multi-turn interaction. Fully open-source with Docker support; free to self-host.
Vidu AI
Vidu is a Chinese AI video generation platform supporting text-to-video, image-to-video, and reference-to-video creation modes, with an official MaaS API launched in February 2025. It generates videos in as little as 10 seconds using advanced inference acceleration. Consumer plans range from a free tier to $8–$79/month (yearly billing) with 800 to 8,000 credits. The API is hosted at platform.vidu.com and supports enterprise partnerships for advertising, e-commerce, and high-volume use cases.
SWE-bench
Benchmark for evaluating AI agents on real-world software engineering tasks drawn from GitHub issues across Django, Matplotlib, SymPy, and other open-source projects. Agents must generate patches that resolve described bugs; performance is measured by resolution rate, API cost, and per-instance detail. Hosts public leaderboards and supports custom evaluation runs. Free and open-access.
Stable Video Diffusion (Stability AI)
Stable Video Diffusion (SVD) is Stability AI's open-weight image-to-video diffusion model, capable of generating short video clips (14–25 frames) from a single still image. The model weights are available on Hugging Face for self-hosting, and Stability AI previously offered an API endpoint (since discontinued as of 2025). Developers can self-host SVD under Stability AI's self-hosted license or access it via third-party inference providers like Replicate and Fal.ai. SVD remains relevant as a lightweight, controllable baseline for building image-animation and product visualization features without cloud API costs.
Point-E (OpenAI)
Point-E is OpenAI's open-source system for generating 3D point clouds from text prompts or images using a two-stage diffusion pipeline: first generating a synthetic 2D view, then converting it into a 3D point cloud. The repository includes notebooks for text-to-point-cloud, image-to-point-cloud, and point-cloud-to-mesh conversion using SDF regression. Available on GitHub and Hugging Face under an open license, Point-E is best suited for developers doing 3D prototyping research or building pipelines that accept coarse 3D estimates as a starting point for downstream refinement.
KubeAI
Open-source AI inference operator for Kubernetes that deploys and scales LLMs, embeddings, reranking models, and speech-to-text services with zero-to-demand autoscaling. Exposes an OpenAI-compatible API, uses prefix-aware load balancing to optimize KV cache hit rates across replicas, and handles model downloading and volume mounting automatically. Works without Istio or Knative dependencies.
Frequently asked questions
What is an MCP server?+
An MCP (Model Context Protocol) server exposes tools, resources, and prompts that AI agents can call at runtime. It is the standard interface for connecting language models to external capabilities — databases, APIs, file systems, and custom logic.
How do I choose an MCP server for production?+
Check the Agentifact trust score. Look specifically at the Security and Interop dimensions. Verify the server handles authentication, has error recovery, and has been tested beyond demo scenarios. Most of the 5,800+ community servers are not production-ready.
Are MCP servers free?+
Many MCP servers are open source and free to use. However, production-grade servers from vendors like Anthropic, Stripe, and Cloudflare may require API keys or paid accounts for the underlying services they connect to.
Can I build my own MCP server?+
Yes. The MCP specification is open and SDKs exist for TypeScript and Python. Building a custom server is the right move when you need to expose internal APIs or proprietary data to agents in a controlled way.