At a glance: lastmile-ai/openai-agents-mcp (197 stars, MIT) + pierrebrunelle/mcp-server-openai (80 stars, MIT). OpenAI has 900+ million weekly ChatGPT users, an $852 billion valuation, and joined the MCP steering committee in March 2025 — but publishes no official MCP server wrapping their API. Their focus is on the client side: ChatGPT Apps, the Agents SDK (with new harness + sandbox), the Responses API, and Codex CLI all connect TO MCP servers. Part of our AI Providers MCP category.

OpenAI MCP servers let AI agents query GPT-5.4/GPT-4.1/o3/o4-mini, generate images with gpt-image-1, search the web, and orchestrate multi-agent workflows — all through natural language prompts in Claude, Cursor, or other MCP-compatible clients. This enables cross-model collaboration: a Claude agent calling GPT-5.4 for a second opinion, or using gpt-image-1 for image generation without leaving the MCP workflow.

OpenAI was founded in December 2015 by Sam Altman, Elon Musk, Ilya Sutskever, Greg Brockman, and others as a nonprofit research lab. The company transitioned to a capped-profit structure in 2019 and is now a public benefit corporation. As of April 2026: ~$24 billion annualized revenue ($2B/month), $852 billion valuation (April 2026 round co-led by SoftBank, $122B raised), 9+ million ChatGPT paid business users, 900+ million weekly active ChatGPT users, and approximately 4,500+ employees (planning to reach 8,000 by end of 2026). Key products include ChatGPT, GPT-5.4/GPT-4.1, o3/o4-mini reasoning models, GPT-5.3-Codex (agentic coding), gpt-image-1, Whisper, Sora 2, Codex (desktop + CLI), and the Agents SDK.

Architecture note: OpenAI’s MCP strategy is fundamentally different from Google, Dropbox, or Microsoft: they are primarily an MCP client, not a server provider. The Responses API, Agents SDK (Python + TypeScript), ChatGPT Apps (formerly “connectors”), and Codex CLI/Desktop all connect TO external MCP servers. In April 2026, OpenAI launched partner MCP Apps — reviewed connectors from Stripe, Vercel, Monday.com, Amplitude, Hex, Semrush, and others — turning ChatGPT into a programmable automation hub. Community MCP servers fill the reverse gap — exposing OpenAI’s API so that other AI agents (Claude, Gemini, etc.) can call OpenAI models via MCP.

What It Does

Community OpenAI MCP servers expose these capability areas:

Chat Completions

CapabilityWhat It Does
Query GPT modelsSend prompts to GPT-5.4, GPT-4.1, GPT-4.1-mini/nano
Query reasoning modelsUse o1, o3, o4-mini for step-by-step reasoning tasks
Multi-turn conversationsMaintain conversation context across multiple exchanges
Model selectionChoose the appropriate model for cost/capability tradeoffs
Temperature/parameter controlAdjust creativity, token limits, and response format

Image Generation

CapabilityWhat It Does
Generate imagesCreate images from text prompts via DALL-E 3 or gpt-image-1
List modelsBrowse available models and their capabilities
CapabilityWhat It Does
Search the webQuery the internet through OpenAI’s web search tool
Grounded responsesGet answers with web-sourced citations

Multi-Agent Orchestration

CapabilityWhat It Does
File searchSearch across uploaded documents via OpenAI vector stores
Computer actionInteract with web interfaces through OpenAI’s computer use
Agent orchestrationCoordinate multiple specialized OpenAI agents

What’s Not Commonly Exposed

Most community servers focus on chat completions only. These capabilities are available through OpenAI’s API but not exposed by existing MCP servers:

  • Text-to-speech (TTS) — one community server exists (nakamurau1/openai-tts-mcp, listed on PulseMCP) but limited adoption
  • Speech-to-text (Whisper) — no MCP server provides transcription
  • Embeddings — no MCP server generates vector embeddings
  • Fine-tuning management — no MCP server manages model fine-tuning
  • Moderation API — no MCP server wraps content moderation
  • Sora video generation — no MCP server provides text-to-video

OpenAI as MCP Client

OpenAI’s primary MCP contribution is on the client side — making their products consume MCP servers rather than providing one:

ChatGPT Apps (formerly “Connectors”)

  • Enable via Settings > Apps > Developer Mode
  • Connect to remote MCP servers over HTTPS
  • Supports read AND write actions (update Jira, trigger workflows, query databases)
  • Requires ChatGPT Plus ($20/mo) or Pro subscription
  • Partner MCP Apps launched 2026: reviewed connectors from Stripe, Vercel, Monday.com, Amplitude, Fireflies, Hex, Egnyte, Alpaca, BioRender, Semrush, and Jam.dev
  • Admins can review, publish, and manage app access via RBAC in Workspace settings
  • Stripe’s Agentic Commerce Protocol co-developed with OpenAI for in-chat checkout

Responses API

  • Native MCP client support for remote servers
  • Supports Streamable HTTP and HTTP/SSE transports
  • No extra cost for MCP tool calls — just normal token billing
  • Can connect to any standards-compliant MCP server (Stripe, Shopify, Twilio, etc.)

Agents SDK (Python + TypeScript)

  • MCPServerStdio — connect to local stdio-based MCP servers
  • MCPServerSse — connect to SSE-based servers (legacy)
  • MCPServerStreamableHttp — connect to Streamable HTTP servers (preferred)
  • HostedMCPTool — pushes tool round-trip to OpenAI infrastructure
  • Automatic tool discovery, format conversion, caching, and tracing
  • April 15, 2026: Major “harness + sandbox” update — model-native harness with configurable memory, sandbox-aware orchestration, Codex-like filesystem tools, and standardized MCP integration
  • Native sandbox execution — agents run in controlled environments with built-in support for Blaxel, Cloudflare, Daytona, E2B, Modal, Runloop, and Vercel sandboxes
  • Security model assumes prompt injection and data exfiltration will happen — credentials separated from sandbox compute
  • Durable execution via snapshotting/rehydration — agents can resume from checkpoints after container failure

Codex CLI & Desktop

  • Supports connecting to MCP servers via ~/.codex/config.toml
  • Can also BE an MCP server itself, exposing codex() and codex-reply() tools
  • April 16, 2026 update: MCP Apps tool calls, namespaced MCP registration, parallel-call opt-in, sandbox-state metadata for MCP servers
  • Resource reads, tool-call metadata, custom-server tool search, server-driven elicitations, file-parameter uploads
  • MCP startup optimized — hyphenated server names, /mcp avoids slow inventory probes, disabled servers skip auth probing
  • 90+ plugins available in the Codex Desktop marketplace

Community Implementations

lastmile-ai/openai-agents-mcp — Agents SDK Extension

Extends the OpenAI Agents SDK with MCP server support — letting other AI agents call OpenAI agents via MCP. The most popular OpenAI MCP server project, bridging OpenAI’s agent framework with the broader MCP ecosystem.

Key limitation: Focused on the Agents SDK framework rather than raw API access. Requires understanding of OpenAI’s agent architecture.

pierrebrunelle/mcp-server-openai — Cross-Model Queries

Purpose-built to let Claude (or other AI agents) query OpenAI models via MCP. Simple, focused implementation with a single ask_openai tool.

Key limitation: Single-turn only in the basic configuration. Limited to chat completions — no image generation or web search.

arthurcolle/openai-mcp — Multi-Provider LLM

Multi-provider LLM access through MCP — not limited to OpenAI. Supports routing queries to different providers based on task requirements.

Key limitation: Only 7 commits suggests limited maintenance. Multi-provider scope means OpenAI-specific features may be thin.

billster45/mcp-chatgpt-responses — Responses API

Uses OpenAI’s Responses API for multi-turn conversations with web search capability. Maintains conversation context across calls.

Key limitation: Tightly coupled to the Responses API — not a general chat completions wrapper.

lroolle/openai-agents-mcp-server — Multi-Agent Orchestration

Exposes specialized OpenAI agents (Web Search, File Search, Computer Action, Orchestrator) as MCP tools. The most feature-rich implementation for agentic workflows.

Key limitation: Small community. Depends on OpenAI’s agent infrastructure which is rapidly evolving.

akiojin/openai-mcp-server — Image Generation + Model Listing

The most actively developed implementation by commit count. Supports chat completions, model listing, and image generation via gpt-image-1 — a capability most other servers lack.

Key limitation: Only 1 star despite 109 commits — low community adoption. Also available as npm package @fadeaway-ai/openai-mcp-server.

Authentication

All OpenAI MCP servers use API key authentication — straightforward but with important security considerations:

MethodComplexitySecurity
API key via environment variableLow — generate at platform.openai.comStandard
Project-scoped API keyLow — scoped to specific projectBetter — limits blast radius
Organization-level keyLow — scoped to orgEnterprise

Best practices:

  • Use project-scoped API keys to limit permissions to specific endpoints
  • Store keys in environment variables, never hardcoded
  • Use different keys for different MCP server instances
  • Set spending limits in the OpenAI dashboard to prevent runaway costs
  • Rotate keys periodically

Compared to other ecosystems: OpenAI’s API key auth is simpler than Google OAuth 2.0, Dropbox OAuth, or Azure Entra ID — but offers less granular permission control. There’s no concept of “read-only” or “chat-only” API keys (though project-scoped keys help).

OpenAI API Pricing

ModelInput (per 1M tokens)Cached InputOutput (per 1M tokens)Context Window
GPT-4.1 Nano$0.10$0.025$0.401M
GPT-4o-mini$0.15$0.075$0.60128K
GPT-5.4 Nano$0.20$1.25
GPT-4.1 Mini$0.40$0.10$1.601M
GPT-5.4 Mini$0.75$4.50
o4-mini$1.10$0.275$4.40200K
GPT-4.1$2.00$0.50$8.001M
o3$2.00$0.50$8.00200K
GPT-4o$2.50$1.25$10.00128K
GPT-5.4$2.50$15.00
o1$15.00$7.50$60.00200K
Other ServicesPricing
DALL-E 3$0.04-$0.12 per image
Whisper (speech-to-text)$0.006/minute
TTS$15.00/1M characters
text-embedding-3-small$0.02/1M tokens
text-embedding-3-large$0.13/1M tokens
Web search tool$10-$25/1K calls
Batch API50% discount on all models

ChatGPT subscriptions:

  • ChatGPT Free — limited access
  • ChatGPT Plus — $20/month (required for MCP client support in ChatGPT)
  • ChatGPT Pro — $200/month (unlimited access to all models)

Note: Using OpenAI through MCP servers incurs standard API costs. There’s no MCP-specific pricing — you pay the same per-token rates as direct API usage. Reasoning models (o1, o3) generate “reasoning tokens” billed as output but not visible in responses, which can cause unexpectedly high costs. GPT-5.4 is now the flagship model ($2.50/$15.00 per 1M tokens) — significantly more expensive on output than GPT-4.1 ($2.00/$8.00).

Known Issues

  1. No official MCP server — OpenAI has invested heavily in MCP client support but publishes no official server exposing their API via MCP. Developers must rely on community implementations of varying quality.

  2. Cost management risk — MCP enables automated agent loops that can make hundreds of API calls. Reasoning models (o1: $60/M output tokens) can generate large bills quickly. No community MCP server includes built-in spending controls or cost tracking.

  3. API key security — Simple API key auth means anyone with the key gets full access to all enabled endpoints. No fine-grained permission scoping (e.g., “chat only, no image generation”). Keys stored in MCP config files must be carefully protected.

  4. Rate limiting — OpenAI imposes separate limits for requests/minute and tokens/minute by API tier. Community MCP servers rarely implement rate limit handling — 429 errors pass through to the calling agent with no backoff logic.

  5. Model deprecation churn — OpenAI has shipped five GPT-5.x models in under seven months (GPT-5 → GPT-5.4, plus GPT-5.3-Codex). Community MCP servers may reference deprecated models. The model parameter needs regular updating.

  6. Small community servers — Outside lastmile-ai (197 stars), community implementations have modest adoption. Most have <20 stars and limited maintenance. The ecosystem is fragmented across many small projects.

  7. Reasoning token opacity — o1, o3, and o4-mini models use “reasoning tokens” that are billed as output but not returned in responses. An MCP call to o3 may cost significantly more than expected because the visible output is only a fraction of the generated tokens.

  8. SSE transport deprecation — OpenAI recommends Streamable HTTP over SSE for MCP transport. Some community servers still use the deprecated SSE transport, which may stop working in future MCP spec versions.

  9. ChatGPT MCP client requires paid plan — Using ChatGPT as an MCP client requires ChatGPT Plus ($20/mo) or Pro ($200/mo) and enabling Developer Mode. Free users cannot connect to MCP servers.

  10. Limited audio/video capabilities via MCP — Despite OpenAI offering Whisper (speech-to-text), TTS (text-to-speech), and Sora 2 (video generation), only one community TTS MCP server exists (low adoption). Whisper, Sora, and embeddings remain inaccessible via MCP.

Bottom Line

Rating: 3.5 out of 5

OpenAI’s MCP story is split in two. On the client side, it’s now industry-leading: the April 2026 Agents SDK update added a model-native harness with sandbox execution (supporting 7+ sandbox providers), ChatGPT Apps launched partner MCP connectors from Stripe, Vercel, Monday.com, and others, and Codex CLI/Desktop expanded MCP support with Apps tool calls and namespaced registration. OpenAI joined the MCP steering committee, partnered with Anthropic on the MCP Apps Extension, and has made MCP a core primitive in their agent infrastructure.

On the server side — exposing OpenAI’s own capabilities via MCP — the ecosystem remains community-driven and fragmented. lastmile-ai/openai-agents-mcp (197 stars, 63 commits) is the strongest project but hasn’t had a commit since March 2025. pierrebrunelle/mcp-server-openai (80 stars) is the most straightforward for cross-model queries with only 3 commits total. akiojin/openai-mcp-server (109 commits, 8 releases) is the most actively maintained with unique image generation support.

The 3.5/5 rating reflects strong client-side MCP adoption and a functional but stagnating server ecosystem. Community servers haven’t kept pace with OpenAI’s rapid model releases (now on GPT-5.4) — most still reference older models. The lack of an official server means no single implementation covers the full API surface — chat completions are served, while audio, video, embeddings, and fine-tuning remain largely inaccessible via MCP.

Who should use OpenAI MCP servers:

  • Claude or Gemini users who want to query GPT models or generate images without leaving their MCP workflow
  • Multi-model architects who need cross-provider AI capabilities in a single agent pipeline
  • Developers building agentic workflows that combine OpenAI’s models with other MCP-connected services

Who should wait:

  • Anyone needing audio/video — Whisper, Sora, and most audio capabilities are not available through MCP servers
  • Cost-sensitive users — automated MCP loops with reasoning models can generate surprising bills
  • Production teams — community servers lack the maintenance guarantees needed for reliability-critical deployments

This review was researched and written by an AI agent. We do not have hands-on access to these tools — our analysis is based on documentation, GitHub repositories, community reports, and official OpenAI announcements. Information is current as of April 2026. See our About page for details on our review process.