Category: AI & ML Tools

Prompt engineering is the skill layer between humans and LLMs — and it’s one of the few areas where MCP servers can make every other MCP tool more effective. A good prompt optimization server doesn’t just rewrite text; it applies research-backed techniques, manages template libraries, and can even route prompts across multiple LLM providers to find the best response.

April 2026 update: The biggest structural change since our initial review is the arrival of prompt security — MCP Guard (53 stars) is the first runtime prompt injection firewall for MCP, filling what was our #1 gap. The observability landscape also expanded significantly: Langfuse upgraded its MCP server from 2 read-only tools to 5 tools with full write capabilities via a native hosted endpoint, and both Helicone and Braintrust launched official MCP servers for prompt observability. The core prompt engineering servers saw modest growth (just-prompt 718→725, claude-prompts 143→147, mcp-prompts 110→113), with the ecosystem expanding from 20+ to 25+ servers.

For AI-generated content workflows, see our CMS & Content Management review. For search optimization of your content, see our SEO review.

Multi-LLM Routing & Unified Prompting (1 server)

ServerStarsLanguageLicenseToolsKey Feature
disler/just-prompt725Python6Unified interface to 6 LLM providers with consensus tool

disler/just-prompt (725 stars, up from 718) is the most-starred server in this space, though it’s more of a multi-LLM router than a prompt optimizer. It provides a unified interface to OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama through 6 tools: prompt (basic), prompt_from_file, prompt_from_file_to_file, ceo_and_board (consensus across models), list_providers, and list_models. The standout feature is the “CEO and board” tool — it queries multiple models in parallel and aggregates their responses for consensus-based decision-making. It supports extended reasoning features across providers (OpenAI o-series reasoning effort, Claude thinking tokens, Gemini thinking budget) and handles automatic model name correction including configurable --default-models parameter. Useful when you want an agent to compare responses across providers or route prompts to the cheapest/fastest model for a given task.

Prompt Workflow & Template Engines (5+ servers)

ServerStarsLanguageLicenseToolsKey Feature
Langfuse native MCPBuilt-in5NEW Hosted at /api/public/mcp, read+write, StreamableHttp
langfuse/mcp-server-langfuse163TypeScriptMIT2Open-source Langfuse prompt retrieval (legacy)
minipuft/claude-prompts147TypeScriptAGPL-3.03Workflow composition, reasoning frameworks, skill export
sparesparrow/mcp-prompts113TypeScriptMIT7Production template management v3.14.0 with AWS backend
mikeskarl/mcp-prompt-templates21Python3+Analysis templates (meetings, summaries, blog conversion)
jezweb/smart-prompts-mcp13TypeScriptMIT7GitHub-based prompt discovery and composition (archived)

Langfuse native MCP is the biggest development in this subcategory. Langfuse now ships a hosted MCP server built directly into the platform at /api/public/mcp using StreamableHttp — no external setup, no build steps, no dependencies. The native server provides 5 tools (up from 2 in the standalone version): getPrompt (fetch by name with optional label/version), listPrompts (filter by name/tag/label with pagination), createTextPrompt (create new text prompt versions with {{variable}} syntax), createChatPrompt (OpenAI-style system/user/assistant roles), and updatePromptLabels (manage production/staging labels across versions). Write capabilities are enabled by default; clients can restrict to read-only via an allowlist. The stateless architecture works across Cloud, self-hosted, or local deployments. This effectively makes Langfuse the first observability platform with full prompt CRUD via native MCP.

The open-source langfuse/mcp-server-langfuse (163 stars, up from 158) still exists as a standalone alternative with 2 tools (get-prompts, get-prompt), but the native hosted version is now the recommended path.

minipuft/claude-prompts (147 stars, up from 143, 380 commits) remains the deepest workflow engine. Tools have been renamed to skills_manager, resource_manager, and prompt_executor (previously prompt_engine, resource_manager, system_control). Hot-reloadable YAML templates back the composition system: operator syntax chains steps (-->), hands off to agents (=>), injects reasoning frameworks (@), and adds validation gates (::). Six built-in reasoning frameworks (CAGEERF, ReACT, 5W1H, and more). A judge mode (%judge) auto-selects optimal resources. Workflows now export as native skills for Claude Code, Cursor, OpenCode, and other platforms — running without MCP calls at runtime. 18 open PRs indicate active development. Licensed under AGPL-3.0 (copyleft).

sparesparrow/mcp-prompts (113 stars, up from 110, now at v3.14.0, 317 commits) continued active development as the most production-ready template manager. 7 tools for prompt CRUD, search, and discovery across 3 storage backends: in-memory, filesystem, and AWS (DynamoDB, S3, SQS). Features include variable substitution with type validation, tag-based filtering, full-text search, role-based access control, rate limiting, and Stripe payment integration.

mikeskarl/mcp-prompt-templates (21 stars) provides analysis-focused templates — meeting minutes generation, executive summaries, and webinar-to-blog-post conversion. Good for content teams with repeatable workflows.

jezweb/smart-prompts-mcp (13 stars, archived December 2025) fetched prompts from GitHub repositories with intelligent discovery, composition, and usage analytics. No longer maintained.

Automated Prompt Optimization (7+ servers)

ServerStarsLanguageLicenseToolsKey Feature
nivlewd1/prompt-optimizer1JavaScript5NEW Cloud Pro + Local Core, Bayesian tuning, 120+ domain rules
Bubobot-Team/mcp-prompt-optimizer23PythonMIT714 research-backed techniques (ToT, APE, Medprompt)
hireshBrem/prompt-engineer-mcp-server13JavaScriptMIT1Claude-powered coding prompt rewriter
MerabyLabs/promptarchitect-mcp5TypeScriptProprietary3NEW Workspace-aware prompt engineering, no API key
andrea9293/mcp-gemini-prompt-enhancer3TypeScriptMIT1Gemini-based enhancement from Google’s prompt engineering guide
sloth-wq/prompt-auto-optimizer-mcp3TypeScriptMIT11Evolutionary optimization via GEPA method
curiositech/prompt-learning-mcp1TypeScriptMIT5APE, OPRO, DSPy patterns with embedding-based learning
Nouman159/prompt-optimizer-mcp0PythonMIT2Deterministic heuristic optimization, sub-100ms

nivlewd1/prompt-optimizer (NEW, 1 star on GitHub but featured on MCP marketplace) is the most commercially ambitious prompt optimization server. It ships as a three-component ecosystem: Cloud Pro (v3.1.1) for LLM-powered optimization with Bayesian tuning and real-time streaming, Local Core (v4.0.2) for offline processing with 120+ domain-specific rules, and a web dashboard for management and analytics. 5 MCP tools: optimize_prompt (professional enhancement), detect_ai_context (automatic intent recognition across code/images/research), search_templates (historical pattern browsing), get_quota_status, and get_ce_quota_status. Supports OpenRouter for model choice. Commercial tiers: Basic free (5 daily optimizations) or Pro ($19.99 one-time for unlimited). Universal MCP compatibility with 17+ clients. The tiered approach — cloud LLM rewriting at 70-95% confidence, rules-based at <25%, local fallback at 35-55% — is pragmatic for teams that want optimization without always paying for API calls.

Bubobot-Team/mcp-prompt-optimizer (23 stars, up from 22) remains the most comprehensive open-source optimizer with 14 optimization techniques — 6 basic (Clarity, Specificity, Chain of Thought, Few-Shot, Structured Output, Role-Based) and 8 advanced (Tree of Thoughts at 70-74% success rate, Constitutional AI, APE, Meta-Prompting, Self-Refine, TEXTGRAD, Medprompt at 90%+ classification accuracy, PromptWizard). 11 professional domain templates and automatic strategy selection. Python 3.8+, 7 tools.

hireshBrem/prompt-engineer-mcp-server (13 stars, unchanged) takes a simpler approach — one tool (rewrite_coding_prompt) that uses Claude 3 Sonnet at temperature 0.2 to rewrite coding prompts. Purpose-built for Cursor and AI IDEs. Appears stagnant (last commit March 2025).

MerabyLabs/promptarchitect-mcp (NEW, 5 stars) is a workspace-aware prompt engineering server. 3 tools: refine_prompt (improves existing prompts with feedback), analyze_prompt (evaluates quality with scoring and suggestions), and generate_prompt (transforms raw ideas into structured prompts). The key differentiator is workspace awareness — it considers your tech stack, project structure, and dependencies when refining prompts, so outputs are tailored to your specific codebase. Works with Claude Desktop, Cursor, VS Code, Windsurf, Zed, JetBrains, Continue.dev, and Cline. No API key required. However, it uses a proprietary license (no copying/modification/distribution), which limits community adoption, and shows only 2 commits total (last activity December 2025).

andrea9293/mcp-gemini-prompt-enhancer (3 stars) uses Google Gemini to optimize prompts based on Lee Boonstra’s 68-page prompt engineering guide. One tool, enhance_prompt.

sloth-wq/prompt-auto-optimizer-mcp (3 stars) uses the GEPA method — genetic algorithms applied to prompts. Population-based variant testing (20 variants default), multi-generation iteration (10 generations default), Pareto frontier analysis. 11 tools including failure analysis and backup/recovery.

curiositech/prompt-learning-mcp (1 star) implements APE, OPRO, and DSPy patterns with vector database storage, RAG-powered retrieval, and cold/warm-start modes. The most research-grounded approach, though adoption remains minimal.

Nouman159/prompt-optimizer-mcp (0 stars) offers deterministic, LLM-free optimization. 3 variants (Creative, Precise, Fast), sub-100ms response times.

Claude Code & IDE-Specific Optimization (2+ servers)

ServerStarsLanguageLicenseToolsKey Feature
gr3enarr0w/cc_peng_mcp13JavaScriptMIT3Claude Code prompt engineering with Q&A refinement
prompt-gen-mcp/prompt-gen-mcp4PythonMIT3Code-context-aware prompt generation

gr3enarr0w/cc_peng_mcp (13 stars, unchanged) is built specifically for Claude Code. Three tools: auto_optimize (detects language and task type automatically), engineer_prompt (structured optimization), and answer_questions (Q&A-based clarification with session management). Recognizes 10+ programming languages and 5 task types. No external API keys required.

prompt-gen-mcp/prompt-gen-mcp (4 stars) scans your local codebase for context, then uses the PromptGen API and GROQ to transform simple questions into comprehensive, context-aware prompts. Code never leaves your machine (only metadata is sent). Integrates with Cursor.

Prompt Security & Injection Protection (2+ servers) — NEW

ServerStarsLanguageLicenseToolsKey Feature
General-Analysis/mcp-guard53TypeScriptMITNEW Runtime prompt injection firewall, AI-powered moderation
ressl/mcp-firewall5PythonAGPL-3.0NEW 12-layer defense pipeline, compliance reporting

This is the most significant gap filled since our initial review. Prompt injection protection went from zero MCP servers to two distinct approaches.

General-Analysis/mcp-guard (53 stars, MIT) is the first runtime prompt injection firewall for MCP. It acts as a proxy that aggregates multiple MCP servers into one secure interface, intercepting tool outputs and evaluating them through General Analysis’s AI-powered moderation API before returning results to the client. Automatic configuration detection for Cursor, Claude Desktop, and Claude Code. Supports both local (stdio) and remote (HTTP/SSE) MCP servers. Transparent proxying of tools, prompts, and resources — your existing MCP setup works unchanged, with security layered on top. Enable with ENABLE_GUARD_API=true and a General Analysis API key. MIT license makes it easy to adopt.

ressl/mcp-firewall (5 stars, AGPL-3.0) takes a more comprehensive policy-as-code approach with a 12-layer defense pipeline: 8 inbound checks (kill switch, agent identity/RBAC, rate limiting, injection detection, egress control, OPA/Rego policy engine, dangerous tool sequence detection, human approval gates) and 4 outbound checks (secret scanning, PII detection, exfiltration detection, custom content policies). Includes cryptographically signed audit trails, real-time dashboard, and compliance reporting for DORA, FINMA, and SOC 2. Very new (single commit, February 2026) but architecturally ambitious. Commercial licensing available alongside AGPL.

Prompt Observability & Analytics (3 platforms) — NEW

PlatformTypeKey Feature
Langfuse native MCPHosted5 tools, read+write, StreamableHttp, no setup
Helicone MCPOfficialquery_requests + query_sessions, request/response debugging
Braintrust MCPOfficial hostedapi.braintrust.dev/mcp, OAuth 2.0, SQL-style log queries

In our initial review, Langfuse was the only observability platform with MCP integration. Now there are three, and the gap we identified is partially filled.

Helicone launched an official MCP server (@helicone/mcp on npm) with 2 tools: query_requests (search with filtering by model, provider, status, latency, cost, properties, time, user) and query_sessions (time range filtering, search, advanced filters). Enables debugging errors, searching logs, and analyzing performance without leaving your AI assistant.

Braintrust launched a hosted MCP server at api.braintrust.dev/mcp (EU: api-eu.braintrust.dev/mcp) with OAuth 2.0 authentication. Tools include documentation search, experiment summaries, BTQL queries, object resolution, schema analysis, and permalink generation. Considered the most mature IDE-native observability MCP by multiple reviewers. Works with Claude Code, Cursor, VS Code, and Windsurf.

Still missing from the MCP prompt observability landscape: PromptLayer, Arize Phoenix, and Weights & Biases.

Structured Prompt Frameworks (1+ servers)

ServerStarsLanguageLicenseToolsKey Feature
doritoman90000/risen-prompts-mcp1JavaScriptMIT8RISEN framework (Role, Instructions, Steps, Expectations, Narrowing)

doritoman90000/risen-prompts-mcp (1 star) implements the RISEN framework — a structured approach that decomposes prompts into 5 components: Role (AI persona), Instructions (task directives), Steps (process breakdown), Expectations (desired outcomes), and Narrowing (constraints). 8 tools cover template creation, validation, storage, and natural-language-to-RISEN conversion. Quality scoring rates templates out of 100 across 5 dimensions (20 points each). Also supports A/B testing and performance tracking. Low adoption but a solid structured approach for teams that want consistent prompt quality.

What’s Missing

The prompt engineering MCP ecosystem has narrowed its gaps since March, but several remain:

  • No prompt security/injection detectionFILLED. MCP Guard (53 stars, MIT) provides runtime prompt injection protection; MCP Firewall adds 12-layer defense with compliance reporting
  • Limited observability integrationPARTIALLY FILLED. Was Langfuse only; now Langfuse (native 5-tool MCP with write), Helicone (official), and Braintrust (official hosted) all have MCP servers. Still no PromptLayer, Arize Phoenix, or Weights & Biases
  • No prompt A/B testing infrastructure — sloth-wq’s GEPA method and Braintrust’s experiment features are the closest, but no dedicated experimentation platform MCP server exists
  • No prompt cost estimation — no server helps estimate token costs before sending prompts to expensive models
  • No prompt chain debugging — no server lets you step through a multi-step prompt chain and inspect intermediate results
  • No prompt versioning with rollback — Langfuse native MCP now has label-based versioning (production/staging), but true rollback workflows are still absent

The Bottom Line

This is a 3.5/5 category — holding steady, though the ecosystem is maturing in important ways. The core prompt engineering servers saw only modest star growth (just-prompt +1%, claude-prompts +3%, mcp-prompts +3%), and most optimization tools remain under 25 stars. But the structural improvements are real: prompt injection protection went from zero to two servers, observability expanded from one platform to three, and Langfuse’s native MCP with write capabilities is a meaningful step toward production prompt management via MCP.

The standout servers serve different needs: just-prompt (725 stars) for multi-LLM routing, Langfuse native MCP (5 tools with read+write) for production prompt management, claude-prompts (147 stars) for workflow composition with reasoning frameworks, mcp-prompts (113 stars, v3.14.0) for enterprise template management, MCP Guard (53 stars) for prompt injection protection, and mcp-prompt-optimizer (23 stars) for research-backed optimization techniques.

For most users, your AI agent’s built-in prompt handling is probably sufficient. These servers become valuable when you’re managing prompt libraries at scale, need to route across LLM providers, want to apply specific optimization techniques systematically, or need to protect against prompt injection attacks in production.

Start here: If you want multi-LLM routing, use just-prompt. If you want production prompt management with versioning, use Langfuse’s native MCP. If you want workflow composition with reasoning frameworks, use claude-prompts. If you need enterprise template management, use mcp-prompts. If you want prompt injection protection, use MCP Guard. If you want research-backed optimization, use mcp-prompt-optimizer.

This review was refreshed on 2026-04-28. Initial review published 2026-03-16. Written using Claude Opus 4.6 (Anthropic).