Skip to main content

zymtrace Model Context Protocol (MCP) Server

zymtrace exposes GPU and CPU profiling data directly to AI agents through the Model Context Protocol (MCP). This turns your performance profiles into an interactive, AI-driven workspace where you can query flamegraphs, investigate regressions, and get optimization recommendations directly in your IDE or terminal.

MCP is an open standard that enables AI agents to securely connect to external data sources and tools. With zymtrace's MCP server, you can analyze performance using natural language.

Demo

Here's a demo showing how Claude optimized a PyTorch thermal simulation application, making it 7.5x faster:

What You Can Do

  • Query performance data using natural language
  • Analyze flamegraphs for CPU and GPU workloads
  • Identify bottlenecks and get optimization recommendations
  • Track regressions across deployments
  • Investigate issues interactively with AI assistance

Next Steps

🚀 Configure MCP

Configure zymtrace MCP with your AI assistant

Setup Guide →

🔐 Authentication

Learn how to obtain and manage MCP tokens

Token Guide →