Forge MCP Server
F

Forge MCP Server

Forge MCP Server is a tool that automatically optimizes PyTorch models into high-performance CUDA/Triton kernels through 32 parallel AI agents, increasing inference speed by up to 14x and supporting multiple MCP-compatible AI programming assistants.
2.5 points
6.1K

What is Forge MCP Server?

Forge MCP Server is a bridge connecting AI programming assistants with GPU optimization services. It allows you to automatically convert PyTorch code into optimized high-performance GPU kernels through AI assistants such as Claude, Cursor, and VS Code. The service uses 32 parallel AI agents to test and optimize code on real data center GPUs, ensuring optimal performance.

How to use Forge MCP Server?

Using Forge is very simple: 1) Install Forge MCP Server in your AI assistant; 2) Log in and authenticate with one click through the browser; 3) Submit PyTorch code or describe the operations you need; 4) Wait for Forge to optimize and return high-performance kernel code. The entire process is fully automated, eliminating the need to manually write CUDA code.

Applicable scenarios

Forge is most suitable for the following scenarios: accelerating custom PyTorch operations, optimizing the performance of existing GPU kernels, generating new high-performance kernels from scratch, converting research code into production-level performance, and achieving optimal performance on various GPU architectures. It is particularly suitable for deep learning researchers, AI engineers, and developers who need high-performance inference.

Main features

Automatic code optimization
Submit PyTorch code to automatically convert it into optimized Triton or CUDA kernels. Compare with the torch.compile(max-autotune) benchmark to ensure performance improvement.
Intelligent kernel generation
Automatically generate high-performance GPU kernel code by describing operation requirements in natural language, supporting multiple data formats and GPU architectures.
Optimization with 32 parallel AI agents
Use 32 Coder+Judge agents to explore different optimization strategies in parallel, including tensor core utilization, memory coalescing, shared memory tiling, and kernel fusion.
Real GPU benchmark testing
All kernels are compiled, tested for correctness, and analyzed for performance on real data center GPUs, supporting various GPUs such as B200, H200, H100, and A100.
Intelligent detection and suggestions
Automatically identify optimization opportunities in the code, such as custom operations, attention mechanisms, and fused operations, and provide optimization suggestions.
One-click browser authentication
There is no need to manage API keys. Log in securely in the browser through the OAuth 2.0 PKCE process, and the token is automatically refreshed.
Advantages
Up to 14x performance improvement, surpassing torch.compile(max-autotune)
Supports multiple mainstream AI programming assistants (Claude, Cursor, VS Code, etc.)
Tested on real data center GPUs, with reliable results
100% numerical correctness guarantee
Fast result return (in minutes rather than hours)
Obtain high-performance kernels without CUDA programming experience
Pay-as-you-go, with transparent costs
Free trial for one kernel optimization
Limitations
Requires an internet connection to access the optimization service
Each optimization consumes 1 credit (only charged for successful optimizations)
The maximum code input limit is 500KB
Only supports PyTorch code optimization
Requires modern GPU support (T4 and above)

How to use

Install Forge MCP Server
Add Forge MCP Server to the configuration file according to the AI assistant you are using. For example, edit the claude_desktop_config.json file in Claude Desktop.
Authentication and login
Call the forge_auth tool in the AI assistant, and the system will automatically open the browser to complete the login. You only need to log in once, and the token will be automatically refreshed.
Submit code for optimization
When the AI assistant detects PyTorch code that can be optimized, it will automatically call the forge_optimize tool. You can also manually describe the operations to be optimized.
Get optimization results
Wait for the optimization to complete (usually a few minutes), and the system will return the optimized kernel code, performance improvement data, and integration suggestions.
Integrate the optimized code
Integrate the returned optimized kernel code into your project, replacing the original PyTorch implementation, and enjoy the performance improvement.

Usage examples

Optimize custom attention mechanism
Researchers have developed a new attention variant, but the PyTorch implementation is slow. Use Forge to automatically optimize it into a high-performance Triton kernel.
Generate a fused normalization layer
It is necessary to fuse LayerNorm, GELU activation, and Dropout into a single efficient kernel to reduce memory access.
Accelerate existing model inference
The inference speed of the model in the production environment does not meet the requirements, and it is necessary to optimize the key computational bottlenecks.

Frequently Asked Questions

How long does Forge optimization take?
Will I be charged if the optimization fails?
Which GPUs are supported?
How to ensure the correctness of the optimized code?
How large a code can be optimized?
How to purchase credits?
Is local deployment supported?
Are there any licensing restrictions on the optimized code?

Related resources

Official documentation
Detailed technical documentation and benchmark test results of the Forge service
GitHub repository
Source code and issue tracking for Forge MCP Server
MCP protocol official website
Official documentation and specifications of the Model Context Protocol
RightNow AI console
Purchase credits, view usage statistics, and manage accounts
NVIDIA Inception program
RightNow AI is a member of the NVIDIA Inception program

Installation

Copy the following command to your Client for configuration
{
  "mcpServers": {
    "forge": {
      "command": "npx",
      "args": ["-y", "@rightnow/forge-mcp-server"]
    }
  }
}

{
  "mcpServers": {
    "forge": {
      "command": "cmd",
      "args": ["/c", "npx", "-y", "@rightnow/forge-mcp-server"]
    }
  }
}
Note: Your key is sensitive information, do not share it with anyone.

Alternatives

V
Vestige
Vestige is an AI memory engine based on cognitive science. By implementing 29 neuroscience modules such as prediction error gating, FSRS - 6 spaced repetition, and memory dreaming, it provides long - term memory capabilities for AI. It includes a 3D visualization dashboard and 21 MCP tools, runs completely locally, and does not require the cloud.
Rust
4.9K
4.5 points
M
Moltbrain
MoltBrain is a long-term memory layer plugin designed for OpenClaw, MoltBook, and Claude Code, capable of automatically learning and recalling project context, providing intelligent search, observation recording, analysis statistics, and persistent storage functions.
TypeScript
4.4K
4.5 points
B
Bm.md
A feature-rich Markdown typesetting tool that supports multiple style themes and platform adaptation, providing real-time editing preview, image export, and API integration capabilities
TypeScript
4.9K
5 points
S
Security Detections MCP
Security Detections MCP is a server based on the Model Context Protocol that allows LLMs to query a unified security detection rule database covering Sigma, Splunk ESCU, Elastic, and KQL formats. The latest version 3.0 is upgraded to an autonomous detection engineering platform that can automatically extract TTPs from threat intelligence, analyze coverage gaps, generate SIEM-native format detection rules, run tests, and verify. The project includes over 71 tools, 11 pre-built workflow prompts, and a knowledge graph system, supporting multiple SIEM platforms.
TypeScript
6.3K
4 points
P
Paperbanana
Python
6.4K
5 points
B
Better Icons
An MCP server and CLI tool that provides search and retrieval of over 200,000 icons, supports more than 150 icon libraries, and helps AI assistants and developers quickly obtain and use icons.
TypeScript
6.1K
4.5 points
A
Assistant Ui
assistant - ui is an open - source TypeScript/React library for quickly building production - grade AI chat interfaces, providing composable UI components, streaming responses, accessibility, etc., and supporting multiple AI backends and models.
TypeScript
7.6K
5 points
A
Apify MCP Server
The Apify MCP Server is a tool based on the Model Context Protocol (MCP) that allows AI assistants to extract data from websites such as social media, search engines, and e-commerce through thousands of ready-to-use crawlers, scrapers, and automation tools (Apify Actors). It supports OAuth and Skyfire proxy payment and can be integrated into MCP clients such as Claude and VS Code through HTTPS endpoints or local stdio.
TypeScript
7.4K
5 points
N
Notion Api MCP
Certified
A Python-based MCP Server that provides advanced to-do list management and content organization functions through the Notion API, enabling seamless integration between AI models and Notion.
Python
21.5K
4.5 points
G
Gitlab MCP Server
Certified
The GitLab MCP server is a project based on the Model Context Protocol that provides a comprehensive toolset for interacting with GitLab accounts, including code review, merge request management, CI/CD configuration, and other functions.
TypeScript
24.7K
4.3 points
D
Duckduckgo MCP Server
Certified
The DuckDuckGo Search MCP Server provides web search and content scraping services for LLMs such as Claude.
Python
73.3K
4.3 points
M
Markdownify MCP
Markdownify is a multi-functional file conversion service that supports converting multiple formats such as PDFs, images, audio, and web page content into Markdown format.
TypeScript
34.6K
5 points
F
Figma Context MCP
Framelink Figma MCP Server is a server that provides access to Figma design data for AI programming tools (such as Cursor). By simplifying the Figma API response, it helps AI more accurately achieve one - click conversion from design to code.
TypeScript
63.7K
4.5 points
U
Unity
Certified
UnityMCP is a Unity editor plugin that implements the Model Context Protocol (MCP), providing seamless integration between Unity and AI assistants, including real - time state monitoring, remote command execution, and log functions.
C#
32.5K
5 points
G
Gmail MCP Server
A Gmail automatic authentication MCP server designed for Claude Desktop, supporting Gmail management through natural language interaction, including complete functions such as sending emails, label management, and batch operations.
TypeScript
21.1K
4.5 points
C
Context7
Context7 MCP is a service that provides real-time, version-specific documentation and code examples for AI programming assistants. It is directly integrated into prompts through the Model Context Protocol to solve the problem of LLMs using outdated information.
TypeScript
97.5K
4.7 points
AIBase
Zhiqi Future, Your AI Solution Think Tank
© 2026AIBase