Server data from the Official MCP Registry
Complete GPU infrastructure for Claude Code — 192 MCP tools for provisioning, training, inference
Complete GPU infrastructure for Claude Code — 192 MCP tools for provisioning, training, inference
Valid MCP server (3 strong, 1 medium validity signals). 8 known CVEs in dependencies (0 critical, 3 high severity) Package registry verified. Imported from the Official MCP Registry.
5 files analyzed · 9 issues found
Security scores are indicators to help you make informed decisions, not guarantees. Always review permissions before connecting any MCP server.
This plugin requests these system permissions. Most are normal for its category.
Set these up before or after installing:
Environment variable: TERRADEV_API_KEY
Environment variable: TERRADEV_CREDENTIALS_FILE
Environment variable: TERRADEV_PROVIDER
Add this to your MCP configuration file:
{
"mcpServers": {
"io-github-theoddden-terradev": {
"env": {
"TERRADEV_API_KEY": "your-terradev-api-key-here",
"TERRADEV_PROVIDER": "your-terradev-provider-here",
"TERRADEV_CREDENTIALS_FILE": "your-terradev-credentials-file-here"
},
"args": [
"-y",
"terradev-mcp"
],
"command": "npx"
}
}
}From the project's GitHub README.
Complete Agentic GPU Infrastructure for Claude Code — 192 MCP tools: GPU provisioning, vLLM/SGLang/Ollama inference, Arize Phoenix observability, NeMo Guardrails safety, Qdrant vector DB, Ray cluster management, Datadog monitoring, and Terraform-powered parallel provisioning across 20 cloud providers.
Terraform is the fundamental engine - not just a feature. This provides:
pip install terradev-cli
# For all providers + HF Spaces:
pip install "terradev-cli[all]"
export TERRADEV_RUNPOD_KEY=your_runpod_api_key
npm install -g terradev-mcp
Add to your Claude Code MCP configuration:
{
"mcpServers": {
"terradev": {
"command": "terradev-mcp"
}
}
}
Use Terradev from Claude.ai on any device — no local install required.
https://terradev-mcp.terradev.cloud/sse
That's it — GPU provisioning tools are now available in every Claude.ai conversation.
To host your own instance:
# Set required env vars
export TERRADEV_MCP_BEARER_TOKEN=your-secret-token
export TERRADEV_RUNPOD_KEY=your-runpod-key
# Option 1: Run directly
pip install -r requirements.txt
python3 terradev_mcp.py --transport sse --port 8080
# Option 2: Docker
docker-compose up -d
The server exposes:
GET /sse — SSE stream endpoint (Claude.ai connects here)POST /messages — MCP message endpointGET /health — Health check (unauthenticated)See nginx-mcp.conf for reverse proxy configuration with SSL.
The Terradev MCP server provides 192 tools for complete GPU cloud management:
local_scan - Discover local GPU devices and total VRAM pool (NEW in v1.2.2)quote_gpu - Get real-time GPU prices across all cloud providersprovision_gpu - Terraform-powered GPU provisioning with parallel efficiencyterraform_plan - Generate Terraform execution plansterraform_apply - Apply Terraform configurationsterraform_destroy - Destroy Terraform-managed infrastructurek8s_create - Create Kubernetes clusters with GPU nodesk8s_list - List all Kubernetes clustersk8s_info - Get detailed cluster informationk8s_destroy - Destroy Kubernetes clustersinferx_deploy - Deploy models to InferX serverless platforminferx_status - Check inference endpoint statusinferx_list - List deployed inference modelsinferx_optimize - Get cost analysis for inference endpointshf_space_deploy - Deploy models to HuggingFace Spacesdeploy_wide_ep - Deploy MoE model with Wide-EP across multiple GPUs via Ray Serve LLMdeploy_pd - Deploy disaggregated Prefill/Decode serving with NIXL KV transferep_group_status - Health check EP groups (all ranks must be healthy for all-to-all)sglang_start - Start SGLang server with EP/EPLB/DBO flags via SSH/systemdsglang_stop - Stop SGLang server on remote instancestatus - View all instances and costsmanage_instance - Stop/start/terminate GPU instancesanalytics - Get cost analytics and spending trendsoptimize - Find cheaper alternatives for running instancessetup_provider - Get setup instructions for any cloud providerconfigure_provider - Configure provider credentials locallyphoenix_test - Test connection to Phoenix serverphoenix_projects - List Phoenix projectsphoenix_spans - List spans with SpanQuery DSL filtersphoenix_trace - View full execution tree for a tracephoenix_otel_env - Generate OTEL env vars for serving podsphoenix_snippet - Generate Python instrumentation snippetphoenix_k8s - Generate K8s deployment manifestguardrails_test - Test connection to Guardrails serverguardrails_chat - Send message through safety railsguardrails_generate_config - Generate Colang 2.x configguardrails_k8s - Generate K8s deployment manifestqdrant_test - Test connection to Qdrantqdrant_collections - List vector collectionsqdrant_create_collection - Create collection (auto-configures from embedding model)qdrant_info - Get collection statsqdrant_count - Count vectors in collectionqdrant_k8s - Generate K8s StatefulSet manifest# Scan for local GPUs
terradev local scan
# Example output:
# ✅ Found 2 local GPU(s)
# 📊 Total VRAM Pool: 48 GB
#
# Devices:
# • NVIDIA GeForce RTX 4090
# - Type: CUDA
# - VRAM: 24 GB
# - Compute: 8.9
#
# • Apple Metal
# - Type: MPS
# - VRAM: 24 GB
# - Platform: arm64
Hybrid Use Case: Mac Mini (24GB) + Gaming PC with RTX 4090 (24GB) = 48GB local pool for Qwen2.5-72B!
# Get prices for specific GPU type
terradev quote -g H100
# Filter by specific providers
terradev quote -g A100 -p runpod,vastai,lambda
# Quick-provision cheapest option
terradev quote -g H100 --quick
# Provision single GPU via Terraform
terradev provision -g A100
# Provision multiple GPUs in parallel across providers
terradev provision -g H100 -n 4 --providers ["runpod", "vastai", "lambda", "aws"]
# Plan without applying
terradev provision -g A100 -n 2 --plan-only
# Set maximum price ceiling
terradev provision -g A100 --max-price 2.50
# Terraform state is automatically managed
# Generate execution plan
terraform plan -config-dir ./my-gpu-infrastructure
# Apply infrastructure
terraform apply -config-dir ./my-gpu-infrastructure -auto-approve
# Destroy infrastructure
terraform destroy -config-dir ./my-gpu-infrastructure -auto-approve
# Create multi-cloud K8s cluster
terradev k8s create my-cluster --gpu H100 --count 4 --multi-cloud --prefer-spot
# List all clusters
terradev k8s list
# Get cluster details
terradev k8s info my-cluster
# Destroy cluster
terradev k8s destroy my-cluster
# Deploy model to InferX
terradev inferx deploy --model meta-llama/Llama-2-7b-hf --gpu-type a10g
# Check endpoint status
terradev inferx status
# List deployed models
terradev inferx list
# Get cost analysis
terradev inferx optimize
# Deploy LLM template
terradev hf-space my-llama --model-id meta-llama/Llama-2-7b-hf --template llm
# Deploy with custom hardware
terradev hf-space my-model --model-id microsoft/DialoGPT-medium --hardware a10g-large --sdk gradio
# Deploy embedding model
terradev hf-space my-embeddings --model-id sentence-transformers/all-MiniLM-L6-v2 --template embedding
# View all running instances and costs
terradev status --live
# Stop instance
terradev manage -i <instance-id> -a stop
# Start instance
terradev manage -i <instance-id> -a start
# Terminate instance
terradev manage -i <instance-id> -a terminate
# Get 30-day cost analytics
terradev analytics --days 30
# Find cheaper alternatives
terradev optimize
# Get quick setup instructions
terradev setup runpod --quick
terradev setup aws --quick
terradev setup vastai --quick
# Configure credentials (stored locally)
terradev configure --provider runpod
terradev configure --provider aws
terradev configure --provider vastai
This release includes fixes for all known production issues:
| Bug | Fix | Impact |
|---|---|---|
| Wrong import path (terradev_cli.providers) | Changed to providers.provider_factory | ✅ API calls now work |
| list builtin shadowed by Click command | Used type([]) instead of isinstance(r, list) | ✅ No more crashes |
| aiohttp.ClientSession(trust_env=False) | Set trust_env=True for proxy support | ✅ Proxy environments work |
| boto3 not in dependencies | Added boto3>=1.26.0 to requirements | ✅ AWS provider functional |
| Vast.ai GPU name filter exact match | Switched to client-side filtering with "in" | ✅ Vast.ai provider works |
All bugs are now resolved in v1.2.0
The MCP now includes a terraform.tf template for custom infrastructure:
terraform {
required_providers {
terradev = {
source = "theoddden/terradev"
version = "~> 3.0"
}
}
}
resource "terradev_instance" "gpu" {
gpu_type = var.gpu_type
spot = true
count = var.gpu_count
tags = {
Name = "terradev-mcp-gpu"
Provisioned = "terraform"
GPU_Type = var.gpu_type
}
}
Terradev v1.5 integrates the full MoE serving stack:
| Component | What it does | Terradev integration |
|---|---|---|
| Ray Serve LLM | Orchestrates Wide-EP and P/D deployments | build_dp_deployment, build_pd_openai_app |
| Expert Parallelism | Distributes experts across GPUs | EP/DP flags in task.yaml, K8s, Helm, Terraform |
| EPLB | Rebalances experts at runtime | --enable-eplb in vLLM/SGLang serving |
| Dual-Batch Overlap | Overlaps compute with all-to-all | --enable-dbo flag |
| DeepEP kernels | Optimized all-to-all for MoE | VLLM_ALL2ALL_BACKEND=deepep_low_latency |
| DeepGEMM | FP8 GEMM for MoE experts | VLLM_USE_DEEP_GEMM=1 |
| NIXL | Zero-copy KV cache transfer | NixlConnector in P/D tracker |
| EP Group Router | Routes to rank hosting target experts | Expert range tracking per endpoint |
RunPod, Vast.ai, AWS, GCP, Azure, Lambda Labs, CoreWeave, TensorDock, Oracle Cloud, Crusoe Cloud, DigitalOcean, HyperStack, Alibaba Cloud, OVHcloud, FluidStack, Hetzner, SiliconFlow, Baseten, HuggingFace, Paperspace
Minimum setup:
TERRADEV_RUNPOD_KEY: RunPod API keyRemote SSE mode:
TERRADEV_MCP_BEARER_TOKEN: Bearer token for authenticating Claude.ai Connector requests (required in production)Full multi-cloud setup:
TERRADEV_AWS_ACCESS_KEY_ID, TERRADEV_AWS_SECRET_ACCESS_KEY, TERRADEV_AWS_DEFAULT_REGIONTERRADEV_GCP_PROJECT_ID, TERRADEV_GCP_CREDENTIALS_PATHTERRADEV_AZURE_SUBSCRIPTION_ID, TERRADEV_AZURE_CLIENT_ID, TERRADEV_AZURE_CLIENT_SECRET, TERRADEV_AZURE_TENANT_IDHF_TOKEN: For HuggingFace Spaces deployment| Tier | Price | Instances | Seats |
|---|---|---|---|
| Research (Free) | $0 | 1 | 1 |
| Research+ | $49.99/mo | 8 | 1 |
| Enterprise | $299.99/mo | 32 | 5 |
| Enterprise+ | $0.09/GPU-hr (32 GPU min) | Unlimited | Unlimited |
Enterprise+: Metered billing at $0.09 per GPU-hour with a minimum of 32 GPUs. Unlimited provisions, servers, seats, dedicated support, fleet management, and GPU-hour metering. Run
terradev upgrade -t enterprise_plus.
BYOAPI: All API keys stay on your machine. Terradev never proxies credentials through third parties.
Be the first to review this server!
by Modelcontextprotocol · Developer Tools
Read, search, and manipulate Git repositories programmatically
by Modelcontextprotocol · Developer Tools
Web content fetching and conversion for efficient LLM usage
by Toleno · Developer Tools
Toleno Network MCP Server — Manage your Toleno mining account with Claude AI using natural language.
by mcp-marketplace · Developer Tools
Create, build, and publish Python MCP servers to PyPI — conversationally.
by Microsoft · Content & Media
Convert files (PDF, Word, Excel, images, audio) to Markdown for LLM consumption
by mcp-marketplace · Developer Tools
Scaffold, build, and publish TypeScript MCP servers to npm — conversationally