Deploy MCP Server
ML Infrastructure Bearer Token

Anyscale REST API

Scalable ML infrastructure for Ray workloads

Anyscale provides a unified platform for building, deploying, and managing distributed machine learning applications using Ray. The API enables developers to programmatically create compute clusters, deploy ML models, manage workloads, and monitor distributed training jobs. It's designed for teams scaling from prototypes to production ML systems with complex distributed computing requirements.

Base URL https://api.anyscale.com/v2

API Endpoints

MethodEndpointDescription
GET/clustersList all Ray clusters in your organization
POST/clustersCreate a new Ray cluster with specified compute resources
GET/clusters/{cluster_id}Get details about a specific Ray cluster
DELETE/clusters/{cluster_id}Terminate a running Ray cluster
POST/servicesDeploy a Ray Serve application as a managed service
GET/servicesList all deployed Ray Serve services
GET/services/{service_id}Get details and status of a deployed service
PATCH/services/{service_id}Update a deployed service configuration
DELETE/services/{service_id}Delete a deployed Ray Serve service
POST/jobsSubmit a Ray job for execution on a cluster
GET/jobsList all jobs and their execution status
GET/jobs/{job_id}Get detailed information about a specific job
GET/jobs/{job_id}/logsRetrieve logs from a running or completed job
DELETE/jobs/{job_id}Cancel a running job
GET/compute-configsList available compute configurations and instance types

Code Examples

curl -X POST https://api.anyscale.com/v2/clusters \
  -H 'Authorization: Bearer anyscale_your_api_token_here' \
  -H 'Content-Type: application/json' \
  -d '{
    "name": "ml-training-cluster",
    "compute_config": "ml.m5.4xlarge",
    "min_workers": 2,
    "max_workers": 10,
    "ray_version": "2.9.0"
  }'

Connect Anyscale to AI

Deploy a Anyscale MCP server on IOX Cloud and connect it to Claude, ChatGPT, Cursor, or any AI client. Your AI assistant gets direct access to Anyscale through these tools:

create_ray_cluster Create and configure a new Ray cluster with specified compute resources and autoscaling parameters for distributed ML workloads
deploy_ml_service Deploy a trained ML model as a scalable Ray Serve service with automatic load balancing and version management
submit_training_job Submit a distributed training job to an Anyscale cluster with resource specifications and monitoring capabilities
monitor_cluster_metrics Retrieve real-time metrics and resource utilization data from Ray clusters including CPU, GPU, and memory usage
manage_job_lifecycle Start, stop, monitor, and retrieve logs from Ray jobs running on Anyscale infrastructure

Deploy in 60 seconds

Describe what you need, AI generates the code, and IOX deploys it globally.

Deploy Anyscale MCP Server →

Related APIs