Skip to main content
Runpod is a cloud computing platform built for AI, machine learning, and general compute needs. Whether you’re or AI models, or deploying cloud-based applications for , Runpod provides scalable, high-performance GPU and CPU resources to power your workloads.

Get started

If you’re new to Runpod, start here to learn the essentials and deploy your first GPU. You can also watch this video for a high-level overview of our products:

Serverless

Serverless provides pay-per-second computing with automatic scaling for production AI/ML apps. You only pay for actual compute time when your code runs, with no idle costs, making Serverless ideal for variable workloads and cost-efficient production deployments.

Pods

Pods give you dedicated GPU or CPU instances for containerized AI/ML workloads. Pods are billed by the minute and stay available as long as you keep them running, making them perfect for development, training, and workloads that need continuous access.

Public Endpoints

Public Endpoints provide instant API access to pre-deployed AI models for image, video, audio, and text generation. No deployment or infrastructure required. You only pay for what you generate, making it easy to integrate AI capabilities into your applications.

Instant Clusters

Instant Clusters deliver fully managed multi-node compute clusters with high-speed networking (up to 3200 Gbps) for distributed workloads. Run multi-node training, fine-tune large language models, or scale inference across multiple GPUs working in parallel.

Support