Serverless
Create an image generation endpoint
Deploy a Stable Diffusion endpoint and generate your first AI image.
Integrate with a web application
Deploy an image generation endpoint and integrate it into a web app.
Deploy a cached model
Serve an LLM with reduced cost and cold start times.
Deploy a chatbot with Gemma 3
Use vLLM and the OpenAI API to build an interactive chatbot.
Generate images with ComfyUI
Deploy ComfyUI and generate images using JSON workflows.
Flash
Generate images with Flash
Deploy SDXL as a serverless endpoint with Python decorators.
Text generation with Transformers
Deploy a text generation model on Runpod.
Build a load balancing API
Create a REST API with automatic load balancing.
Pods
Run LLM inference with JupyterLab
Launch JupyterLab on a GPU Pod and run inference with Transformers.
Pods + Ollama
Deploy Ollama on a GPU Pod and run inference using the Ollama API.
Build Docker images with Bazel
Build Docker images on Pods using Bazel.
Generate images with ComfyUI
Deploy ComfyUI on a GPU Pod and use the web interface.
Public Endpoints
Build a text-to-video pipeline
Chain multiple Public Endpoints to generate videos from text.