Skip to main content
Step-by-step guides for building and deploying example applications on Runpod.

Serverless

Create an image generation endpoint

Deploy a Stable Diffusion endpoint and generate your first AI image.

Integrate with a web application

Deploy an image generation endpoint and integrate it into a web app.

Deploy a cached model

Serve an LLM with reduced cost and cold start times.

Deploy a chatbot with Gemma 3

Use vLLM and the OpenAI API to build an interactive chatbot.

Generate images with ComfyUI

Deploy ComfyUI and generate images using JSON workflows.

Flash

Generate images with Flash

Deploy SDXL as a serverless endpoint with Python decorators.

Text generation with Transformers

Deploy a text generation model on Runpod.

Build a load balancing API

Create a REST API with automatic load balancing.

Pods

Run LLM inference with JupyterLab

Launch JupyterLab on a GPU Pod and run inference with Transformers.

Pods + Ollama

Deploy Ollama on a GPU Pod and run inference using the Ollama API.

Build Docker images with Bazel

Build Docker images on Pods using Bazel.

Generate images with ComfyUI

Deploy ComfyUI on a GPU Pod and use the web interface.

Public Endpoints

Build a text-to-video pipeline

Chain multiple Public Endpoints to generate videos from text.