Available models
The following models are currently available:Model | Description | Endpoint URL |
---|---|---|
Flux Dev | Offers exceptional prompt adherence, high visual fidelity, and rich image detail. | https://api.runpod.ai/v2/black-forest-labs-flux-1-dev/ |
Flux Schnell | Fastest and most lightweight FLUX model, ideal for local development, prototyping, and personal use. | https://api.runpod.ai/v2/black-forest-labs-flux-1-schnell/ |
Deep Cognito v2 70B | An open-source hybrid reasoning LLM developed under a novel AI paradigm. | https://api.runpod.ai/v2/deep-cogito-v2-llama-70b/ |
Qwen3 32B AWQ | The latest LLM in the Qwen series, offering advancements in reasoning, instruction-following, agent capabilities, and multilingual support. | https://api.runpod.ai/v2/qwen3-32b-awq/ |
Public endpoint playground
The public endpoint playground provides a streamlined way to discover and experiment with AI models. The playground offers:- Interactive parameter adjustment: Modify prompts, dimensions, and model settings in real-time.
- Instant preview: Generate images directly in the browser.
- Cost estimation: See estimated costs before running generation.
- API code generation: Create working code examples for your applications.
Access the playground
- Navigate to the Runpod Hub in the console.
- Select the Public endpoints section.
- Browse the available models and select one that fits your needs.
Test a model
To test a model in the playground:- Select a model from the Runpod Hub.
- Under Input, enter a prompt in the text box.
- Enter a negative prompt if needed. Negative prompts tell the model what to exclude from the output.
- Under Additional settings, you can adjust the seed, aspect ratio, number of inference steps, guidance scale, and output format.
- Click Run to start generating.
Create a code example
After inputting parameters using the playground, you can automatically generate an API request to use in your application.- Select the API tab in the UI (above the Input field).
- Using the dropdown menu, select the programming language (Python, JavaScript, cURL, etc.) and POST command you want to use (
/run
or/runsync
). - Click the Copy icon to copy the code to your clipboard.
Make API requests to public endpoints
You can make API requests to public endpoints using any HTTP client. The endpoint URL is specific to the model you want to use. All requests require authentication using your Runpod API key, passed in theAuthorization
header. You can find and create API keys in the Runpod console under Settings > API Keys.
To learn more about the difference between synchronous and asynchronous requests, see Endpoint operations.
Synchronous request example
Here’s an example of a synchronous request to Flux Dev using the/runsync
endpoint:
curl
Asynchronous request example
Here’s an example of an asynchronous request to Flux Dev using the/run
endpoint:
curl
/status
endpoint, replacing {job-id}
with the job ID returned from the /run
request:
curl
Response format
All endpoints return a consistent JSON response format:Model-specific parameters
Each endpoint accepts a different set of parameters to control the generation process.Flux Dev
Flux Dev is optimized for high-quality, detailed image generation. The model accepts several parameters to control the generation process:Parameter | Type | Required | Default | Range | Description |
---|---|---|---|---|---|
prompt | string | Yes | - | - | Text description of the desired image. |
negative_prompt | string | No | - | - | Elements to exclude from the image. |
width | integer | No | 1024 | 256-1536 | Image width in pixels. Must be divisible by 64. |
height | integer | No | 1024 | 256-1536 | Image height in pixels. Must be divisible by 64. |
num_inference_steps | integer | No | 28 | 1-50 | Number of denoising steps. |
guidance | float | No | 7.5 | 0.0-10.0 | How closely to follow the prompt. |
seed | integer | No | -1 | - | Provide a seed for reproducible results. The default value (-1) will generate a random seed. |
image_format | string | No | ”jpeg" | "png” or “jpeg” | Output format. |
Flux Schnell
Flux Schnell is optimized for speed and real-time applications:Parameter | Type | Required | Default | Range | Description |
---|---|---|---|---|---|
prompt | string | Yes | - | - | Text description of the desired image. |
negative_prompt | string | No | - | - | Elements to exclude from the image. |
width | integer | No | 1024 | 256-1536 | Image width in pixels. Must be divisible by 64. |
height | integer | No | 1024 | 256-1536 | Image height in pixels. Must be divisible by 64. |
num_inference_steps | integer | No | 4 | 1-8 | Number of denoising steps. |
guidance | float | No | 7.5 | 0.0-10.0 | How closely to follow the prompt. |
seed | integer | No | -1 | - | Provide a seed for reproducible results. The default value (-1) will generate a random seed. |
image_format | string | No | ”jpeg" | "png” or “jpeg” | Output format. |
Flux Schnell is optimized for speed and works best with lower step counts. Using higher values may not improve quality significantly.
vLLM endpoints
The following vLLM models are available:Model | Endpoint URL | OpenAI API Model Name |
---|---|---|
Deep Cognito v2 70B | https://api.runpod.ai/v2/deep-cogito-v2-llama-70b/ | deepcogito/cogito-v2-preview-llama-70B |
Qwen3 32B AWQ | https://api.runpod.ai/v2/qwen3-32b-awq/ | Qwen/Qwen3-32B-AWQ |
To learn more about sending requests to vLLM public endpoints, see Send vLLM requests.
OpenAI API code example
Use the OpenAI API Model Name in the table above as the model name parameter. Here’s an example of how to use the Qwen3 32B AWQ model with the OpenAI API:To learn more about sending requests to vLLM public endpoints with the OpenAI-compatible API, see OpenAI API compatibility.
OpenAI API streaming example
You can stream responses from the OpenAI API using thestream
and stream_options
parameters:
stream_options={"include_usage": True}
is required for streaming to work with vLLM public endpoints.vLLM Response format
Python example
Here is an example Python API request to Flux Dev using the/run
endpoint:
JavaScript/TypeScript integration with Vercel AI SDK
For JavaScript and TypeScript projects, you can use the@runpod/ai-sdk-provider
package to integrate Runpod’s public endpoints with the Vercel AI SDK.
Run this command to install the package:
Pricing
Public endpoints use transparent, usage-based pricing:Model | Price | Billing unit |
---|---|---|
Flux Dev | $0.02 | Per megapixel |
Flux Schnell | $0.0024 | Per megapixel |
Pricing is calculated based on the actual output resolution. You will not be charged for failed generations.
Pricing examples
Below are some pricing examples that show how you can estimate costs for different image sizes:-
512×512 image (0.25 megapixels)
- Flux Dev: (512 * 512 / 1,000,000) * $0.02 = $0.00524288
- Flux Schnell: (512 * 512 / 1,000,000) * $0.0024 = $0.0006291456
-
1024×1024 image (1 megapixel)
- Flux Dev: (1024 * 1024 / 1,000,000) * $0.02 = $0.02097152
- Flux Schnell: (1024 * 1024 / 1,000,000) * $0.0024 = $0.0025165824
Runpod’s billing system rounds up after the first 10 decimal places.