AI Inferencefaster than you can type
Build real-time AI applications with lightning fast inference (under ~120ms).
No coldstarts. Pay only for what you use.
Model Gallery
Check out some of the ready-to-use models in our model gallery. They have simple API endpoints ready for you to start your own AI-powered applications.
Stable Diffusion XL
Run SDXL at the speed of light
text-to-image
inference
Stable Diffusion with LoRAs
Run Any Stable Diffusion model with customizable LoRA weights.
text-to-image
inference
stylized
Stable Cascade
Stable Cascade: Image generation on a smaller & cheaper latent space.
text-to-image
inference
stylized
Stable Video Diffusion
Generate short video clips from your images using SVD v1.1
image-to-video
inference
ControlNet Scribble
Generate images from scribbled conditioned images.
image-to-image
inference
utility
LLaVA v1.5 13B
Vision
vision
inference
Pricing
Ship custom model endpoints with fine-grained control over idle timeout, max concurrency and autoscaling.
- CPU: 10
- Memory: 64GB
- GPU: A100 (40GB VRAM)
- Total Price: $0.00111/s
Unit Price
CPU
$0.00003/s
Memory
$0.000004/s
GPU A100
$0.001/s
GPU A10G
$0.0002/s
GPU T4
$0.00009/s
Storage
$1/GB/month
Join our community
Join the discussion around our product and help shape the future of AI.