- Edit
- Text To Image
Endpoint:
POST https://fal.run/fal-ai/bytedance/seedream/v4.5/edit
Endpoint ID: fal-ai/bytedance/seedream/v4.5/editTry it in the Playground
Run this model interactively with your own prompts.
Quick Start
Input Schema
The text prompt used to edit the image
The size of the generated image. Width and height must be between 1920 and 4096, or total number of pixels must be between 25601440 and 40964096.Possible values:
square_hd, square, portrait_4_3, portrait_16_9, landscape_4_3, landscape_16_9, auto_2K, auto_4KNumber of separate model generations to be run with the prompt. Default value:
1Range: 1 to 6If set to a number greater than one, enables multi-image generation. The model will potentially return up to
max_images images every generation, and in total, num_images generations will be carried out. In total, the number of images generated will be between num_images and max_images*num_images. The total number of images (image inputs + image outputs) must not exceed 15 Default value: 1Range: 1 to 6Random seed to control the stochasticity of image generation.
If
True, the media will be returned as a data URI and the output data won’t be available in the request history.If set to true, the safety checker will be enabled. Default value:
trueList of URLs of input images for editing. Presently, up to 10 image inputs are allowed. If over 10 images are sent, only the last 10 will be used.
Output Schema
Generated images
Input Example
Output Example
Natural Language Editing Without Layers
Seedream 4.5 consolidates image generation and editing into a single architecture that interprets spatial references directly from your prompt. Instead of requiring layer masks or selection tools, you describe edits using natural language - “replace the product in Figure 1 with that in Figure 2” or “copy the text from Figure 3 to the top with clear contrast.” What this means for you:- Multi-source composition: Reference up to 10 images per edit, enabling complex workflows like product swaps, text overlay copying, and element positioning across multiple source files
- Context-aware transformations: The model maintains depth, perspective, and lighting consistency when integrating elements from different sources - no manual blending required
- Resolution flexibility: Output up to 4 megapixels (2048x2048 maximum) with configurable dimensions between 1920px and 4096px on either axis
- Batch generation control: Run 1-6 separate generations per request, with optional multi-image output (up to 6 images per generation) for exploring variations
Performance That Scales
Seedream 4.5 processes edits in approximately 60 seconds on fal infrastructure, with pricing structured for production workflows requiring multiple reference images.| Metric | Result | Context |
|---|---|---|
| Inference Speed | ~60 seconds | Standard processing time per edit on fal |
| Cost per Edit | $0.04 | 25 edits per $1.00 on fal |
| Max Reference Images | 10 images | Multi-source composition capability (last 10 used if more provided) |
| Max Resolution | 4MP (2048x2048) | Configurable dimensions between 1920-4096px per axis |
Technical Specifications
| Spec | Details |
|---|---|
| Architecture | Seedream 4.5 |
| Input Formats | Image URLs (up to 10), text prompt |
| Output Formats | PNG images via URL or data URI |
| Resolution Range | 1920-4096px per axis, 2560×1440 to 4096×4096 total pixels |
| License | Commercial use via fal Partner agreement |
How It Stacks Up
Bytedance Seedream v4 Edit - Seedream 4.5 expands multi-image input capacity from v4’s baseline while maintaining the unified editing architecture. Both versions handle natural language spatial instructions, with v4.5 prioritizing higher reference image limits for complex composition workflows. Bytedance Seededit v3 - Seedream 4.5 consolidates generation and editing into a single model architecture, trading v3’s specialized editing focus for broader capability coverage. Seededit v3 remains purpose-built for pure image-to-image transformation workflows without generation requirements. NAFNet-deblur - Seedream 4.5 handles multi-image composition and semantic editing through natural language, making it ideal for layout assembly and element integration. NAFNet-deblur specializes in single-image restoration tasks like blur removal and artifact correction where semantic understanding isn’t required.Related
- Bytedance — Image Generation
Limitations
image_sizerestricted to:square_hd,square,portrait_4_3,portrait_16_9,landscape_4_3,landscape_16_9,auto_2K,auto_4Knum_imagesrange: 1 to 6max_imagesrange: 1 to 6- Content moderation via safety checker