Run the latest models all in one Sandbox 🏖️

Illusion Diffusion Text to Image

fal-ai/illusion-diffusion
Create illusions conditioned on image.
Inference
Commercial use

Input

Additional Settings

Customize your input with more control.

Result

Idle

What would you like to do next?

Your request will cost $0 per compute second.

Logs

Illusion Diffusion | [text-to-image]

Illusion Diffusion extends Stable Diffusion with ControlNet conditioning to create optical illusion images from pattern inputs and text prompts. Trading straightforward text-to-image generation for pattern-aware composition, this model embeds your prompts into geometric patterns, creating images that reveal different interpretations based on viewing distance or angle. Built for designers and artists who need controllable visual illusions without manual composition work.

Use Cases: Optical Illusion Art | Pattern-Based Design | Hidden Image Marketing | Visual Puzzle Creation


Performance

Illusion Diffusion runs on fal's infrastructure with configurable inference steps (up to 80) and multiple scheduler options for quality-speed tradeoffs.

MetricResultContext
Inference Steps40 (default)Configurable 0-80 range for quality control
Output ResolutionUp to 1024x1024Square HD default, custom multiples of 8 supported
ControlNet Scale0-1 adjustableControls pattern influence strength

Pattern-Conditioned Generation Architecture

Illusion Diffusion extends Stable Diffusion with ControlNet conditioning, allowing you to input both a base pattern image and a text prompt. The model interprets your prompt through the geometric constraints of the pattern, creating outputs where the pattern remains visible while the prompted content emerges at different viewing distances.

What this means for you:

  • Dual-input control: Provide a pattern image URL alongside your text prompt, the model blends both inputs rather than choosing between them

  • Adjustable pattern strength: ControlNet conditioning scale (0-1) lets you dial pattern visibility up or down without regenerating from scratch

  • Guidance timing control: Set when ControlNet influence starts and ends during generation (control_guidance_start/end parameters) for precise pattern integration

  • Multiple scheduler options: Choose between DPM++ Karras SDE and Euler samplers to optimize for your specific pattern complexity and prompt requirements


Technical Specifications

SpecDetails
ArchitectureStable Diffusion with ControlNet
Input FormatsPattern image URL (JPEG, PNG, WebP, GIF, AVIF) + text prompt
Output FormatsPNG images with embedded pattern conditioning
Resolution OptionsSquare HD default (1024x1024), custom dimensions as multiples of 8
LicenseCommercial use permitted

API Documentation | Quickstart Guide | Enterprise Pricing


How It Stacks Up

AuraFlow Text to Image – Illusion Diffusion specializes in pattern-conditioned generation for optical illusion workflows. AuraFlow focuses on straightforward text-to-image generation with high prompt adherence, ideal for standard image generation without pattern constraints.