Nano Banana 2 is here 🍌 4x faster, lower cost, better quality

fal-ai/qwen-image-2/text-to-image

Qwen-Image-2.0 is a next-generation foundational unified generation-and-editing model
Inference
Commercial use
Partner

Input

Additional Settings

Customize your input with more control.

Result

Idle

What would you like to do next?

Your request will cost $0.035 per image.

Logs

Qwen Image 2.0 — Text to Image

Qwen Image 2.0 is Alibaba's unified image generation model, running a 7B-parameter architecture that delivers native 2K resolution and built-in typography rendering. This standard endpoint is optimized for speed, making it ideal for rapid prototyping, prompt exploration, and iterative workflows at $0.035 per image.

Standard vs Pro

The standard tier prioritizes generation speed for fast iteration. When you're ready for final production assets with maximum detail and text accuracy, switch to the Pro endpoint ($0.075/image).

Key Parameters

ParameterDefaultRangeNotes
promptup to 1,000 tokensDescribe subject, style, and composition
negative_promptstringExclude unwanted elements
image_sizesquareenum or custom`square`, `square_hd`, `landscape_4_3`, `landscape_16_9`, `portrait_4_3`, `portrait_16_9`
guidance_scale50–20Lower = creative; higher = strict adherence
num_inference_steps281–5015–20 for drafts, 25–30 for quality
seedrandomintegerFix for reproducible outputs
num_images11–4Batch multiple images in one request
output_formatpngpng / jpeg / webpChoose based on file size needs

Quick Start

python
import fal_client

result = fal_client.subscribe(
    "fal-ai/qwen-image-2/text-to-image",
    arguments={
        "prompt": "Watercolor illustration of a cozy Japanese ramen shop at night, warm lantern glow, rain-slick street reflections, Studio Ghibli atmosphere",
        "image_size": "landscape_4_3",
        "num_inference_steps": 25,
        "guidance_scale": 5,
    }
)

print(result["images"][0]["url"])
javascript
import { fal } from "@fal-ai/client";

const result = await fal.subscribe("fal-ai/qwen-image-2/text-to-image", {
  input: {
    prompt: "Watercolor illustration of a cozy Japanese ramen shop at night, warm lantern glow, rain-slick street reflections, Studio Ghibli atmosphere",
    image_size: "landscape_4_3",
    num_inference_steps: 25,
    guidance_scale: 5,
  },
});

console.log(result.data.images[0].url);

Iteration Workflow

A practical approach for finding the right output:

  1. Explore — generate at 15–20 steps to test prompt ideas quickly
  2. Refine — bump to 25–28 steps once you have a direction
  3. Finalize — switch to the Pro endpoint at 35+ steps for production renders

Use a fixed `seed` while adjusting prompt wording to isolate what each change does.

Prompting Tips

  • Front-load subjects: "Red sports car on mountain road, aerial drone shot, golden hour" beats "A nice landscape photo with a car somewhere."
  • Be specific about style: "editorial photography", "watercolor illustration", "low-poly 3D render" — avoid vague terms like "beautiful" or "amazing."
  • One style per prompt: "Photorealistic oil painting" confuses the model. Pick one direction.
  • Negative prompts help: `"blurry, distorted, deformed, watermark, text artifacts"` cleans up common issues.

Best Use Cases

  • Concept exploration — quickly test ideas before committing to Pro-quality renders
  • Social media content — fast turnaround for posts and stories
  • Mood boards and references — generate visual directions for creative projects
  • Batch generation — produce multiple variations cheaply to find the best composition

Related Endpoints