# LTX-2.3 22B Distilled

> Generate video with audio from audio, text and images using LTX-2 Distilled


## Overview

- **Endpoint**: `https://fal.run/fal-ai/ltx-2.3-22b/distilled/audio-to-video`
- **Model ID**: `fal-ai/ltx-2.3-22b/distilled/audio-to-video`
- **Category**: audio-to-video
- **Kind**: inference


## Pricing

Your request will cost $0.001205 per megapixel of generated video data (width × height × frames), rounded up. For example, if you generate a video that is 121 frames long at 1280 × 720, your total generated video is ≈112 MP, and your request will cost $0.1344.

For more details, see [fal.ai pricing](https://fal.ai/pricing).

## API Information

This model can be used via our HTTP API or more conveniently via our client libraries.
See the input and output schema below, as well as the usage examples.


### Input Schema

The API accepts the following input parameters:


- **`prompt`** (`string`, _required_):
  The prompt to generate the video from.
  - Examples: "A woman speaks to the camera"

- **`audio_url`** (`string`, _required_):
  The URL of the audio to generate the video from.
  - Examples: "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-audio.mp3"

- **`image_url`** (`string`, _optional_):
  Optional URL of an image to use as the first frame of the video.
  - Examples: "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-image.png"

- **`end_image_url`** (`string`, _optional_):
  The URL of the image to use as the end of the video.

- **`match_audio_length`** (`boolean`, _optional_):
  When enabled, the number of frames will be calculated based on the audio duration and FPS. When disabled, use the specified num_frames. Default value: `true`
  - Default: `true`

- **`num_frames`** (`integer`, _optional_):
  The number of frames to generate. Default value: `121`
  - Default: `121`
  - Range: `9` to `481`

- **`video_size`** (`ImageSize | Enum`, _optional_):
  The size of the generated video. Use 'auto' to match the input image dimensions if provided. Default value: `landscape_16_9`
  - Default: `"landscape_16_9"`
  - One of: ImageSize | Enum

- **`use_multiscale`** (`boolean`, _optional_):
  Whether to use multi-scale generation. If True, the model will generate the video at a smaller scale first, then use the smaller video to guide the generation of a video at or above your requested size. This results in better coherence and details. Default value: `true`
  - Default: `true`

- **`fps`** (`float`, _optional_):
  The frames per second of the generated video. Default value: `24`
  - Default: `24`
  - Range: `1` to `60`

- **`scheduler`** (`SchedulerEnum`, _optional_):
  The scheduler to use. Default value: `"beta"`
  - Default: `"beta"`
  - Options: `"ltx2"`, `"linear_quadratic"`, `"beta"`
  - Examples: "beta"

- **`acceleration`** (`AccelerationEnum`, _optional_):
  The acceleration level to use. Default value: `"none"`
  - Default: `"none"`
  - Options: `"none"`, `"regular"`, `"high"`, `"full"`
  - Examples: "none"

- **`camera_lora`** (`CameraLoRAEnum`, _optional_):
  The camera LoRA to use. This allows you to control the camera movement of the generated video more accurately than just prompting the model to move the camera. Default value: `"none"`
  - Default: `"none"`
  - Options: `"dolly_in"`, `"dolly_out"`, `"dolly_left"`, `"dolly_right"`, `"jib_up"`, `"jib_down"`, `"static"`, `"none"`
  - Examples: "none"

- **`camera_lora_scale`** (`float`, _optional_):
  The scale of the camera LoRA to use. This allows you to control the camera movement of the generated video more accurately than just prompting the model to move the camera. Default value: `1`
  - Default: `1`
  - Range: `0` to `1`

- **`negative_prompt`** (`string`, _optional_):
  The negative prompt to generate the video from. Default value: `"news broadcast, 3d animation, computer graphics, pc game, console game, video game, cartoon, childish, watermark, logo, text, on screen text, subtitles, titles, signature, slowmo, static"`
  - Default: `"news broadcast, 3d animation, computer graphics, pc game, console game, video game, cartoon, childish, watermark, logo, text, on screen text, subtitles, titles, signature, slowmo, static"`

- **`seed`** (`integer`, _optional_):
  The seed for the random number generator.

- **`enable_prompt_expansion`** (`boolean`, _optional_):
  Whether to enable prompt expansion. Default value: `true`
  - Default: `true`

- **`enable_safety_checker`** (`boolean`, _optional_):
  Whether to enable the safety checker. Default value: `true`
  - Default: `true`

- **`video_output_type`** (`VideoOutputTypeEnum`, _optional_):
  The output type of the generated video. Default value: `"X264 (.mp4)"`
  - Default: `"X264 (.mp4)"`
  - Options: `"X264 (.mp4)"`, `"VP9 (.webm)"`, `"PRORES4444 (.mov)"`, `"GIF (.gif)"`

- **`video_quality`** (`VideoQualityEnum`, _optional_):
  The quality of the generated video. Default value: `"high"`
  - Default: `"high"`
  - Options: `"low"`, `"medium"`, `"high"`, `"maximum"`

- **`video_write_mode`** (`VideoWriteModeEnum`, _optional_):
  The write mode of the generated video. Default value: `"balanced"`
  - Default: `"balanced"`
  - Options: `"fast"`, `"balanced"`, `"small"`

- **`sync_mode`** (`boolean`, _optional_):
  If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
  - Default: `false`

- **`distill_lora_second_pass_scale`** (`float`, _optional_):
  The scale of the distill LoRA to use for the second and subsequent passes. Default value: `0.5`
  - Default: `0.5`
  - Range: `0` to `1`

- **`image_strength`** (`float`, _optional_):
  The strength of the image to use for the video generation. Default value: `1`
  - Default: `1`
  - Range: `0` to `1`

- **`end_image_strength`** (`float`, _optional_):
  The strength of the end image to use for the video generation. Default value: `1`
  - Default: `1`
  - Range: `0` to `1`

- **`audio_strength`** (`float`, _optional_):
  Audio conditioning strength. Values below 1.0 will allow the model to change the audio, while a value of exactly 1.0 will use the input audio without modification. Default value: `1`
  - Default: `1`
  - Range: `0` to `1`

- **`preprocess_audio`** (`boolean`, _optional_):
  Whether to preprocess the audio before using it as conditioning. Default value: `true`
  - Default: `true`



**Required Parameters Example**:

```json
{
  "prompt": "A woman speaks to the camera",
  "audio_url": "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-audio.mp3"
}
```

**Full Example**:

```json
{
  "prompt": "A woman speaks to the camera",
  "audio_url": "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-audio.mp3",
  "image_url": "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-image.png",
  "match_audio_length": true,
  "num_frames": 121,
  "video_size": "landscape_16_9",
  "use_multiscale": true,
  "fps": 24,
  "scheduler": "beta",
  "acceleration": "none",
  "camera_lora": "none",
  "camera_lora_scale": 1,
  "negative_prompt": "news broadcast, 3d animation, computer graphics, pc game, console game, video game, cartoon, childish, watermark, logo, text, on screen text, subtitles, titles, signature, slowmo, static",
  "enable_prompt_expansion": true,
  "enable_safety_checker": true,
  "video_output_type": "X264 (.mp4)",
  "video_quality": "high",
  "video_write_mode": "balanced",
  "distill_lora_second_pass_scale": 0.5,
  "image_strength": 1,
  "end_image_strength": 1,
  "audio_strength": 1,
  "preprocess_audio": true
}
```


### Output Schema

The API returns the following output format:

- **`video`** (`VideoFile`, _required_):
  The generated video.
  - Examples: {"file_name":"ltx-2-a2v-output.mp4","content_type":"video/mp4","url":"https://storage.googleapis.com/falserverless/example_outputs/ltx-2-a2v-output.mp4"}

- **`seed`** (`integer`, _required_):
  The seed used for the random number generator.
  - Examples: 175932751

- **`prompt`** (`string`, _required_):
  The prompt used for the generation.
  - Examples: "A woman speaks to the camera"



**Example Response**:

```json
{
  "video": {
    "file_name": "ltx-2-a2v-output.mp4",
    "content_type": "video/mp4",
    "url": "https://storage.googleapis.com/falserverless/example_outputs/ltx-2-a2v-output.mp4"
  },
  "seed": 175932751,
  "prompt": "A woman speaks to the camera"
}
```


## Usage Examples

### cURL

```bash
curl --request POST \
  --url https://fal.run/fal-ai/ltx-2.3-22b/distilled/audio-to-video \
  --header "Authorization: Key $FAL_KEY" \
  --header "Content-Type: application/json" \
  --data '{
     "prompt": "A woman speaks to the camera",
     "audio_url": "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-audio.mp3"
   }'
```

### Python

Ensure you have the Python client installed:

```bash
pip install fal-client
```

Then use the API client to make requests:

```python
import fal_client

def on_queue_update(update):
    if isinstance(update, fal_client.InProgress):
        for log in update.logs:
           print(log["message"])

result = fal_client.subscribe(
    "fal-ai/ltx-2.3-22b/distilled/audio-to-video",
    arguments={
        "prompt": "A woman speaks to the camera",
        "audio_url": "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-audio.mp3"
    },
    with_logs=True,
    on_queue_update=on_queue_update,
)
print(result)
```

### JavaScript

Ensure you have the JavaScript client installed:

```bash
npm install --save @fal-ai/client
```

Then use the API client to make requests:

```javascript
import { fal } from "@fal-ai/client";

const result = await fal.subscribe("fal-ai/ltx-2.3-22b/distilled/audio-to-video", {
  input: {
    prompt: "A woman speaks to the camera",
    audio_url: "https://storage.googleapis.com/falserverless/example_inputs/ltx-2-a2v-input-audio.mp3"
  },
  logs: true,
  onQueueUpdate: (update) => {
    if (update.status === "IN_PROGRESS") {
      update.logs.map((log) => log.message).forEach(console.log);
    }
  },
});
console.log(result.data);
console.log(result.requestId);
```


## Additional Resources

### Documentation

- [Model Playground](https://fal.ai/models/fal-ai/ltx-2.3-22b/distilled/audio-to-video)
- [API Documentation](https://fal.ai/models/fal-ai/ltx-2.3-22b/distilled/audio-to-video/api)
- [OpenAPI Schema](https://fal.ai/api/openapi/queue/openapi.json?endpoint_id=fal-ai/ltx-2.3-22b/distilled/audio-to-video)

### fal.ai Platform

- [Platform Documentation](https://docs.fal.ai)
- [Python Client](https://docs.fal.ai/clients/python)
- [JavaScript Client](https://docs.fal.ai/clients/javascript)
