Model Gallery
Featured Models
Check out some of our most popular models
Train Hunyuan Video lora on people, objects, characters and more!
Generate video clips from your images using Kling 1.6 (pro)
LatentSync is a video-to-video model that generates lip sync animations from audio using advanced algorithms for high-quality synchronization.
Search Results
61 models found
The video upscaler endpoint uses RealESRGAN on each frame of the input video to upscale the video to a higher resolution.
Generate videos from prompts using CogVideoX-5B
Generate videos from videos and prompts using CogVideoX-5B
Generate videos from images and prompts using CogVideoX-5B
Hunyuan Video is an Open video generation model with high visual quality, motion diversity, text-video alignment, and generation stability
Hunyuan Video is an Open video generation model with high visual quality, motion diversity, text-video alignment, and generation stability
Generate videos from prompts using LTX Video
Generate videos from images using LTX Video
Generate short video clips from your prompts using SVD v1.1
Generate short video clips from your images using SVD v1.1 at Lightning Speed
Generate short video clips from your images using SVD v1.1 at Lightning Speed
Re-animate your videos with evolved consistency!
Re-animate your videos with evolved consistency!
Generate short video clips from your prompts
Sa2VA is an MLLM capable of question answering, visual prompt understanding, and dense object segmentation at both image and video levels
Sa2VA is an MLLM capable of question answering, visual prompt understanding, and dense object segmentation at both image and video levels
Generate video clips from your prompts using MiniMax model
Generate video clips from your images using MiniMax Video model
Transform text into hyper-realistic videos with Haiper 2.0. Experience industry-leading resolution, fluid motion, and rapid generation for stunning AI videos.
Generate video clips from your images using MiniMax Video model
Generate video clips from your prompts using MiniMax model
Transform text into hyper-realistic videos with Haiper 2.0. Experience industry-leading resolution, fluid motion, and rapid generation for stunning AI videos.
Generate short video clips from your images using SVD v1.1
Generate video clips from your prompts using Luma Dream Machine v1.5
Generate video clips from your prompts using Kling 1.0
Generate video clips from your images using Kling 1.0
Generate video clips from your prompts using Kling 1.0 (pro)
Generate video clips from your images using Kling 1.0 (pro)
Generate video clips from your images using Kling 1.5 (pro)
Generate video clips from your prompts using Kling 1.5 (pro)
Generate video clips from your images using Kling 1.6 (std)
Generate video clips from your prompts using Kling 1.6 (std)
Generate video clips from your images using Luma Dream Machine v1.5
MMAudio generates synchronized audio given video and/or text inputs. It can be combined with video models to get videos with audio.
Transform text into stunning videos with TransPixar - an AI model that generates both RGB footage and alpha channels, enabling seamless compositing and creative video effects.
Interpolate between video frames
Re-animate your videos in lightning speed!
This endpoint delivers seamlessly localized videos by generating lip-synced dubs in multiple languages, ensuring natural and immersive multilingual experiences
Mochi 1 preview is an open state-of-the-art video generation model with high-fidelity motion and strong prompt adherence in preliminary evaluation.
Automatically generates text captions for your videos from the audio as per text colour/font specifications
Re-animate your videos!
Transfer expression from a video to a portrait.
Transfer expression from a video to a portrait.
SAM 2 is a model for segmenting images and videos in real-time.
SAM 2 is a model for segmenting images and videos in real-time.
Multimodal vision-language model for video understanding
Animate a reference image with a driving video using ControlNeXt.
Sa2VA is an MLLM capable of question answering, visual prompt understanding, and dense object segmentation at both image and video levels
Sa2VA is an MLLM capable of question answering, visual prompt understanding, and dense object segmentation at both image and video levels
Bria RMBG 2.0 enables seamless removal of backgrounds from images, ideal for professional editing tasks. Trained exclusively on licensed data for safe and risk-free commercial use. Model weights for commercial use are available here: https://share-eu1.hsforms.com/2GLpEVQqJTI2Lj7AMYwgfIwf4e04?utm_campaign=RMBG%202.0&utm_source=RMBG%20image%20and%20video%20page&utm_medium=button&utm_content=rmbg%20image%20pricing%20form
Generate realistic lipsync animations from audio using advanced algorithms for high-quality synchronization.
Interpolate between image frames
Animate your ideas!
Animate your ideas in lightning speed!
Animate Your Drawings with Latent Consistency Models!
Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation
MuseTalk is a real-time high quality audio-driven lip-syncing model. Use MuseTalk to animate a face with your own audio.
Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation