At CES 2026, Lightricks released the highly anticipated open weights of the LTX-2 audio-video model, marking a major step forward in AI video and audio generation. Optimized for NVIDIA GPUs, LTX-2 is the leading open-weights audio-video model, capable of generating clips of up to 4K resolution, 50 FPS, and up to 20 seconds long.
The models are now available for download in BF16 precision. The base model is also available in quantized NVFP8 weights that cut model size by roughly 30% and can deliver up to 2x faster performance on RTX GPUs.
This guide gets you running with an RTX-optimized ComfyUI workflow in minutes.
LTX-2 is a family of audio-video models that generate videos with audio. There are five checkpoints coming at launch:
As a frontier model, LTX-2 uses significant amounts of video memory (VRAM) to deliver quality results. Memory use goes up as we increase resolution, framerate, length, or steps. Fortunately for users, ComfyUI and NVIDIA have collaborated to optimize a weight streaming feature, allowing users to offload parts of the workflow to system memory if your GPU runs out of VRAM, but this will come at a cost in performance.
Depending on your GPU and use case, you may want to constrain these factors to ensure reasonable generation times. For example, GeForce RTX 5090 GPUs have 32GB of VRAM, and can generate a 720p 24fps 4-second clip within GPU memory in about 25 seconds. However, if a user wants a longer 8-second video, the generation time will increase to three minutes because it will require more than 32GB of VRAM and automatically engage weight streaming.
Recommendation: use lower settings to iterate on your video, then increase the settings to tune the quality to what you want. In our experience it’s best to:
LTX-2 is an advanced model capable of generating amazing videos. But as with any model, tweaking the settings will have a big impact on quality. The community will come up with fantastic recommendations as the model weights become available, but here are pro tips we’ve found help the most in our testing: