
LTX-2
Video
LTX-2 (Lightricks)
Text-and-image-to-video generation with synchronized audio, built as a single unified model.
Audio + video together. Generates visuals and matching sound (dialogue, music, ambience) in one pass instead of stitching separate models.
High-resolution output. Designed for native high-res video (up to 4K) with strong temporal consistency across frames.
Unified diffusion architecture. One foundation model handles motion, scene dynamics, and sound alignment, reducing pipeline complexity.
Developer-oriented. Designed to respond well to structured prompts, enabling easier testing,Open weights. Fully open-source release intended for local runs, customization, fine-tuning, and research use.
Production-aware design. Built with real creative workflows in mind, supporting structured prompting and extension via LoRAs.
Why pick it for Norman AI?
LTX-2 is a strong choice when you want full audio-visual generation in a single stack. It simplifies experimentation with video + sound, reduces multi-model orchestration, and gives developers direct control over performance, quality, and customization.
