Training LTX-2
The LTX-2 Trainer enables researchers and advanced users to fine-tune the LTX-2 model on custom datasets. You can train LoRAs, perform full fine-tuning, or create video-to-video transformations (IC-LoRA).
Who Should Use the Trainer?
The trainer is designed primarily for:
- Researchers exploring video generation model training techniques
- Advanced users building specialized video generation models
Requirements
Before starting, ensure you have:
- LTX-2 Model Checkpoint - Local
.safetensorsfile - Gemma Text Encoder - Local Gemma model directory (download from HuggingFace)
- Linux with CUDA - CUDA 13+ recommended
- Nvidia H100 GPU with 80GB+ VRAM - Lower VRAM may work with gradient checkpointing and reduced resolutions
Additional Documentation
For comprehensive training documentation, including setup instructions, dataset preparation guides, configuration options, and troubleshooting, visit the LTX-2 Trainer GitHub repository.
The GitHub documentation includes:
- Quick Start Guide - Get training in minutes
- Dataset Preparation - Video preprocessing and captioning
- Training Modes - Detailed explanations of each training approach
- Configuration Reference - All available training parameters
- Troubleshooting - Solutions to common issues
- Utility Scripts - Tools for dataset management