Improved Training Technique for Latent Consistency Models
Abstract
Consistency models are a new family of generative models capable of producing high-quality samples in either a single step or multiple steps. Recently, consistency models have demonstrated impressive performance, achieving results on par with diffusion models in the pixel space. However, the success of scaling consistency training to large-scale datasets, particularly for text-to-image and video generation tasks, is determined by performance in the latent space. In this work, we analyze the statistical differences between pixel and latent spaces, discovering that latent data often contains highly impulsive outliers, which significantly degrade the performance of iCT in the latent space. To address this, we replace Pseudo-Huber losses with Cauchy losses, effectively mitigating the impact of outliers. Additionally, we introduce a diffusion loss at early timesteps and employ optimal transport (OT) coupling to further enhance performance. Lastly, we introduce the adaptive scaling-c scheduler to manage the robust training process and adopt Non-scaling LayerNorm in the architecture to better capture the statistics of the features and reduce outlier impact. With these strategies, we successfully train latent consistency models capable of high-quality sampling with one or two steps, significantly narrowing the performance gap between latent consistency and diffusion models. The implementation is released here: https://github.com/quandao10/sLCT/
Community
Robust and stable framework for training latent consistency model from scratch without any pretrained diffusion model
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Self-Corrected Flow Distillation for Consistent One-Step and Few-Step Text-to-Image Generation (2024)
- Robust Representation Consistency Model via Contrastive Denoising (2025)
- Exploring Representation-Aligned Latent Space for Better Generation (2025)
- Weak Supervision Dynamic KL-Weighted Diffusion Models Guided by Large Language Models (2025)
- Improving Source Extraction with Diffusion and Consistency Models (2024)
- From Slow Bidirectional to Fast Autoregressive Video Diffusion Models (2024)
- Taming Feed-forward Reconstruction Models as Latent Encoders for 3D Generative Models (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper