DLF: Extreme Image Compression with Dual-generative Latent Fusion

Naifu Xue1* , Zhaoyang Jia2* , Jiahao Li3 , Bin Li3 , Yuan Zhang1 , Yan Lu3
1 Communication University of China   2 University of Science and Technology of China
3 Microsoft Research Asia
ICCV 2025

*Naifu Xue and Zhaoyang Jia are visiting students of MSRA.
Teaser

Instead of directly compressing the tokenizer's latent, our DLF splits it into the semantic and detail parts for flexible compression. Meanwhile, we propose a cross-branch interactive design to reduce redundancy between the two branches, thereby minimizing the overall bit cost.

Codes coming soon!

Abstract

Recent studies in extreme image compression have achieved remarkable performance by compressing the tokens from generative tokenizers. However, these methods often prioritize clustering common semantics within the dataset, while overlooking the diverse details of individual objects. Consequently, this results in suboptimal reconstruction fidelity, especially at low bitrates. To address this issue, we introduce a Dual-generative Latent Fusion (DLF) paradigm. DLF decomposes the latent into semantic and detail elements, compressing them through two distinct branches. The semantic branch clusters high-level information into compact tokens, while the detail branch encodes perceptually critical details to enhance the overall fidelity. Additionally, we propose a cross-branch interactive design to reduce redundancy between the two branches, thereby minimizing the overall bit cost. Experimental results demonstrate the impressive reconstruction quality of DLF even below 0.01 bits per pixel (bpp). On the CLIC2020 test set, our method achieves bitrate savings of up to 27.93% on LPIPS and 53.55% on DISTS compared to MS-ILLM. Furthermore, DLF surpasses recent diffusion-based codecs in visual fidelity while maintaining a comparable level of generative realism.

BibTeX


@article{xue2025dlf,
  title={DLF: Extreme Image Compression with Dual-generative Latent Fusion},
  author={Xue, Naifu and Jia, Zhaoyang and Li, Jiahao and Li, Bin and Zhang, Yuan and Lu, Yan},
  journal={arXiv preprint arXiv:2503.01428},
  year={2025}
}