CycleDiff: Cycle Diffusion Models
for Unpaired Image-to-image Translation

Shilong Zou, Yuhang Huang, Renjiao Yi, Chenyang Zhu*, Kai Xu*
National University of Defense and Technology
TIP under review

Indicates Equal Contribution, * Co-corresponding Author
Teaser image demonstrating Marigold depth estimation.

The proposed CycleDiff consists of two domain-specific diffusion models and a cycle translator, and learns the diffusion and translation processes jointly. The cycle translator consists of two translation network used for performing cycle translation between two domains: Gϕ:ST and Fψ:TS. We employ the cycle consistency constrain to regularize the forward and backward translation mappings. Utilizing only unpaired images, CycleDiff can synthesize structure-consistent and photo-realistic results across different modalities of images.

Overview

We introduce a diffusion-based cross-domain image translator in the absence of paired training data. Unlike GAN-based methods, our approach integrates diffusion models to learn the image translation process, allowing for more coverable modeling of the data distribution and performance improvement of the cross-domain translation. However, incorporating the translation process within the diffusion process is still challenging since the two processes are not aligned exactly, i.e., the diffusion process is applied to the noisy signal while the translation process is conducted on the clean signal. As a result, recent diffusion-based studies employ separate training or shallow integration to learn the two processes, yet this may cause the local minimal of the translation optimization, constraining the effectiveness of diffusion models.

To address the problem, we propose a novel joint learning framework that aligns the diffusion and the translation process, thereby improving the global optimality. Specifically, we propose to extract the image components with diffusion models to represent the clean signal and employ the translation process with the image components, enabling an end-to-end joint learning manner. On the other hand, we introduce a time-dependent translation network to learn the complex translation mapping, resulting in effective translation learning and significant performance improvement. Benefiting from the design of joint learning, our method enables global optimization of both processes, enhancing the optimality and achieving improved fidelity and structural consistency.

We have conducted extensive experiments on RGB⇆RGB and cross-modality translation tasks including RGB⇆Edge, RGB⇆Semantics and RGB⇆Depth, showcasing better generative performances than the state of the arts. Especially, our method achieves the best FID score in widely-adopted tasks and outperforms the second-best method with an improved FID of 19.61 and 19.67 on Dog⇆Cat and Dog⇆Wild respectively.

Method and main results

Method

MY ALT TEXT

The overall architecture of CycleDiff. CycleDiff comprises two parts: the diffusion models and the cycle translator. The diffusion models are employed to extract image components, which are then fed into the cycle translator for unpaired translation between two domains. The diffusion and translation processes are learned jointly.

Comparison with other methods

MY ALT TEXT

Quantitative comparison of unpaired image-to-image translation methods. The best results are shown in bold, and the second-best results are underline. Note that the KID metric is multiplied by 100.

More Visual Results

Other interesting results

    ☑ Training in a 640 * 320 resolution on RGB⇆Dpeth

BibTeX

@article{zou2025cyclediff,
    title    = {CycleDiff: Cycle Diffusion Models for Unpaired Image-to-image Translation},
    author   = {Zou, Shilong and Huang, Yuhang and Yi, Renjiao and Zhu, Chenyang and Xu, Kai},
    journal  = {arXiv preprint arXiv:2508.06625},
    year     = {2025}
}