Mitigating Occlusions in Virtual Try-On via A Simple-Yet-Effective Mask-Free Framework

The Thirty-Ninth Annual Conference on Neural Information Processing Systems
🔥(NeurIPS 2025)🔥
1Wuhan University of Technology (武汉理工大学),
2Interdisciplinary Artificial Intelligence Research Institute,
3Sanya Science and Education Innovation Park,
4Shanghai AI Laboratory


Abstract

This work tackles occlusion issues in Virtual Try-On (VTON).
We taxonomize failures into:

1. Inherent Occlusions – “ghost” garments from the reference image that remain in the result.

2. Acquired Occlusions – distorted human anatomy that visually blocks the new outfit.

To remove both, we propose a mask-free VTON framework with two plug-and-play operations:

- Background Pre-Replacement – swaps the background before generation so the model never confuses clothes with body/background, suppressing inherent occlusions.

- Covering-and-Eliminating – enforces human-aware semantics, yielding anatomically plausible shapes and thus fewer acquired occlusions.

The operations are architecture-agnostic: drop them into GANs or diffusion models without re-design.


Paper and Supplementary Material

[Paper] [Supplementary Material] [Code]

NeurIPS, 2023.
Chenghu Du, Shengwu Xiong, Junyin Wang, and others.
"Mitigating Occlusions in Virtual Try-On via A
Simple-Yet-Effective Mask-Free Framework"


Experiments

The test pair and test results on VITON dataset are shown this and here, from left to right are reference person, target clothes, try-on results of five baseline methods including CP-VITON+ (CVPRW 2020), ACGPN (CVPR 2020), DCTON (CVPR 2021), RT-VITON (CVPR 2022), and USC-PFN.


Results (Dress)

Results (Upper)

Results (Lower)





BibTeX


    @article{du2025mitigating,
      title={Mitigating Occlusions in Virtual Try-On via A Simple-Yet-Effective Mask-Free Framework},
      author={Du, Chenghu and Xiong, Shengwu and Wang, Junyin and Rong, Yi and Xiong, Shili},
      journal={Advances in Neural Information Processing Systems},
      year={2025}
    }