Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Virtual try-on technology seeks to seamlessly integrate an image of a specified garment onto the target person, generating a synthesized image that realistically depicts the person wearing the clothing. Existing methods based on generative adversarial network (GAN) for clothing warping in the generation process usually use human pose- and body parsing-based features to guide the distortion of a flattened clothing item. However, it is hard using these approaches to accurately capture the spatial characteristics of the distorted clothing (e.g., wrinkles on the clothing). In this research, we propose a Wrinkle-Preserving Virtual Try-On Network, named WP-VTON, to address the aforementioned issues exist in the virtual try-on task. Specifically, in the clothing warping stage, we incorporate the normal features extracted from spatial attributes of both clothing and the human body to learn about the clothing deformation caused by warping; in the try-on generation stage, we leverage a pre-trained StyleGAN, called clothing texture book, to optimize the try-on image, with the aim of further improving the generation capability of WP-VTON with regard to texture details. Experimental results in public datasets demonstrate the effectiveness of our method by outperforming the state-of-the-art GAN-based virtual try-on models.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.neunet.2025.107546DOI Listing

Publication Analysis

Top Keywords

virtual try-on
20
clothing
10
wrinkle-preserving virtual
8
try-on network
8
clothing texture
8
texture book
8
clothing warping
8
try-on
7
virtual
5
wp-vton wrinkle-preserving
4

Similar Publications

Image-based virtual try-on aims to generate realistic images of individuals wearing target garments by synthesizing input clothing and person images. Traditional methods often follow separate stages, including garment warping, segmentation map generation, and final image synthesis. However, the lack of interaction between these stages frequently causes misalignments and visual artifacts, particularly in scenarios involving occlusions or complex poses.

View Article and Find Full Text PDF

We introduce a monocular-to-3D virtual try-on network based on a conditional 3D-aware Generative Adversarial Network (3D-GAN) for synthesizing multi-view try-on results from single monocular images. In contrast to previous 3D virtual try-on methods that rely on costly scanned meshes or pseudo-depth maps for supervision, our approach utilizes a conditional 3D-GAN trained solely on 2D images, greatly simplifying dataset construction and enhancing model scalability. Specifically, we propose a Generative monocular-to-3D Virtual Try-ON network (G3D-VTON) that integrates a 3D-aware conditional Parsing Module (3DPM), a U-Net Refinement Module (URM), and a Flow-based 2D Virtual Try-On Module (FTM).

View Article and Find Full Text PDF

Enhancing image-based virtual try-on with Multi-Controlled Diffusion Models.

Neural Netw

September 2025

Key Laboratory of Textile Science & Technology, Ministry of Education, College of Textiles, Donghua University, Shanghai, 201620, China. Electronic address:

Image-based virtual try-on technology digitally overlays clothing onto images of individuals, enabling users to preview how garments fit without physical trial, thus enhancing the online shopping experience. While current diffusion-based virtual try-on networks produce high-quality results, they struggle to accurately render garments with textual designs such as logos or prints which are widely prevalent in the real world, often carrying significant brand and cultural identities. To address this challenge, we introduce the Multi-Controlled Diffusion Models for Image-based Virtual Try-On (MCDM-VTON), a novel approach that synergistically incorporates global image features and local textual features extracted from garments to control the generation process.

View Article and Find Full Text PDF

Virtual try-on technology seeks to seamlessly integrate an image of a specified garment onto the target person, generating a synthesized image that realistically depicts the person wearing the clothing. Existing methods based on generative adversarial network (GAN) for clothing warping in the generation process usually use human pose- and body parsing-based features to guide the distortion of a flattened clothing item. However, it is hard using these approaches to accurately capture the spatial characteristics of the distorted clothing (e.

View Article and Find Full Text PDF

Image-based virtual try-on aims to fit garments onto a target person accurately and naturally while preserving the textural details of the garment. Inspired by the dynamic perception process of the human visual system, which transitions from global perception to local details, we propose a novel multigrained adaptive fusion network for virtual try-on framework named MA-VITON. MA-VITON precisely aligns clothing semantic features with human body parts across different scales, reduces unrealistic textures caused by garment distortion, and employs coarse-to-fine clothing features to progressively guide the generation of try-on results.

View Article and Find Full Text PDF