Please use this identifier to cite or link to this item: https://elib.vku.udn.vn/handle/123456789/6212
Title: Auto-ARM: An Autonomous Adaptive Mask Refinement Mechanism for Enhancing Naturalness in Virtual Try-On Models
Authors: Bui, D. Hung
Nguyen, P. H. Phuc
Dang, Q. Vinh
Huynh, Nga
Vo, D. Hung
Nguyen, S. T. Long
Quan, T. Tho
Keywords: Virtual try-on
Mask refinement
Attention U-Net
Issue Date: Jan-2026
Publisher: Springer Nature
Abstract: Virtual Try-on (VTON or VITON) technology has become a cornerstone of e-commerce, offering users an immersive and personalized shopping experience. Recent advancements in diffusion models have improved the quality of try-on images. However, these models still rely heavily on the accuracy of input try-on masks, which are the masked regions used to instruct VTON models to generate the target clothing on. Furthermore, such approaches apply rule-based methods to produce try-on masks, which lack flexibility and can lead to distortion or incomplete clothing replacement, especially with unnatural poses or mixed clothes. To address these limitations, we introduce Auto-ARM, an innovative framework that employs an attention-based U-Net architecture with Attention Gate (AG) to dynamically refine the try-on masks based on the target outfit. This novel approach not only significantly enhances the generalization of mask-dependent VTON models but also delivers superior qualitative and quantitative results. Auto-ARM achieves state-of-the-art performance on benchmarks such as VITON-HD and DressCode, proving its potential for high-quality, real-world VTON applications.
Description: Lecture Notes in Networks and Systems (LNNS,volume 1581); The 14th Conference on Information Technology and Its Applications (CITA 2025) ; pp: 303-316
URI: https://doi.org/10.1007/978-3-032-00972-2_23
https://elib.vku.udn.vn/handle/123456789/6212
ISBN: 978-3-032-00971-5 (p)
978-3-032-00972-2 (e)
Appears in Collections:CITA 2025 (International)

Files in This Item:

 Sign in to read



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.