Authors
Wei-Lin Hsiao, Isay Katsman, Chao-Yuan Wu, Devi Parikh, Kristen Grauman
Publication date
2019/4/19
Conference
International Conference on Computer Vision (ICCV)
Description
Given an outfit, what small changes would most improve its fashionability? This question presents an intriguing new computer vision challenge. We introduce Fashion++, an approach that proposes minimal adjustments to a full-body clothing outfit that will have maximal impact on its fashionability. Our model consists of a deep image generation neural network that learns to synthesize clothing conditioned on learned per-garment encodings. The latent encodings are explicitly factorized according to shape and texture, thereby allowing direct edits for both fit/presentation and color/patterns/material, respectively. We show how to bootstrap Web photos to automatically train a fashionability model, and develop an activation maximization-style approach to transform the input image into its more fashionable self. The edits suggested range from swapping in a new garment to tweaking its color, how it is worn (eg, rolling up sleeves), or its fit (eg, making pants baggier). Experiments demonstrate that Fashion++ provides successful edits, both according to automated metrics and human opinion.
Total citations
20192020202120222023202451827202110
Scholar articles
WL Hsiao, I Katsman, CY Wu, D Parikh, K Grauman - Proceedings of the IEEE/CVF International Conference …, 2019