Back to Search Start Over

Contrastive language and vision learning of general fashion concepts

Authors :
Patrick John Chia
Giuseppe Attanasio
Federico Bianchi
Silvia Terragni
Ana Rita Magalhães
Diogo Goncalves
Ciro Greco
Jacopo Tagliabue
Source :
Scientific Reports, Vol 12, Iss 1, Pp 1-13 (2022)
Publication Year :
2022
Publisher :
Nature Portfolio, 2022.

Abstract

Abstract The steady rise of online shopping goes hand in hand with the development of increasingly complex ML and NLP models. While most use cases are cast as specialized supervised learning problems, we argue that practitioners would greatly benefit from general and transferable representations of products. In this work, we build on recent developments in contrastive learning to train FashionCLIP, a CLIP-like model adapted for the fashion industry. We demonstrate the effectiveness of the representations learned by FashionCLIP with extensive tests across a variety of tasks, datasets and generalization probes. We argue that adaptations of large pre-trained models such as CLIP offer new perspectives in terms of scalability and sustainability for certain types of players in the industry. Finally, we detail the costs and environmental impact of training, and release the model weights and code as open source contribution to the community.

Subjects

Subjects :
Medicine
Science

Details

Language :
English
ISSN :
20452322
Volume :
12
Issue :
1
Database :
Directory of Open Access Journals
Journal :
Scientific Reports
Publication Type :
Academic Journal
Accession number :
edsdoj.3100ce49df1460988e379136d7eaad3
Document Type :
article
Full Text :
https://doi.org/10.1038/s41598-022-23052-9