OpenFashionCLIP: Vision-and-Language Contrastive Learning with Open-Source Fashion Data

Giuseppe Cartella, Alberto Baldrati, Davide Morelli, Marcella Cornia, Marco Bertini, Rita Cucchiara

Research output: Book chapterConference contributionpeer-review

1 Citation (Scopus)

Abstract

The inexorable growth of online shopping and e-commerce demands scalable and robust machine learning-based solutions to accommodate customer requirements. In the context of automatic tagging classification and multimodal retrieval, prior works either defined a low generalizable supervised learning approach or more reusable CLIP-based techniques while, however, training on closed source data. In this work, we propose OpenFashionCLIP, a vision-and-language contrastive learning method that only adopts open-source fashion data stemming from diverse domains, and characterized by varying degrees of specificity. Our approach is extensively validated across several tasks and benchmarks, and experimental results highlight a significant out-of-domain generalization capability and consistent improvements over state-of-the-art methods both in terms of accuracy and recall. Source code and trained models are publicly available at: https://github.com/aimagelab/open-fashion-clip.
Original languageEnglish
Title of host publicationImage Analysis And Processing, Iciap 2023, Pt I
EditorsGL Foresti, A Fusiello, E Hancock
PublisherSpringer Nature
Pages245-256
Number of pages12
Volume14233
ISBN (Electronic)978-3-031-43148-7
ISBN (Print)978-3-031-43147-0
DOIs
Publication statusPublished - 2023
Externally publishedYes
Event22nd International Conference on Image Analysis and Processing (ICIAP) - Udine, Italy
Duration: 11 Sept 202315 Sept 2023

Publication series

NameLecture Notes In Computer Science

Conference

Conference22nd International Conference on Image Analysis and Processing (ICIAP)
Country/TerritoryItaly
CityUdine
Period11/09/2315/09/23

Keywords

  • Fashion Domain
  • Open-Source Datasets
  • Vision-and-Language Pre-Training

Fingerprint

Dive into the research topics of 'OpenFashionCLIP: Vision-and-Language Contrastive Learning with Open-Source Fashion Data'. Together they form a unique fingerprint.

Cite this