Published December 12, 2021 | Version v1
Conference paper Open

Conditioned Image Retrieval for Fashion using Contrastive Learning and CLIP-based Features

  • 1. University of Florence, Italy

Description

Building on the recent advances in multimodal zero-shot represen-
tation learning, in this paper we explore the use of features obtained
from the recent CLIP model to perform conditioned image retrieval.
Starting from a reference image and an additive textual description
of what the user wants with respect to the reference image, we
learn a Combiner network that is able to understand the image
content, integrate the textual description and provide combined
feature used to perform the conditioned image retrieval. Starting
from the bare CLIP features and a simple baseline, we show that
a carefully crafted Combiner network, based on such multimodal
features, is extremely effective and outperforms more complex state
of the art approaches on the popular FashionIQ dataset.

Files

3469877.3493593.pdf

Files (566.2 kB)

Name Size Download all
md5:7f582bbbdedea53ae1b0570ee32e7819
566.2 kB Preview Download

Additional details

Funding

AI4Media – A European Excellence Centre for Media, Society and Democracy 951911
European Commission