Scalable Visual Attribute Extraction through Hidden Layers of a Residual ConvNet

Andres Baloian
DCC University of Chile

abaloian@dcc.uchile.cl


Nils Murrugarra-Llerena
Snap Research

mnurrugarraller@snap.com

Jose M. Saavedra
Impresee Inc.

 jose.saavedra@impresee.com

Abstract

Visual attributes play an essential role in real applications based on image retrieval. For instance, the extraction of attributes from images allows an eCommerce search engine to produce retrieval results with higher precision. The traditional manner to build an attribute extractor is by training a convnet-based classifier with a fixed number of classes. However, this approach does not scale for real applications where the number of attributes changes frequently. Therefore in this work, we propose an approach for extracting visual attributes from images, leveraging the learned capability of the hidden layers of a general convolutional network to discriminate among different visual features. We run experiments with a resnet-50 trained on Imagenet, on which we evaluate the output of its different blocks to discriminate between colors and textures. Our results show that the second block of the resnet is appropriate for discriminating colors, while the fourth block can be used for textures. In both cases, the achieved accuracy of attribute classification is superior to 93%. We also show that the proposed embeddings form local structures in the underlying feature space, which makes it possible to apply reduction techniques like UMAP, maintaining high accuracy and widely reducing the size of the feature space.

Would you like to collaborate?

Join us and transform the world!

 

6 + 5 =

Soluciones

Creative Search Bar & Filters
Creative Memory

Nosotros

Nuestra historia
Partners
Precios
Política de privacidad

Plataformas

WooCommerce
Shopify
Magento
Others

Recursos

Blog
Tutoriales
eBooks

Contacto

Agendar demo