Learning semantic representations of objects and their parts

  • Authors:
  • Grégoire Mesnil;Antoine Bordes;Jason Weston;Gal Chechik;Yoshua Bengio

  • Affiliations:
  • LISA, Université de Montréal, Montreal, Canada and LITIS, Université de Rouen, Rouen, France;CNRS--Heudiasyc UMR 7253, Université de Technologie de Compiègne, Compiégne, France;Google, New York, USA;Google, Mountain View, USA and Gonda Brain research center, Bar-Ilan University, Ramat Gan, Israel;LISA, Université de Montréal, Montreal, Canada

  • Venue:
  • Machine Learning
  • Year:
  • 2014

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recently, large scale image annotation datasets have been collected with millions of images and thousands of possible annotations. Latent variable models, or embedding methods, that simultaneously learn semantic representations of object labels and image representations can provide tractable solutions on such tasks. In this work, we are interested in jointly learning representations both for the objects in an image, and the parts of those objects, because such deeper semantic representations could bring a leap forward in image retrieval or browsing. Despite the size of these datasets, the amount of annotated data for objects and parts can be costly and may not be available. In this paper, we propose to bypass this cost with a method able to learn to jointly label objects and parts without requiring exhaustively labeled data. We design a model architecture that can be trained under a proxy supervision obtained by combining standard image annotation (from ImageNet) with semantic part-based within-label relations (from WordNet). The model itself is designed to model both object image to object label similarities, and object label to object part label similarities in a single joint system. Experiments conducted on our combined data and a precisely annotated evaluation set demonstrate the usefulness of our approach.