Command Palette
Search for a command to run...
Fast Zero-Shot Image Tagging
Fast Zero-Shot Image Tagging
Zhang Yang Gong Boqing Shah Mubarak
Abstract
The well-known word analogy experiments show that the recent word vectorscapture fine-grained linguistic regularities in words by linear vector offsets,but it is unclear how well the simple vector offsets can encode visualregularities over words. We study a particular image-word relevance relation inthis paper. Our results show that the word vectors of relevant tags for a givenimage rank ahead of the irrelevant tags, along a principal direction in theword vector space. Inspired by this observation, we propose to solve imagetagging by estimating the principal direction for an image. Particularly, weexploit linear mappings and nonlinear deep neural networks to approximate theprincipal direction from an input image. We arrive at a quite versatile taggingmodel. It runs fast given a test image, in constant time w.r.t.\ the trainingset size. It not only gives superior performance for the conventional taggingtask on the NUS-WIDE dataset, but also outperforms competitive baselines onannotating images with previously unseen tags