Abstract
We propose an end-to-end deep convolutional network to simultaneously localize and rank relative visual attributes, given only weakly-supervised pairwise image comparisons. Unlike previous methods, our network jointly learns the attribute’s features, localization, and ranker. The localization module of our network discovers the most informative image region for the attribute, which is then used by the ranking module to learn a ranking model of the attribute. Our end-to-end framework also significantly speeds up processing and is much faster than previous methods. We show state-of-the-art ranking results on various relative attribute datasets, and our qualitative localization results clearly demonstrate our network’s ability to learn meaningful image patches.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Visual attributes are mid-level representations that describe semantic properties (e.g., ‘furry’, ‘natural’, ‘short’) of objects and scenes, and have been explored extensively for various applications including zero-shot learning [1–3], image retrieval [4–6], fine-grained recognition [7–9], and human computer interaction [10]. Attributes have been studied in the binary [7, 11]—describing their presence or absence—and relative [3, 12]—describing their relative strength—settings.
Recent work on visual attributes have shown that local representations often lead to better performance compared to global representations [9, 11, 13]. These methods use pre-trained part detectors to bring the candidate object parts into correspondence to model the attribute, with the assumption that there is at least one well-defined part that corresponds to the attribute. However, this assumption does not always hold; for example, the exact spatial extent of the attribute bald head can be ambiguous, which means that training a bald head detector itself can be difficult. Furthermore, since the part detectors are trained independently of the attribute, their learned parts may not necessarily be useful for modeling the attribute. Finally, these methods are designed for binary attributes and are not applicable for relative attributes; however, relative attributes have been shown to be equally or more useful in many settings [3, 12].
Recently, Xiao and Lee [14] proposed an algorithm that overcomes the above drawbacks by automatically discovering the relevant spatial extent of relative attributes given only weakly-supervised (i.e., image-level) pairwise comparisons. The key idea is to transitively connect “visual chains” that localize the same visual concept (e.g., object part) across the attribute spectrum, and then to select the chains that together best model the attribute. The approach produces state-of-the-art performance for relative attribute ranking. Despite these qualities, it has three main limitations due to its pipeline nature: (1) The various components of the approach, including the feature learning and ranker, are not optimized jointly, which can lead to sub-optimal performance; (2) It is slow due to time-consuming intermediate modules of the pipeline; (3) In order to build the visual chains, the approach assumes the existence of a visual concept that undergoes a gradual visual change along with the change in attribute strength; however, this does not always hold. For example, for the natural attribute for outdoor scenes, there are various visual concepts (e.g., forests and mountains) that are relevant but not consistently present across the images.
To address these issues, in this paper, we propose an end-to-end deep convolutional network that simultaneously learns to rank and localize relative attributes. Our setting is the same as in [14]: we are given only weakly-supervised image-level pairwise attribute comparisons, but no supervision on where in each image the attribute is present nor what the attribute looks like. Thus, the main challenge is to automatically localize the relevant regions in each image pair simultaneously, such that an accurate relative ranking of the attribute for each image pair can be produced. We tackle this challenge by designing a new architecture that combines a localization network with a ranking network, and optimize a pairwise image ranking loss. In this way, our approach learns to focus on the regions in each image that are optimal for attribute ranking. Furthermore, our network is optimized end-to-end, and jointly learns an attribute’s features, localization, and ranker, which mutually benefit each other. This end-to-end framework also significantly speeds up processing. Finally, unlike [14], we do not assume that the attribute must be conditioned on the same visual concept across the attribute spectrum. Instead, our network is free to identify discriminative patterns in each image that are most relevant for attribute ranking (e.g., for the natural attribute, localizing trees for forest images and buildings for city images). We demonstrate that all of these lead to improved performance over standard pipeline approaches.
Briefly, our method works as follows: we train a Siamese network [15] with a pairwise ranking loss, which takes as input a pair of images and a weak-label that compares the relative strength of an attribute for the image pair. A Siamese network consists of two identical parallel branches with shared parameters, and during testing, either branch can be used to assign a ranking score to a single image. Each branch consists of a localization module and a ranking module. The localization module is modeled with spatial transformer [16] layers, which discover the most relevant part of the image corresponding to the attribute (see Fig. 1 for examples of localized patches). The output patch of the localization module is then fed into the ranking module for fine-grained attention for ranking.
Contributions. To our knowledge, this is the first attempt to learn an end-to-end network to rank and localize relative attributes. To accomplish this, we make two main contributions: (1) a new deep convolutional network that learns to rank attributes given pairwise relative comparisons, and (2) integrating the spatial transformer into our network to discover the image patches that are most relevant to an attribute. We demonstrate state-of-the-art results on the LFW-10 [13] face, UT-Zap50K [17] shoe, and OSR [18] outdoor scene datasets.
2 Related Work
Visual attributes. Visual attributes serve as an informative and compact representation for visual data. Earlier work relied on hand-crafted features like SIFT and HOG to model the attributes [1–3, 6, 19–23]. More recent work use deep convolutional networks to learn the attribute representations, and achieve superior performance [9, 24–26]. While these approaches learn deep representations for binary attributes, we instead learn deep representations for relative attributes. Concurrent work [27] also trains a deep CNN for the relative setting; however, it does not perform localization as we do.
Attribute localization. Learning attribute models conditioned on local object parts or keypoints have shown to produce superior performance for various recognition tasks [8, 9, 11, 13, 14, 28]. Most existing work rely on pre-trained part/keypoint detectors or crowd-sourcing to localize the attributes [8, 9, 11, 20, 28, 29]. Recently, Xiao and Lee [14] proposed a method to automatically discover the spatial extent of relative attributes. Since it does not rely on pre-trained detectors, it can model attributes that are not clearly tied to object-parts (e.g., open for shoes). However, the approach consists of several sequential independently-optimized modules in a pipeline system. As a result, it can be suboptimal and slow. In contrast, we propose to localize the attributes and train the attribute models simultaneously in an end-to-end learning framework. Similar to [14], our approach automatically discovers the relevant attribute regions, but is more accurate in ranking and faster since everything is learned jointly.
Attention modeling. Attention models selectively attend to informative locations in the visual data to process at higher resolution. The key idea is to neglect irrelevant regions (like background clutter) and instead focus on the important regions that are relevant to the task at hand. Earlier work on bottom-up saliency (e.g., [30, 31]) focus on interesting regions in the image, while high-level object proposal methods (e.g., [32–35]) generate candidate object-like regions. Recent work use deep networks for attention modeling in various tasks including fine-grained classification [36, 37], image caption generation [38], and image generation [39, 40]. In particular, Spatial Transformer Networks [16] spatially transform the input image to focus on the task-relevant regions, and have shown to improve performance on digit and fine-grained bird recognition. In our work, we show how to integrate spatial transformer layers into a deep pairwise ranking network in order to automatically localize and rank relative attributes for faces, shoes, and outdoor scenes, in the more challenging ranking setting.
3 Approach
Given pairs of training images, with each pair ordered according to relative strength of an attribute, our goal is to train a deep convolutional network that learns a function \(f:\mathbb {R}^M \rightarrow \mathbb {R}\) to simultaneously discover where in each image the attribute is present and rank each image (with M pixels) according to predicted attribute strength. Importantly, the only supervision we have are the pairwise image comparisons; i.e., there is no supervision on where in each image the attribute is present nor prior information about the visual appearance of the attribute.
3.1 Input
For training, the input to our network is an image pair \((I_1,I_2)\) and a corresponding label L for a given attribute (e.g., smile) indicating whether the image pair belongs to set E or Q. \((I_1,I_2){\in }E\) means that the ground-truth attribute strengths of \(I_1\) and \(I_2\) are similar to each other, while \((I_1,I_2){\in }Q\) means that the ground-truth attribute strength of \(I_1\) is greater than that of \(I_2\). (If attribute strength of \(I_1\) is less than that of \(I_2\), we simply reorder the two as \((I_2,I_1){\in }Q\).)
For testing, the input is a single image \(I_{test}\), and our network uses its learned function f (i.e., network weights) to predict the attribute strength \(v=f(I_{test})\).
3.2 Architecture
Figure 2 shows the overall architecture of our network. We have a Siamese network [41], which takes as input an image pair \((I_1, I_2)\) along with its label L, and outputs two scalar values \(v_1=f(I_1)\) and \(v_2=f(I_2)\), which are fed into our loss function. A Siamese network consists of two identical branches \(S_1\) and \(S_2\) with shared parameters. Each branch consists of a Spatial Transformer Network (STN) and a Ranker Network (RN). The STN is responsible for localizing the relevant image patch corresponding to the visual attribute, while the RN is responsible for generating a scalar value v that denotes the input image’s attribute strength. During testing, only one of the branches is used to produce the attribute strength v for the input test image.
Spatial Transformer Network (STN). Intuitively, in order to discover the regions in each training image pair that are relevant to the attribute, we could apply a ranking function to various pairs of regions (one region from each image in the pair), and then select the pair that leads to the best agreement with the ground-truth pairwise rankings. STNs [16] provide an elegant framework for doing so. An STN learns an explicit spatial image (or feature map) transformation for each image that is optimal for the task at hand [16]. It has two main advantages: (1) fully-differentiable and can be trained with backpropagation; and (2) can learn to translate, crop, rotate, scale, or warp an image without any explicit supervision for the transformation. By attending to the most relevant image regions, the STN allows the ensuing computation to be dedicated to those regions. In [16], STNs were shown to learn meaningful transformations for digit classification and fine-grained bird classification.
In this work, we incorporate an STN as part of our end-to-end ranking system, in order to discover the region-of-interest for each relative attribute. The STN’s output can then be fed into the ensuing Ranker network, easing its task. For example, for attribute visible-teeth, it will be easier to optimize the ranking function if the Ranker network receives the mouth region. While there is no explicit human supervision denoting that the mouth is in fact the most relevant for visible-teeth, the STN can learn to attend to it via its contribution to the ranking loss function. The architecture of an STN has three main blocks (see Fig. 3, orange box): it first takes the input image and passes it through a convolutional network to obtain the transformation parameters \(\theta \), and then a grid generator creates a sampling grid, which provides the set of input points that should be sampled to produce the transformed output. With the generated grid and the transformation parameters \(\theta \), a bilinear interpolation kernel is applied to sample the input values to produce the output image. See [16] for more details.
In our STN, we have three transformation parameters \(\theta = [s, t_x, t_y]\), representing isotropic scaling s and horizontal and vertical translation \(t_x\) and \(t_y\).Footnote 1 The transformation is applied as an inverse warp to generate the output image:
where \(x_i^{in}\) and \(y_i^{in}\) are the input image coordinates, \(x_i^{out}\) and \(y_i^{out}\) are the output image coordinates, and i indexes the pixels in the output image.
The convolutional network of our STN has six convolutional layers and two fully-connected layers (see Fig. 3, orange box). The first five layers are equivalent to those of AlexNet [42]. After the max-pooling layer of the 5th convolutional layer (i.e., pool5), we add a convolutional layer consisting of 128 filters of size \( 1 \times 1\) to reduce feature dimensionality. After the convolutional layer, we add two fully-connected layers; the first takes in 4608 values as input and outputs 128 values, while the second takes the 128 values as input and outputs the final 3 transformation parameters (i.e., scale, vertical and horizontal translation). We find these hyperparmeters to provide a good balance of high enough capacity to learn the transformation while minimizing overfitting.
Figure 4 shows the change in the STN output over the training epochs of our network. For each attribute, the first row shows an example image with the STN’s localized patch in the red box, while the second row shows the distribution of the STN’s output over the entire training data overlaid onto the example image (this visualization works because the face images in LFW-10 are roughly aligned). The STN is initially unsure of the attribute’s location and thus has high spatial variance. It then proceeds to search over the various regions in each image and converges to the top of the head for dark-hair and to the mouth for smile.
Ranker Network (RN). The RN takes the output of the STN (i.e., an image patch) and the original image as input, and computes and combines their features to generate a scalar attribute strength v as output (see Fig. 3, blue box). The key idea of combining the two inputs is to provide the ranker with both the high-resolution image patch that is focused on the visual attribute as well as the entire image to provide global context. We demonstrate in our experiments that the two sources of information are indeed complementary.
For image pair \((I_1,I_2){\in }E\), the RN will learn to generate similar values for \(I_1\) and \(I_2\), while for image pair \((I_1,I_2){\in }Q\) the RN will learn to generate a higher value for \(I_1\) than \(I_2\). Ultimately, the RN will learn a global ranking function that tries to satisfy all such pairwise constraints in the training data. Our RN is a Siamese network [41], with each branch consisting of AlexNet [42] without the last fully connected classification layer (i.e., all layers up through fc7). This generates 4096-D feature vectors for both the image patch and global image, which are concatenated to produce the final 8192-D feature. A linear layer takes the 8192-D feature to generate a final single value v for the image. Note that each branch of the RN has shared weights, which reduces the number of parameters by half and helps in reducing overfitting.
3.3 Localization and Ranking Loss Function
To learn the parameters of our network, we train it using the loss function introduced in the seminal work of RankNet [43]. Specifically, we map the outputs \(v_1\) and \(v_2\) (corresponding to \(I_1\) and \(I_2\)), to a probability P via a logistic function \(P= e^{(v_1-v_2)}/(1 + e^{(v_1-v_2)})\), and then optimize the standard cross-entropy loss:
where if \((I_1,I_2){\in }Q\) then \(L=1\), else if \((I_1,I_2){\in }E\) then \(L=0.5\). This loss function enforces \(v_1 > v_2\) when \(I_1\) has a higher ground-truth attribute strength than \(I_2\), and enforces \(v_1 = v_2\) when \(I_1\) and \(I_2\) have similar ground-truth attribute strengths. As described in [43], a nice property of this loss function is that it asymptotes to a linear function, which makes it more robust to noise compared to a quadratic function, and handles input pairs with similar ground-truth attribute strengths in a principled manner as it becomes symmetric with minimum value at 0 when \(L=0.5\).
In our initial experiments, we found that large magnitudes for the translation parameters of the STN can lead to its output patch going beyond the input image’s boundaries (resulting in a black patch with all 0-valued pixels). One reason for this is because for any similar pair \((I_1,I_2){\in }E\), its ranking loss can be minimized when the STN produces identical patches for both \(I_1\) and \(I_2\). This makes learning difficult because the resulting gradient direction of the ranking loss with respect to the transformation parameters becomes uninformative, since the same black patch will be produced in all nearby spatial directions. To handle this, we introduce a simple loss that updates the transformation parameters to bring the STN’s output patch back within the image boundaries if it goes outside:
where \(t_x\) and \(t_y\) are horizontal and vertical translation, respectively, s is isotropic scaling, and \(C_x\) and \(C_y\) are the center x and y pixel-coordinates of the input image I, respectively. \(ST_{loss}\) is simply the squared distance of the center coordinates of the output patch from the input image’s center, and is differentiable. The loss increases as \((s \cdot t_x,s \cdot t_y)\) moves farther away from the image center, so the output patch will be forced to move back toward the image. Importantly, we do not apply this loss if the output patch’s center coordinates are within the image’s boundaries (i.e., this loss does not bias the STN to produce regions that are near the image center).
Putting Eqs. 2 and 3 together, our final loss function is:
where N is the total number of training image pairs, i indexes over the training pairs, \(\lambda _1^i\)=1 (\(\lambda _2^i\)=1) if the center coordinates of the STN’s output patch of \(I_1^i\) (\(I_2^i\)) falls outside of the image’s boundaries. We optimize Eq. 4 with backpropagation to learn the entire network’s weights f. Note that the gradient computed for \(ST_{loss}\) is only backpropagated through the STN and does not affect the RN. If both \(\lambda _1^i\)=\(\lambda _2^i\)=0 (the STN’s output patches for both \(I_1^i\) and \(I_2^i\) are within their image boundaries), then the gradient computed for \(Rank_{loss}\) is backpropagated through the entire network (i.e., both the RN and STN are updated).
4 Results
In this section, we analyze our network’s attribute localization and ranking accuracy through both qualitative and quantitative results.
Datasets. LFW-10 [13]: It consists of 10 face attributes (bald-head, dark-hair, eyes-open, good-looking, masculine-looking, mouth-open, smile, visible-teeth, visible-forehead, and young). There are 1000 training images and 1000 test images, with 500 pairs per attribute for both training and testing. We use the same train-test split used in [13].
UT-Zap50K-1 [17]: It consists of 4 shoe attributes (open, pointy, sporty and comfort). There are 50,025 shoe images, and 1388 training and 300 testing pairs per attribute. We use the train-test splits provided by [17].
OSR [18]: It consists of outdoor scene attributes (natural, open, perspective, large-objects, diagonal-plane, and close-depth). There are 2688 images, and we use the same train/test split as in [3, 17].
Implementation Details. We train a separate network for each visual attribute. We initialize the STN and RN weights with those of AlexNet [42] pre-trained on ImageNet classification up through conv5 and fc7, respectively. We first train our network without the global image; i.e., the RN only receives the output of the STN to generate the attribute score v. We then retrain the entire network—in which we use both the STN output and the global image to train the RN—with the STN weights initialized with the initially learned weights. We find that this setup helps the STN localize the attribute more accurately, since initially the STN is forced to find the optimal image region without being able to rely on the global image when predicting the attribute strength.
For training, we use a mini-batch size of 25 image pairs for SGD, and train the network for 400, 200, and 15 epochs for LFW-10, UT-Zap50K, and OSR, respectively. We set the learning rate for the RN and STN to be 0.001 and 0.0001, respectively, and fix momentum to 0.9. Also, we set the relative learning rate for the scale parameter to be one-tenth of that of the translation parameters, as we find scaling to be more sensitive and can transform the image drastically. We initialize the scale to be 1/3 of the image size for LFW-10 and 1/2 for UT-Zap50K and OSR, based on initial qualitative observations. Translation is initialized randomly for all datasets. We use random crops of size \(227 \times 227\) from our \(256 \times 256\) input image during training, and average the scores for 10 crops (4 corners plus center, and same with horizontal flip) during testing.
Baselines. We compare against the state-of-the art method of Xiao and Lee [14], which uses a pipeline system to discover the spatial extent of relative attributes and trains an SVM ranker to rank the images. We report the method’s results obtained by combining the features computed over the global image and discovered patches, which produce the best accuracy.
We also compare against [13], which computes dense SIFT features on keypoints detected using a supervised facial keypoint detector [44] to train an SVM ranker, and [3], which trains an SVM ranker with global image features. For [3], we compare against the results reported using CNN (pre-trained AlexNet pool5) features in [14]. Finally, we compare against the local learning method of [17]. We report its numbers generated using GIST+color-histogram features.
4.1 Qualitative Results of Localized Attributes
We first visualize our attribute localization results. In Fig. 5, we show the results for the face attributes on the LFW-10 test images. Each row corresponds to a face attribute, and the red box in each image indicates the output of the STN. The images in each row are uniformly sampled after sorting them according to the attribute strength predicted by our network. We can see that our network localizes the relevant regions for the various face attributes. For example, it localizes the mouth region for mouth-open, smile, and visible-teeth; the top of the head for bald-head and dark-hair; near the eyes for eyes-open; and the forehead for visible-forehead. For more global attributes like good-looking, masculine-looking, and young, there is no definite answer but our network tends to localize larger portions of the face.
In Fig. 6, we show the results for the shoe attributes on the UT-Zap50K-1 test images. Again, our network localizes the relevant regions for the different shoe attributes. It localizes the heel for comfort; the toe end for pointy; and the top opening for open. Finally, our network is able to produce accurate image rankings using the localized image parts, as shown in both Figs. 5 and 6. For example, we can see the progression of light-to-dark hair for dark-hair, and closed-to-open shoes for open.
Overall, these qualitative results demonstrate that our method is able to produce accurate localizations.
4.2 Quantitative Results for Attribute Ranking
We next evaluate quantitative ranking accuracy. We report the percentage of test image pairs whose relative attribute ranking is predicted correctly.
Table 1 shows the results on LFW-10. First, the baseline method of [3] uses a global representation instead of a local one to model the attributes and produces the lowest accuracy. The state-of-the-art method of Xiao and Lee [14] outperforms the baseline method of [13] because it automatically discovers the relevant regions of an attribute without relying on pretrained keypoint detectors whose detected parts may be irrelevant to the attribute. Still, across all attributes, we improve on average by \(2.25\,\%\) absolute over the method of Xiao and Lee [14], and produce the best results on seven attributes. This shows the benefit of our end-to-end network, which learns the attribute features, localization, and ranker jointly. In contrast, the method of Xiao and Lee [14] optimizes each step independently, which leads to sub-optimal results.
Table 2 shows the results on UT-Zap50K-1. For this dataset, our improvement over the baselines is marginal because the shoe images are so well-aligned, centered, and have clear backgrounds. Consequently, all baselines obtain very high accuracy and the results are nearly saturated. Thus, following [14], we perform a cross-dataset experiment on the more challenging Shoes-with-Attribute [5] dataset, whose shoe images are not as well-aligned and have more variation in style and scale. Shoes-with-Attribute has three overlapping attributes with UT-Zap50K-1 (open, pointy, and sporty) with 140 annotated image pairs per attribute. We take our models trained on UT-Zap50K-1, and test on Shoes-with-Attribute in order to evaluate cross-dataset generalization ability. Table 3 shows the results. We get a significant boost of \(4.88\,\%\) absolute over the method of Xiao and Lee [14]. This demonstrates that our joint training of the attribute features, localization, and ranker lead to more robustness to appearance and scale variations, compared to training them independently.
Finally, Table 4 shows the results on OSR, which contains very different looking images for the same relative attribute (e.g., forest vs. city images for natural). We compare against the previous state-of-the-art local learning method of Yu and Grauman [17], the relative forest method of [45], and global image-based method of [3]. We obtain an overall accuracy of 97.02 %, which is better than all baselines. More importantly, this result shows our advantage over the method of Xiao and Lee [14], which is not applicable to this dataset, since it requires a visually consistent concept (e.g., the same object) for an attribute to build its visual chains. In contrast, our method can handle drastically different visual concepts for the same attribute (e.g., forest image with trees being more natural than city image with buildings) since the STN only needs to localize the most relevant region in each image without requiring that the concept be present in other images.
4.3 Ablation Study
We study the contribution that the global image versus the output region of the STN has on ranking performance. For this, we train and compare two baseline networks: (1) the RN is trained with only the global image as input (Global image), i.e., we do not have the STN as part of the network; and (2) the RN is trained with only the STN output region as input without the global image (STN output). For both baseline networks, the final linear layer uses the newly-learned 4096-D feature output of the RN to generate the attribute strength v.
Table 5 shows quantitative ranking accuracy of these baseline networks on LFW-10. Overall, the STN output baseline outperforms the Global image baseline. It performs especially well for attributes that are conditioned on small facial parts like eyes-open, mouth-open, and visible-teeth. This is mainly because the Global image baseline needs to process the entire input image, which means that small object parts like the eyes or mouth have very low resolution, and thus, cannot receive the fine-grained attention that they need. In contrast, the STN can attend to those small informative parts, and provide a high-resolution (cropped-out) image for the RN to learn the ranking function. Finally, the third row in Table 5 shows the result of our full model, i.e., combining both the global input image and STN output to train the RN. Our full model produces the best accuracy for eight out of the 10 attributes, which shows that the global contextual information from the input image and the fine-grained information from the localized STN output are complementary. The boost is especially significant for mid-sized and global attributes like dark-hair, visible-forehead, and young, likely because they require both fine-grained information of specific parts as well as more global contextual information of the entire face.
4.4 Computational Speed Analysis
Finally, we analyze the computational speed of our network. Our approach is significantly faster than the related baseline method of Xiao and Lee [14]. Since that method has to process a sequence of time-consuming modules including feature extraction, nearest neighbor matching, iterative SVM classifier training to build the visual chains, and training an SVM ranker to rank the chains, it takes \(\sim \)10 h to train one attribute model on LFW-10 using a cluster of 20 CPU nodes with 2 cores each. In contrast, our end-to-end network only takes \(\sim \)3 h to train one attribute model using a single Titan X GPU. For testing, our approach takes only 0.011 s per image compared to 1.1 s per image for [14].
5 Discussion
We presented a novel end-to-end deep network that combines a localization module with a ranking module to jointly localize and rank relative attributes. Our qualitative results showed our network’s ability to accurately localize the meaningful image patches corresponding to an attribute. We demonstrated state-of-the-art attribute ranking performance on benchmark face, shoe, and outdoor scene datasets. One limitation of our approach is that it can only localize one image part for a visual attribute. However, for certain attributes there can be multiple relevant parts. We would like to explore this issue further either by having multiple spatial transformers or directly predicting pixel-level relevance.
Notes
- 1.
More complex transformations (e.g., affine, thin plate spline) are possible, but we find this transformation to be sufficient for our datasets.
References
Palatucci, M., Pomerleau, D., Hinton, G., Mitchell, T.: Zero-shot learning with semantic output codes. In: NIPS (2009)
Lampert, C.H., Nickisch, H., Harmeling, S.: Learning to detect unseen object classes by between-class attribute transfer. In: CVPR (2009)
Parikh, D., Grauman, K.: Relative attributes. In: ICCV (2011)
Siddiquie, B., Feris, R.S., Davis, L.S.: Image ranking and retrieval based on multi-attribute queries. In: CVPR (2011)
Kovashka, A., Parikh, D., Grauman, K.: Whittlesearch: image search with relative attribute feedback. In: CVPR (2012)
Kovashka, A., Grauman, K.: Attribute adaptation for personalized image search. In: ICCV (2013)
Kumar, N., Berg, A.C., Belhumeur, P.N., Nayar, S.K.: Attribute and simile classifiers for face verification. In: ICCV (2009)
Duan, K., Parikh, D., Crandall, D., Grauman, K.: Discovering localized attributes for fine-grained recognition. In: CVPR (2012)
Zhang, N., Paluri, M., Ranzato, M., Darrell, T., Bourdev, L.: PANDA: pose aligned networks for deep attribute modeling. In: CVPR (2014)
Branson, S., Wah, C., Schroff, F., Babenko, B., Welinder, P., Perona, P., Belongie, S.: Visual recognition with humans in the loop. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part IV. LNCS, vol. 6314, pp. 438–451. Springer, Heidelberg (2010)
Bourdev, L., Maji, S., Malik, J.: Describing people: poselet-based approach to attribute classification. In: ICCV (2011)
Shrivastava, A., Singh, S., Gupta, A.: Constrained semi-supervised learning using attributes and comparative attributes. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012, Part III. LNCS, vol. 7574, pp. 369–383. Springer, Heidelberg (2012)
Sandeep, R.N., Verma, Y., Jawahar, C.V.: Relative parts: distinctive parts for learning relative attributes. In: CVPR (2014)
Xiao, F., Lee, Y.J.: Discovering the spatial extent of relative attributes. In: ICCV (2015)
Chopra, S., Hadsell, R., LeCun, Y.: Learning a similarity metric discriminatively, with application to face verification. In: CVPR (2005)
Jaderberg, M., Simonyan, K., Zisserman, A., Kavukcuoglu, K.: Spatial transformer networks. In: NIPS (2015)
Yu, A., Grauman, K.: Fine-grained visual comparisons with local learning. In: CVPR (2014)
Oliva, A., Torralba, A.: Modeling the shape of the scene: a holistic representation of the spatial envelope. In: IJCV (2001)
Farhadi, A., Endres, I., Hoiem, D., Forsyth, D.: Describing objects by their attributes. In: CVPR (2009)
Kumar, N., Berg, A., Belhumeur, P., Nayar, S.: Attribute and simile classifiers for face verification. In: ICCV (2009)
Farhadi, A., Endres, I., Hoiem, D.: Attribute-centric recognition for cross-category generalization. In: CVPR (2010)
Rastegari, M., Farhadi, A., Forsyth, D.: Attribute discovery via predictable discriminative binary codes. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012, Part VI. LNCS, vol. 7577, pp. 876–889. Springer, Heidelberg (2012)
Saleh, B., Farhadi, A., Elgammal, A.: Object-centric anomaly detection by attribute-based reasoning. In: CVPR (2013)
Chung, J., Lee, D., Seo, Y., Yoo, C.D.: Deep attribute networks. In: NIPS Workshop on Deep Learning and Unsupervised Feature Learning (2012)
Escorcia, V., Niebles, J.C., Ghanem, B.: On the relationship between visual attributes and convolutional networks. In: CVPR (2015)
Shankar, S., Garg, V.K., Cipolla, R.: Deep-carving: Discovering visual attributes by carving deep neural nets. arXiv:1504.04871 (2015)
Souri, Y., Noury, E., Adeli-Mosabbeb, E.: Deep relative attributes. arXiv:1512.04103 (2015)
Kumar, N., Belhumeur, P.N., Nayar, S.K.: FaceTracer: a search engine for large collections of images with faces. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008, Part IV. LNCS, vol. 5305, pp. 340–353. Springer, Heidelberg (2008)
Kiapour, M.H., Yamaguchi, K., Berg, A.C., Berg, T.L.: Hipster wars: discovering elements of fashion styles. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part I. LNCS, vol. 8689, pp. 472–488. Springer, Heidelberg (2014)
Itti, L., Baldi, P.F.: A principled approach to detecting surprising events in video. In: CVPR (2005)
Gao, D., Mahadevan, V., Vasconcelos, N.: The discriminant center-surround hypothesis for bottom-up saliency. In: NIPS (2007)
Alexe, B., Deselaers, T., Ferrari, V.: What is an object? In: CVPR (2010)
Carreira, J., Sminchisescu, C.: Constrained parametric min-cuts for automatic object segmentation. In: CVPR (2010)
Uijlings, J., Sande, K., Gevers, T., Smeulders, A.: Selective search for object recognition. IJCV 104(2), 154–171 (2013)
Zitnick, C.L., Dollár, P.: Edge boxes: locating object proposals from edges. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part V. LNCS, vol. 8693, pp. 391–405. Springer, Heidelberg (2014)
Mnih, V., Heess, N., Graves, A., Kavukcuoglu, K.: Recurrent models of visual attention. In: NIPS (2014)
Xiao, T., Xu, Y., Yang, K., Zhang, J., Peng, Y., Zhang, Z.: The application of two-level attention models in deep convolutional neural network for fine-grained image classification. In: CVPR (2015)
Xu, K., Ba, J., Kiros, R., Courville, A., Salakhutdinov, R., Zemel, R., Bengio, Y.: Show, attend and tell: Neural image caption generation with visual attention. arXiv:1502.03044 (2015)
Gregor, K., Danihelka, I., Graves, A., Wierstra, D.: Draw: A recurrent neural network for image generation. arXiv:1502.04623 (2015)
Tang, Y., Srivastava, N., Salakhutdinov, R.R.: Learning generative models with visual attention. In: NIPS (2014)
Chopra, S., Hadsell, R., LeCun, Y.: Learning a similarity metric discriminatively with application to face verification. In: CVPR (2005)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: NIPS (2012)
Burges, C.J., Shaked, T., Renshaw, E., Lazier, A., Deeds, M., Hamilton, N., Hullender, G.: Learning to rank using gradient descent. In: ICML (2005)
Zhu, X., Ramanan, D.: Face detection, pose estimation, and landmark localization in the wild. In: CVPR (2012)
Li, S., Shan, S., Chen, X.: Relative forest for attribute prediction. In: Lee, K.M., Matsushita, Y., Rehg, J.M., Hu, Z. (eds.) ACCV 2012, Part I. LNCS, vol. 7724, pp. 316–327. Springer, Heidelberg (2013)
Acknowledgements
This work was supported in part by an Amazon Web Services Education Research Grant and GPUs donated by NVIDIA.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Singh, K.K., Lee, Y.J. (2016). End-to-End Localization and Ranking for Relative Attributes. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds) Computer Vision – ECCV 2016. ECCV 2016. Lecture Notes in Computer Science(), vol 9910. Springer, Cham. https://doi.org/10.1007/978-3-319-46466-4_45
Download citation
DOI: https://doi.org/10.1007/978-3-319-46466-4_45
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46465-7
Online ISBN: 978-3-319-46466-4
eBook Packages: Computer ScienceComputer Science (R0)