Image annotation using metric learning in semantic neighbourhoods
Citations
2 citations
2 citations
2 citations
Cites background or methods from "Image annotation using metric learn..."
...Another important difference between our method and existing methods [13, 14, 52] is the number of nearest-neighbors used to propagate the tags....
[...]
...Verma and Jawahar [52] presented two-pass kNN to find neighbors in semantic neighborhoods besides metric learning which learns weights for combining different features....
[...]
1 citations
Cites background or methods from "Image annotation using metric learn..."
...For IAPR TC-12 dataset, only 2PKNN-ML [107] performs slightly better than our system while being highly imbalanced in terms of precision and recall....
[...]
...Annotation systems like [27], [34], [59], [77], [107] identify ‘nearest neighbors’ of the test image from which to propagate the labels to the test image....
[...]
...Therefore, we used the same datasets and evaluation measures as used by the previously proposed images annotation systems designed to work under these condition such as [29], [30], [34], [52], [97], [107]....
[...]
...Previously, semantic information has been quantified and incorporated in image annotation systems in different ways [24], [37], [107]....
[...]
References
4,433 citations
4,157 citations
"Image annotation using metric learn..." refers background or methods in this paper
...With this goal, we perform metric learning over 2PKNN by generalizing the LMNN [11] algorithm for multi-label prediction....
[...]
...In such a scenario, (i) since each base distance contributes differently, we can learn appropriate weights to combine them in the distance space [2, 3]; and (ii) since every feature (such as SIFT or colour histogram) itself is represented as a multidimensional vector, its individual elements can also be weighted in the feature space [11]....
[...]
...Our extension of LMNN conceptually differs from its previous extensions such as [21] in at least two significant ways: (i) we adapt LMNN in its choice of target/impostors to learn metrics for multi-label prediction problems, whereas [21] uses the same definition of target/impostors as in LMNN to address classification problem in multi-task setting, and (ii) in our formulation, the amount of push applied on an impostor varies depending on its conceptual similarity w.r.t. a given sample, which makes it suitable for multi-label prediction tasks....
[...]
...Our metric learning framework extends LMNN in two major ways: (i) LMNN is meant for single-label classification (or simply classification) problems, while we adapt it for images annotation which is a multi-label classification task; and (ii) LMNN learns a single Mahalanobis metric in the feature space, while we extend it to learn linear metrics for multi- Image Annotation Using Metric Learning in Semantic Neighbourhoods 3 ple features as well as distances together....
[...]
...For this purpose, we extend the classical LMNN [11] algorithm for multi-label prediction....
[...]
2,365 citations
"Image annotation using metric learn..." refers background in this paper
...ESP Game contains images annotated using an on-line game, where two (mutually unknown) players are randomly given an image for which they have to predict same keyword(s) to score points [22]....
[...]
2,037 citations
"Image annotation using metric learn..." refers methods in this paper
...To overcome this issue, we solve it by alternatively using stochastic sub-gradient descent and projection steps (similar to Pegasos [12])....
[...]
...To address this, we implement metric learning by alternating between stochastic sub-gradient descent and projection steps (similar to Pegasos [12])....
[...]
1,765 citations
"Image annotation using metric learn..." refers background in this paper
...translation models [13, 14] and nearest-neighbour based relevance models [1, 8]....
[...]
...Corel 5K was first used in [14], and since then it has become a benchmark for comparing annotation performance....
[...]