scispace - formally typeset
Search or ask a question

Showing papers on "Pairwise comparison published in 2014"


Journal ArticleDOI
26 Sep 2014-PLOS ONE
TL;DR: Sequence Demarcation Tool (SDT) as discussed by the authors is a free user-friendly computer program that aims to provide a robust and highly reproducible means of objectively using pairwise genetic identity calculations to classify any set of nucleotide or amino acid sequences.
Abstract: The perpetually increasing rate at which viral full-genome sequences are being determined is creating a pressing demand for computational tools that will aid the objective classification of these genome sequences. Taxonomic classification approaches that are based on pairwise genetic identity measures are potentially highly automatable and are progressively gaining favour with the International Committee on Taxonomy of Viruses (ICTV). There are, however, various issues with the calculation of such measures that could potentially undermine the accuracy and consistency with which they can be applied to virus classification. Firstly, pairwise sequence identities computed based on multiple sequence alignments rather than on multiple independent pairwise alignments can lead to the deflation of identity scores with increasing dataset sizes. Also, when gap-characters need to be introduced during sequence alignments to account for insertions and deletions, methodological variations in the way that these characters are introduced and handled during pairwise genetic identity calculations can cause high degrees of inconsistency in the way that different methods classify the same sets of sequences. Here we present Sequence Demarcation Tool (SDT), a free user-friendly computer program that aims to provide a robust and highly reproducible means of objectively using pairwise genetic identity calculations to classify any set of nucleotide or amino acid sequences. SDT can produce publication quality pairwise identity plots and colour-coded distance matrices to further aid the classification of sequences according to ICTV approved taxonomic demarcation criteria. Besides a graphical interface version of the program for Windows computers, command-line versions of the program are available for a variety of different operating systems (including a parallel version for cluster computing platforms).

1,068 citations


Journal ArticleDOI
03 Jul 2014-PLOS ONE
TL;DR: An approach to determining confidence in the output of a network meta-analysis is proposed based on methodology developed by the Grading of Recommendations Assessment, Development and Evaluation Working Group for pairwise meta-analyses and applied to a systematic review comparing topical antibiotics without steroids for chronically discharging ears with underlying eardrum perforations.
Abstract: Systematic reviews that collate data about the relative effects of multiple interventions via network meta-analysis are highly informative for decision-making purposes. A network meta-analysis provides two types of findings for a specific outcome: the relative treatment effect for all pairwise comparisons, and a ranking of the treatments. It is important to consider the confidence with which these two types of results can enable clinicians, policy makers and patients to make informed decisions. We propose an approach to determining confidence in the output of a network meta-analysis. Our proposed approach is based on methodology developed by the Grading of Recommendations Assessment, Development and Evaluation (GRADE) Working Group for pairwise meta-analyses. The suggested framework for evaluating a network meta-analysis acknowledges (i) the key role of indirect comparisons (ii) the contributions of each piece of direct evidence to the network meta-analysis estimates of effect size; (iii) the importance of the transitivity assumption to the validity of network meta-analysis; and (iv) the possibility of disagreement between direct evidence and indirect evidence. We apply our proposed strategy to a systematic review comparing topical antibiotics without steroids for chronically discharging ears with underlying eardrum perforations. The proposed framework can be used to determine confidence in the results from a network meta-analysis. Judgements about evidence from a network meta-analysis can be different from those made about evidence from pairwise meta-analyses.

853 citations


Posted Content
TL;DR: This work specifies a graphical model for human pose which exploits the fact the local image measurements can be used both to detect parts (or joints) and also to predict the spatial relationships between them (Image Dependent Pairwise Relations).
Abstract: We present a method for estimating articulated human pose from a single static image based on a graphical model with novel pairwise relations that make adaptive use of local image measurements. More precisely, we specify a graphical model for human pose which exploits the fact the local image measurements can be used both to detect parts (or joints) and also to predict the spatial relationships between them (Image Dependent Pairwise Relations). These spatial relationships are represented by a mixture model. We use Deep Convolutional Neural Networks (DCNNs) to learn conditional probabilities for the presence of parts and their spatial relationships within image patches. Hence our model combines the representational flexibility of graphical models with the efficiency and statistical power of DCNNs. Our method significantly outperforms the state of the art methods on the LSP and FLIC datasets and also performs very well on the Buffy dataset without any training.

474 citations


Journal ArticleDOI
TL;DR: This work integrates a Random Forest classifier into a Conditional Random Field framework, a flexible approach for obtaining a reliable classification result even in complex urban scenes, and investigates the relevance of different features for the LiDAR points as well as for the interaction of neighbouring points.
Abstract: In this work we address the task of the contextual classification of an airborne LiDAR point cloud. For that purpose, we integrate a Random Forest classifier into a Conditional Random Field (CRF) framework. It is a flexible approach for obtaining a reliable classification result even in complex urban scenes. In this way, we benefit from the consideration of context on the one hand and from the opportunity to use a large amount of features on the other hand. Considering the interactions in our experiments increases the overall accuracy by 2%, though a larger improvement becomes apparent in the completeness and correctness of some of the seven classes discerned in our experiments. We compare the Random Forest approach to linear models for the computation of unary and pairwise potentials of the CRF, and investigate the relevance of different features for the LiDAR points as well as for the interaction of neighbouring points. In a second step, building objects are detected based on the classified point cloud. For that purpose, the CRF probabilities for the classes are plugged into a Markov Random Field as unary potentials, in which the pairwise potentials are based on a Potts model. The 2D binary building object masks are extracted and evaluated by the benchmark ISPRS Test Project on Urban Classification and 3D Building Reconstruction. The evaluation shows that the main buildings (larger than 50 m 2 ) can be detected very reliably with a correctness larger than 96% and a completeness of 100%.

455 citations


Journal ArticleDOI
TL;DR: Based on a new effective and feasible representation of uncertain information, called D numbers, a D-AHP method is proposed for the supplier selection problem, which extends the classical analytic hierarchy process (AHP) method.
Abstract: Supplier selection is an important issue in supply chain management (SCM), and essentially is a multi-criteria decision-making problem. Supplier selection highly depends on experts' assessments. In the process of that, it inevitably involves various types of uncertainty such as imprecision, fuzziness and incompleteness due to the inability of human being's subjective judgment. However, the existing methods cannot adequately handle these types of uncertainties. In this paper, based on a new effective and feasible representation of uncertain information, called D numbers, a D-AHP method is proposed for the supplier selection problem, which extends the classical analytic hierarchy process (AHP) method. Within the proposed method, D numbers extended fuzzy preference relation has been involved to represent the decision matrix of pairwise comparisons given by experts. An illustrative example is presented to demonstrate the effectiveness of the proposed method.

419 citations


Proceedings ArticleDOI
24 Feb 2014
TL;DR: The experiments indicate that the proposed adaptive sampler improves the state-of-the art learning algorithm largely in convergence without negative effects on prediction quality or iteration runtime.
Abstract: Pairwise algorithms are popular for learning recommender systems from implicit feedback. For each user, or more generally context, they try to discriminate between a small set of selected items and the large set of remaining (irrelevant) items. Learning is typically based on stochastic gradient descent (SGD) with uniformly drawn pairs. In this work, we show that convergence of such SGD learning algorithms slows down considerably if the item popularity has a tailed distribution. We propose a non-uniform item sampler to overcome this problem. The proposed sampler is context-dependent and oversamples informative pairs to speed up convergence. An efficient implementation with constant amortized runtime costs is developed. Furthermore, it is shown how the proposed learning algorithm can be applied to a large class of recommender models. The properties of the new learning algorithm are studied empirically on two real-world recommender system problems. The experiments indicate that the proposed adaptive sampler improves the state-of-the art learning algorithm largely in convergence without negative effects on prediction quality or iteration runtime.

352 citations


Journal ArticleDOI
TL;DR: This work formally introduces a Pairwise Transform Invariance (PTI) principle, and proposes a novel Pairwise Rotation Invariant Co-occurrence Local Binary Pattern (PRICoLBP) feature, and extends it to incorporate multi-scale, multi-orientation, and multi-channel information.
Abstract: Designing effective features is a fundamental problem in computer vision However, it is usually difficult to achieve a great tradeoff between discriminative power and robustness Previous works shown that spatial co-occurrence can boost the discriminative power of features However the current existing co-occurrence features are taking few considerations to the robustness and hence suffering from sensitivity to geometric and photometric variations In this work, we study the Transform Invariance (TI) of co-occurrence features Concretely we formally introduce a Pairwise Transform Invariance (PTI) principle, and then propose a novel Pairwise Rotation Invariant Co-occurrence Local Binary Pattern (PRICoLBP) feature, and further extend it to incorporate multi-scale, multi-orientation, and multi-channel information Different from other LBP variants, PRICoLBP can not only capture the spatial context co-occurrence information effectively, but also possess rotation invariance We evaluate PRICoLBP comprehensively on nine benchmark data sets from five different perspectives, eg, encoding strategy, rotation invariance, the number of templates, speed, and discriminative power compared to other LBP variants Furthermore we apply PRICoLBP to six different but related applications-texture, material, flower, leaf, food, and scene classification, and demonstrate that PRICoLBP is efficient, effective, and of a well-balanced tradeoff between the discriminative power and robustness

289 citations


Proceedings Article
08 Dec 2014
TL;DR: In this paper, a graphical model for human pose estimation from a single static image is proposed, which exploits the fact the local image measurements can be used both to detect parts and also to predict the spatial relationships between them.
Abstract: We present a method for estimating articulated human pose from a single static image based on a graphical model with novel pairwise relations that make adaptive use of local image measurements. More precisely, we specify a graphical model for human pose which exploits the fact the local image measurements can be used both to detect parts (or joints) and also to predict the spatial relationships between them (Image Dependent Pairwise Relations). These spatial relationships are represented by a mixture model. We use Deep Convolutional Neural Networks (DCNNs) to learn conditional probabilities for the presence of parts and their spatial relationships within image patches. Hence our model combines the representational flexibility of graphical models with the efficiency and statistical power of DCNNs. Our method significantly outperforms the state of the art methods on the LSP and FLIC datasets and also performs very well on the Buffy dataset without any training.

228 citations


Journal ArticleDOI
TL;DR: This paper proposes a global and local structure preservation framework for feature selection (GLSPFS) which integrates both global pairwise sample similarity and local geometric data structure to conduct feature selection and shows that the best feature selection performance is always obtained when the two factors are appropriately integrated.
Abstract: The recent literature indicates that preserving global pairwise sample similarity is of great importance for feature selection and that many existing selection criteria essentially work in this way. In this paper, we argue that besides global pairwise sample similarity, the local geometric structure of data is also critical and that these two factors play different roles in different learning scenarios. In order to show this, we propose a global and local structure preservation framework for feature selection (GLSPFS) which integrates both global pairwise sample similarity and local geometric data structure to conduct feature selection. To demonstrate the generality of our framework, we employ methods that are well known in the literature to model the local geometric data structure and develop three specific GLSPFS-based feature selection algorithms. Also, we develop an efficient optimization algorithm with proven global convergence to solve the resulting feature selection problem. A comprehensive experimental study is then conducted in order to compare our feature selection algorithms with many state-of-the-art ones in supervised, unsupervised, and semisupervised learning scenarios. The result indicates that: 1) our framework consistently achieves statistically significant improvement in selection performance when compared with the currently used algorithms; 2) in supervised and semisupervised learning scenarios, preserving global pairwise similarity is more important than preserving local geometric data structure; 3) in the unsupervised scenario, preserving local geometric data structure becomes clearly more important; and 4) the best feature selection performance is always obtained when the two factors are appropriately integrated. In summary, this paper not only validates the advantages of the proposed GLSPFS framework but also gains more insight into the information to be preserved in different feature selection tasks.

209 citations


Journal ArticleDOI
TL;DR: This study proposes the concept of the information granularity being regarded as an important and useful asset supporting the goal to reach consensus in group decision making by using fuzzy preference relations to represent the opinions of the decision makers.

207 citations


Journal ArticleDOI
TL;DR: In this article, a pairwise censored likelihood is used for consistent estimation of the extremes of space-time data under mild mixing conditions, and illustrates this by fitting an extension of a model of Schlather (2002) to hourly rainfall data.
Abstract: Max-stable processes are the natural analogues of the generalized extreme-value distribution when modelling extreme events in space and time. Under suitable conditions, these processes are asymptotically justified models for maxima of independent replications of random fields, and they are also suitable for the modelling of extreme measurements over high thresholds. This paper shows how a pairwise censored likelihood can be used for consistent estimation of the extremes of space-time data under mild mixing conditions, and illustrates this by fitting an extension of a model of Schlather (2002) to hourly rainfall data. A block bootstrap procedure is used for uncertainty assessment. Estimator efficiency is considered and the choice of pairs to be included in the pairwise likelihood is discussed. The proposed model fits the data better than some natural competitors.

Book ChapterDOI
06 Sep 2014
TL;DR: A network consistent re-identification (NCR) framework is proposed, which is formulated as an optimization problem that not only maintains consistency in re- identification results across the network, but also improves the camera pairwise re-Identification performance between all the individual camera pairs.
Abstract: Most existing person re-identification methods focus on finding similarities between persons between pairs of cameras (camera pairwise re-identification) without explicitly maintaining consistency of the results across the network. This may lead to infeasible associations when results from different camera pairs are combined. In this paper, we propose a network consistent re-identification (NCR) framework, which is formulated as an optimization problem that not only maintains consistency in re-identification results across the network, but also improves the camera pairwise re-identification performance between all the individual camera pairs. This can be solved as a binary integer programing problem, leading to a globally optimal solution. We also extend the proposed approach to the more general case where all persons may not be present in every camera. Using two benchmark datasets, we validate our approach and compare against state-of-the-art methods.

Journal ArticleDOI
TL;DR: A maximum eigenvalue threshold is proposed as the consistency index for the ANP in risk assessment and decision analysis and a block diagonal matrix is introduced for the RPCM to conduct consistency tests simultaneously for all comparison matrices.

Proceedings Article
21 Jun 2014
TL;DR: This paper shows that, under a 'time-reversibility' or Bradley-Terry-Luce (BTL) condition on the distribution, the rank centrality (PageRank) and least squares (HodgeRank) algorithms both converge to an optimal ranking.
Abstract: There has been much interest recently in the problem of rank aggregation from pairwise data. A natural question that arises is: under what sorts of statistical assumptions do various rank aggregation algorithms converge to an 'optimal' ranking? In this paper, we consider this question in a natural setting where pairwise comparisons are drawn randomly and independently from some underlying probability distribution. We first show that, under a 'time-reversibility' or Bradley-Terry-Luce (BTL) condition on the distribution, the rank centrality (PageRank) and least squares (HodgeRank) algorithms both converge to an optimal ranking. Next, we show that a matrix version of the Borda count algorithm, and more surprisingly, an algorithm which performs maximum likelihood estimation under a BTL assumption, both converge to an optimal ranking under a 'low-noise' condition that is strictly more general than BTL. Finally, we propose a new SVM-based algorithm for rank aggregation from pairwise data, and show that this converges to an optimal ranking under an even more general condition that we term 'generalized low-noise'. In all cases, we provide explicit sample complexity bounds for exact recovery of an optimal ranking. Our experiments confirm our theoretical findings and help to shed light on the statistical behavior of various rank aggregation algorithms.

Journal ArticleDOI
TL;DR: A mathematical proof that both approaches to network meta-analysis lead to identical estimates is given, with one additional parameter for a common heterogeneity variance, applied to a systematic review in depression.
Abstract: Network meta-analysis is a statistical method combining information from randomised trials that compare two or more treatments for a given medical condition. Consistent treatment effects are estimated for all possible treatment comparisons. For estimation, weighted least squares regression that in a natural way generalises standard pairwise meta-analysis can be used. Typically, as part of the network, multi-arm studies are found. In a multi-arm study, observed pairwise comparisons are correlated, which must be accounted for. To this aim, two methods have been proposed, a standard regression approach and a new approach coming from graph theory and based on contrast-based data (Rucker 2012). In the standard approach, the dimension of the design matrix is appropriately reduced until it is invertible ('reduce dimension'). In the alternative approach, the weights of comparisons coming from multi-arm studies are appropriately reduced ('reduce weights'). As it was unclear, to date, how these approaches are related to each other, we give a mathematical proof that both approaches lead to identical estimates. The 'reduce weights' approach can be interpreted as the construction of a network of independent two-arm studies, which is basically equivalent to the given network with multi-arm studies. Thus, a simple random-effects model is obtained, with one additional parameter for a common heterogeneity variance. This is applied to a systematic review in depression.

Journal ArticleDOI
TL;DR: A fuzzy AHP variant is proposed, wherein pairwise comparison of decision elements by domain experts is expressed with triangular fuzzy numbers, which allows the degree of confidence of the expert to be quantified explicitly and inconsistencies in judgment to be reconciled within the bounds of the fuzzy numbers to generate reasonable values for the weighting factors.

Journal ArticleDOI
TL;DR: The CRS4EAs was empirically compared to NHST within a computational experiment conducted on 16 evolutionary algorithms and a benchmark suite of 20 numerical minimisation problems, and the analysis of the results shows that the CRS3EAs is comparable with NHST but may also have many additional benefits.

Journal ArticleDOI
TL;DR: This paper studies the active learning problem of selecting pairwise must-link and cannot-link constraints for semi-supervised clustering and introduces a selection criterion that trades off the amount of uncertainty of each data point with the expected number of queries (the cost) required to resolve this uncertainty.
Abstract: Semi-supervised clustering aims to improve clustering performance by considering user supervision in the form of pairwise constraints. In this paper, we study the active learning problem of selecting pairwise must-link and cannot-link constraints for semi-supervised clustering. We consider active learning in an iterative manner where in each iteration queries are selected based on the current clustering solution and the existing constraint set. We apply a general framework that builds on the concept of neighborhood, where neighborhoods contain "labeled examples" of different clusters according to the pairwise constraints. Our active learning method expands the neighborhoods by selecting informative points and querying their relationship with the neighborhoods. Under this framework, we build on the classic uncertainty-based principle and present a novel approach for computing the uncertainty associated with each data point. We further introduce a selection criterion that trades off the amount of uncertainty of each data point with the expected number of queries (the cost) required to resolve this uncertainty. This allows us to select queries that have the highest information rate. We evaluate the proposed method on the benchmark data sets and the results demonstrate consistent and substantial improvements over the current state of the art.

Journal ArticleDOI
TL;DR: One of the aims of the paper is to provide ranking of the selected players, however, the analysis of incomplete pairwise comparison matrices is also in the focus, and the eigenvector method and the logarithmic least squares method were used.

Journal ArticleDOI
TL;DR: It is concluded that psychological models of choice should be based on single-attribute pairwise comparisons made in each choice, with a pair of alternatives compared on a single attribute dimension in each comparison.

Journal ArticleDOI
TL;DR: The current paper integrates the AHP with stochastic multicriteria acceptability analysis (SMAA), an inverse-preference method, to allow the pairwise comparisons to be uncertain, to assess how the consistency of judgements and the ability of the SMAA-AHP model to discern the best alternative deteriorates as uncertainty increases.

Journal Article
TL;DR: This work develops the first algorithms for learning Mallows models (and mixtures thereof) from pairwise comparison data, and develops approximate samplers that are exact for many important special cases and have provable bounds with pairwise evidence.
Abstract: Learning preference distributions is a critical problem in many areas (e.g., recommender systems, IR, social choice). However, many existing learning and inference methods impose restrictive assumptions on the form of user preferences that can be admitted as evidence. We relax these restrictions by considering as data arbitrary pairwise comparisons of alternatives, which represent the fundamental building blocks of ordinal rankings. We develop the first algorithms for learning Mallows models (and mixtures thereof) from pairwise comparison data. At the heart of our technique is a new algorithm, the generalized repeated insertion model (GRIM), which allows sampling from arbitrary ranking distributions, and conditional Mallows models in particular. While we show that sampling from a Mallows model with pairwise evidence is computationally difficult in general, we develop approximate samplers that are exact for many important special cases|and have provable bounds with pairwise evidence--and derive algorithms for evaluating log-likelihood, learning Mallows mixtures, and non-parametric estimation. Experiments on real-world data sets demonstrate the effectiveness of our approach.

Journal ArticleDOI
TL;DR: The pairwise approach does not analyse matrix-level structure and thus views a species pair as the fundamental unit of co-occurrence, and may make this task easier by simplifying the analysis and resulting inferences to associations between paired species.
Abstract: The analysis of species co-occurrence patterns continues to be a main pursuit of ecologists, primarily because the coexistence of species is fundamentally important in evaluating various theories, principles and concepts. Examples include community assembly, equilibrium versus non-equilibrium organization of communities, resource partitioning and ecological character displacement, the local–regional species diversity relationship, and the metacommunity concept. Traditionally, co-occurrence has been measured and tested at the level of an entire species presence–absence matrix wherein various algorithms are used to randomize matrices and produce statistical null distributions of metrics that quantify structure in the matrix. This approach implicitly recognizes a presence–absence matrix as having some real ecological identity (e.g. a set of species exhibiting nestedness among a set of islands) in addition to being a unit of statistical analysis. An emerging alternative is to test for non-random co-occurrence between paired species. The pairwise approach does not analyse matrix-level structure and thus views a species pair as the fundamental unit of co-occurrence. Inferring process from pattern is very difficult in analyses of co-occurrence; however, the pairwise approach may make this task easier by simplifying the analysis and resulting inferences to associations between paired species.

Proceedings ArticleDOI
03 Nov 2014
TL;DR: Experimental results are presented that demonstrate that both team draft multileave and optimized multileaved can accurately determine all pairwise preferences among a set of rankers using far less data than the interleaving methods that they extend.
Abstract: Evaluation methods for information retrieval systems come in three types: offline evaluation, using static data sets annotated for relevance by human judges; user studies, usually conducted in a lab-based setting; and online evaluation, using implicit signals such as clicks from actual users. For the latter, preferences between rankers are typically inferred from implicit signals via interleaved comparison methods, which combine a pair of rankings and display the result to the user. We propose a new approach to online evaluation called multileaved comparisons that is useful in the prevalent case where designers are interested in the relative performance of more than two rankers. Rather than combining only a pair of rankings, multileaved comparisons combine an arbitrary number of rankings. The resulting user clicks then give feedback about how all these rankings compare to each other. We propose two specific multileaved comparison methods. The first, called team draft multileave, is an extension of team draft interleave. The second, called optimized multileave, is an extension of optimized interleave and is designed to handle cases where a large number of rankers must be multileaved. We present experimental results that demonstrate that both team draft multileave and optimized multileave can accurately determine all pairwise preferences among a set of rankers using far less data than the interleaving methods that they extend.

Proceedings Article
08 Dec 2014
TL;DR: It is shown that even if one applies the mismatched maximum likelihood estimator that assumes independence (on pairwise comparisons that are now dependent due to rank-breaking), minimax optimal performance is still achieved up to a logarithmic factor.
Abstract: This paper studies the problem of rank aggregation under the Plackett-Luce model. The goal is to infer a global ranking and related scores of the items, based on partial rankings provided by multiple users over multiple subsets of items. A question of particular interest is how to optimally assign items to users for ranking and how many item assignments are needed to achieve a target estimation error. Without any assumptions on how the items are assigned to users, we derive an oracle lower bound and the Cramer-Rao lower bound of the estimation error. We prove an upper bound on the estimation error achieved by the maximum likelihood estimator, and show that both the upper bound and the Cramer-Rao lower bound inversely depend on the spectral gap of the Laplacian of an appropriately defined comparison graph. Since random comparison graphs are known to have large spectral gaps, this suggests the use of random assignments when we have the control. Precisely, the matching oracle lower bound and the upper bound on the estimation error imply that the maximum likelihood estimator together with a random assignment is minimax-optimal up to a logarithmic factor. We further analyze a popular rank-breaking scheme that decompose partial rankings into pairwise comparisons. We show that even if one applies the mismatched maximum likelihood estimator that assumes independence (on pairwise comparisons that are now dependent due to rank-breaking), minimax optimal performance is still achieved up to a logarithmic factor.

Proceedings ArticleDOI
06 Oct 2014
TL;DR: A novel Gradient Boosting Factorization Machine (GBFM) model is proposed to incorporate feature selection algorithm with Factorization Machines into a unified framework and the efficiency and effectiveness of the algorithm compared to other state-of-the-art methods are demonstrated.
Abstract: Recommendation techniques have been well developed in the past decades. Most of them build models only based on user item rating matrix. However, in real world, there is plenty of auxiliary information available in recommendation systems. We can utilize these information as additional features to improve recommendation performance. We refer to recommendation with auxiliary information as context-aware recommendation. Context-aware Factorization Machines (FM) is one of the most successful context-aware recommendation models. FM models pairwise interactions between all features, in such way, a certain feature latent vector is shared to compute the factorized parameters it involved. In practice, there are tens of context features and not all the pairwise feature interactions are useful. Thus, one important challenge for context-aware recommendation is how to effectively select "good" interaction features. In this paper, we focus on solving this problem and propose a greedy interaction feature selection algorithm based on gradient boosting. Then we propose a novel Gradient Boosting Factorization Machine (GBFM) model to incorporate feature selection algorithm with Factorization Machines into a unified framework. The experimental results on both synthetic and real datasets demonstrate the efficiency and effectiveness of our algorithm compared to other state-of-the-art methods.

Proceedings ArticleDOI
01 Jun 2014
TL;DR: The experimental results show that TrueSkill outperforms other recently proposed models on accuracy, and also can significantly reduce the number of pairwise annotations that need to be collected by sampling non-uniformly from the space of system competitions.
Abstract: A main output of the annual Workshop on Statistical Machine Translation (WMT) is a ranking of the systems that participated in its shared translation tasks, produced by aggregating pairwise sentencelevel comparisons collected from human judges. Over the past few years, there have been a number of tweaks to the aggregation formula in attempts to address issues arising from the inherent ambiguity and subjectivity of the task, as well as weaknesses in the proposed models and the manner of model selection. We continue this line of work by adapting the TrueSkill TM algorithm — an online approach for modeling the relative skills of players in ongoing competitions, such as Microsoft’s Xbox Live — to the human evaluation of machine translation output. Our experimental results show that TrueSkill outperforms other recently proposed models on accuracy, and also can significantly reduce the number of pairwise annotations that need to be collected by sampling non-uniformly from the space of system competitions.

Journal ArticleDOI
TL;DR: This study examines the notion of inconsistency in pairwise comparisons for providing an axiomatization for it and proposes two inconsistency indicators for pairwise comparison.
Abstract: This study examines the notion of inconsistency in pairwise comparisons for providing an axiomatization for it. It also proposes two inconsistency indicators for pairwise comparisons. The primary motivation for the inconsistency reduction is expressed by a computer industry concept “garbage in, garbage out”. The quality of the output depends on the quality of the input.

OtherDOI
29 Sep 2014
TL;DR: The Analytic Hierarchy Process (AHP) as mentioned in this paper is a theory of relative measurement of intangible criteria, where a scale of priorities is derived from pairwise comparison measurements only after the elements to be measured are known.
Abstract: The Analytic Hierarchy Process (AHP) is a theory of relative measurement of intangible criteria. With this approach to relative measurement, a scale of priorities is derived from pairwise comparison measurements only after the elements to be measured are known. The ability to do pairwise comparisons is our biological heritage and we need it to cope with a world where everything is relative and constantly changing and thus, there are no fixed standards to measure things on. In traditional measurement, one has a scale that one applies to measure any element that comes along that has the property the scale is for, and the elements are measured one by one, not by comparing them with each other. In the AHP, paired comparisons are made with judgments using numerical values taken from the AHP absolute fundamental scale of 1 to 9. A scale of relative values is derived from all these paired comparisons and it also belongs to an absolute scale that is invariant under the identity transformation like the system of real numbers. The AHP is useful for making multicriteria decisions involving benefits, opportunities, costs, and risks. The ideas are developed in stages and illustrated with examples of real-life decisions. The subject is transparent and easy to understand why it is done the way it is along the lines discussed here. The AHP has a generalization to dependence and feedback; the Analytic Network Process (ANP) is not discussed here. Keywords: analytic hierarchy process; decision making; prioritization; benefits; costs; complexity

Proceedings ArticleDOI
06 Oct 2014
TL;DR: It is shown how BPR can be extended to deal with more fine-granular, graded preference relations and an empirical analysis shows that this extension can help to measurably increase the predictive accuracy of BPR on realistic e-commerce datasets.
Abstract: In many application domains of recommender systems, explicit rating information is sparse or non-existent. The preferences of the current user have therefore to be approximated by interpreting his or her behavior, i.e., the implicit user feedback. In the literature, a number of algorithm proposals have been made that rely solely on such implicit feedback, among them Bayesian Personalized Ranking (BPR).In the BPR approach, pairwise comparisons between the items are made in the training phase and an item i is considered to be preferred over item j if the user interacted in some form with i but not with j. In real-world applications, however, implicit feedback is not necessarily limited to such binary decisions as there are, e.g., different types of user actions like item views, cart or purchase actions and there can exist several actions for an item over time.In this paper we show how BPR can be extended to deal with such more fine-granular, graded preference relations. An empirical analysis shows that this extension can help to measurably increase the predictive accuracy of BPR on realistic e-commerce datasets.