Potentially all pairwise rankings of all possible alternatives

Potentially all pairwise rankings of all possible alternatives (PAPRIKA) is a method for multicriteria decision making (MCDM) or conjoint analysis based on decisionmakers’ preferences as expressed using pairwise rankings of alternatives.^{[1]}^{[2]}
The PAPRIKA method – implemented via a specific type of decisionmaking software – is used to calculate point values (or ‘weights’) on the criteria or attributes for decision problems involving ranking, prioritising or choosing between alternatives. Point values represent the relative importance of the criteria or attributes to decisionmakers.
As well as representing decisionmakers’ preferences, the point values are used to rank alternatives – enabling decisionmakers to prioritise or choose between them (perhaps subject to a budget constraint). Examples of applications of the PAPRIKA method appear in the next section.
Contents
Applications
Applications of the PAPRIKA method in the area of health care decisionmaking include:

 Prioritising patients for access to elective (nonurgent) surgery in New Zealand^{[3]}^{[4]}
 Referring patients for rheumatology, nephrology, geriatrics and gastroenterology services in Canada^{[5]}^{[6]}^{[7]}
 Classifying individuals by their risks of suffering from rheumatoid arthritis^{[8]}
 Health technology prioritisation^{[9]}
 Measuring patient responses in clinical trials for chronic gout^{[10]}
Applications in other areas include:

 Corporate strategic management^{[11]}
 Revealing central bankers’ monetarypolicy preferences with respect to tradeoffs between inflation and volatility in GDP, interest rates and exchange rates^{[12]}
 Discovering agronomists’ preferred grass breeding traits for pasture plant species in Australia^{[13]}
 Characterising the ‘ideal’ sheep to breed in terms of traits preferred by sheep farmers in Ireland^{[14]}
 Incorporating climate change and mitigation information into environmental resources management for the ocean^{[15]}
Additive multiattribute value models (or ‘points systems’)
As for other methods for MCDM or conjoint analysis, the PAPRIKA method specifically applies to additive multiattribute value models with performance categories^{[16]} – also commonly known as ‘points’, ‘scoring’, ‘pointcount’ or ‘linear’ systems or models.
As the name implies, additive multiattribute value models with performance categories – hereinafter referred to simply as ‘value models’ – consist of multiple criteria (or ‘attributes’), with two or more performance categories (or ‘levels’) within each criterion, that are combined additively. Each category is worth a certain number of points that is intended to reflect both the relative importance (‘weight’) of the criterion and its degree of achievement. For each alternative being considered, the point values are summed across the criteria to get a total score (thus these are additive value models), by which the alternatives are prioritised or ranked (or otherwise classified) relative to each other.
Thus, a value model (or ‘points system’) is simply a schedule of criteria and point values (for an example, see Table 1 in the subsection below) for the decision problem at hand. This representation is equivalent to the more traditional approach involving normalised criterion weights and ‘singlecriterion value functions’ to represent the relative importance of the criteria and to combine values overall. The unweighted (‘points system’) representation is, arguably, easier to use and it simplifies the explication of the PAPRIKA method later below.
The ‘process’ of applying a point system is illustrated next via the example of ranking candidates for a job.
An example
Imagine that ‘Tom’, ‘Dick’ and ‘Harry’ are three job candidates and that they are to be ranked using the value model in Table 1 below. Suppose that after being assessed they are scored on the five criteria (see Table 1) like this:

 Tom’s education is excellent, he has > 5 years of experience, but his references, social skills and enthusiasm are all poor.
 Dick’s education is poor, he has 2  5 years of experience, and his references, social skills and enthusiasm are all good.
 Harry’s education is good, he has < 2 years of experience, and his references, social skills and enthusiasm are all good.
Table 1: Example of a value model (points system) for ranking job candidates
Criterion Category Points Education poor 0 good 8 very good 20 excellent 40 Experience < 2 years 0 2  5 years 3 > 5 years 10 References poor 0 good 27 Social skills poor 0 good 10 Enthusiasm poor 0 good 13 Summing the point values in Table 1 corresponding to the descriptions for Tom, Dick and Harry gives their total scores:

 Tom’s total score = 40 + 10 + 0 + 0 + 0 = 50 points
 Dick’s total score = 0 + 3 + 27 + 10 + 13 = 53 points
 Harry’s total score = 8 + 0 + 27 + 10 + 13 = 58 points
Clearly, Harry has the highest total score. Therefore, according to the value model (and how the candidates were assessed) he is the best of the three candidates. (Though, clearly, relative to other candidates who could potentially have applied for the job, Harry is not as good as the best hypotheticallypossible candidate – who would score a ‘perfect’ 40 + 10 + 27 + 10 + 13 = 100 points.)
In general terms, having specified the criteria and categories for a given value model, the challenge is to derive point values that accurately reflect the relative importance of the criteria and categories to the decisionmaker. Deriving valid and reliable point values is arguably the most difficult task when creating a value model. The PAPRIKA method does this based on decisionmakers’ preferences as expressed using pairwise rankings of alternatives.
Overview of the PAPRIKA method
As mentioned at the start of the article, PAPRIKA is a (partial) acronym for ‘Potentially All Pairwise RanKings of all possible Alternatives’.
The PAPRIKA method pertains both to value models for ranking particular alternatives that are known to decisionmakers (e.g. as in the job candidates example above) and to models for ranking potentially all hypothetically possible alternatives in a pool that is changing over time (e.g. patients presenting for medical care). The following explanation is centred on this second type of application because it is more general.
PAPRIKA is based on the fundamental principle that an overall ranking of all possible alternatives representable by a given value model – i.e. all possible combinations of the categories on the criteria – is defined when all pairwise rankings of the alternatives visàvis each other are known (provided the rankings are consistent).
(As an analogy, suppose you wanted to rank all competitors at the next Olympic Games from the youngest to the oldest. If you knew how each person was pairwise ranked relative to everyone else with respect to their ages – i.e. for each possible pair of individuals, you identified who is the younger of the two individuals or that they’re the same age – then you could produce an overall ranking of competitors from the youngest to the oldest.)
However, depending on the number of criteria and categories, the number of pairwise rankings of all possible alternatives is potentially in the millions or even billions. Of course, though, many of these pairwise rankings are automatically resolved due to one alternative in the pair having a higher category for at least one criterion and none lower for the other criteria than for the other alternative – known as ‘dominated pairs’. But this still leaves potentially millions or billions of ‘undominated pairs’ – pairs of alternatives where one has a higher ranked category for at least one criterion and a lower ranked category for at least one other criterion than the other alternative, and hence a judgement is required for the alternatives to be pairwise ranked. With reference to the example of ranking job candidates in the previous section, an example of an undominated pair (of candidates) would be where one person in the pair is, say, highly educated but inexperienced whereas the other person is uneducated but highly experienced, and so a judgement is required to pairwise rank this pair.
For example, for a value model with eight criteria and four categories within each criterion, and hence 48 = 65,536 (n) possible alternatives, there are n(n−1)/2 = 2,147,450,880 pairwise rankings. Even after eliminating the 99,934,464 dominated pairs, there are still 2,047,516,416 undominated pairs to be ranked.^{[1]} Clearly, performing anywhere near this number of pairwise rankings is impossible without a special method.
PAPRIKA solves this problem by ensuring that the number of pairwise rankings that decisionmakers need to perform is kept to a minimum – only a small fraction of the potentially millions or billions of undominated pairs – so that the method is practicable. It does this by, for each undominated pair explicitly ranked by decisionmakers, identifying (and eliminating) all undominated pairs implicitly ranked as corollaries of this and other explicitly ranked pairs (via the transitivity property of additive value models, as illustrated in the simple demonstration later below).
The method begins with the decisionmaker pairwise ranking undominated pairs defined on just two criteria atatime (where, in effect, all other criteria’s categories are pairwise identical). Again with reference to the example of ranking job candidates, an example of such a pairwiseranking question (corresponding to the undominated pair mentioned earlier) is: “Who would you prefer to hire, someone who’s highly educated but inexperienced or other person who’s uneducated but highly experienced, all else the same?” Each time the decisionmaker ranks a pair, all undominated pairs implicitly ranked as corollaries are identified and discarded.
Having completed ranking undominated pairs defined on just two criteria atatime, this is followed, if the decisionmaker chooses to continue (she can stop at any time), by pairs with successively more criteria, until potentially all undominated pairs are ranked. Thus, Potentially All Pairwise RanKings of all possible Alternatives (hence the PAPRIKA acronym) are identified: as either dominated pairs (given) or undominated pairs explicitly ranked by the decisionmaker or implicitly ranked as corollaries. From the explicitly ranked pairs, point values are obtained via linear programming; although multiple solutions to the linear program are possible, the resulting point values all reproduce the same overall ranking of alternatives.
Simulations of PAPRIKA’s use reveal that if the decisionmaker stops after having ranked undominated pairs defined on just two criteria atatime, the resulting overall ranking of all possible alternatives is very highly correlated with the decisionmaker’s ‘true’ overall ranking obtained if all undominated pairs (involving more than two criteria) were ranked.^{[1]}
Therefore, for most practical purposes decisionmakers are unlikely to need to rank pairs defined on more than two criteria, thereby reducing the elicitation burden. For example, approximately 95 pairwise rankings are required for the value model with eight criteria and four categories each referred to above; 25 pairwise rankings for a model with five criteria and three categories each; and so on.^{[1]} The realworld applications of PAPRIKA referred to earlier suggest that decisionmakers are able to rank comfortably more than 50 and up to at least 100 pairs, and relatively quickly, and that this is sufficient for most applications.
Theoretical antecedents
The PAPRIKA method’s closest theoretical antecedent is Pairwise Tradeoff Analysis,^{[17]} a precursor to Adaptive Conjoint Analysis in marketing research.^{[18]} Like the PAPRIKA method, Pairwise Tradeoff Analysis is based on the idea that undominated pairs that are explicitly ranked by the decisionmaker can be used to implicitly rank other undominated pairs. Pairwise Tradeoff Analysis was abandoned in the late 1970s, however, because it lacked a method for systematically identifying implicitly ranked pairs.
The ZAPROS method (from Russian for ‘Closed Procedure Near References Situations’) was also proposed;^{[19]} however, with respect to pairwise ranking all undominated pairs defined on two criteria “it is not efficient to try to obtain full information”.^{[20]} As explained in the present article, the PAPRIKA method overcomes this efficiency problem.
A simple demonstration of the PAPRIKA method
The PAPRIKA method can be easily demonstrated via the simple example of determining the point values for a value model with just three criteria – denoted by ‘a’, ‘b’ and ‘c’ – and two categories within each criterion – ‘1’ and ‘2’, where 2 is the higher ranked category.^{[1]}
This value model’s six point values (two for each criterion) can be represented by the variables a1, a2, b1, b2, c1, c2 (a2 > a1, b2 > b1, c2 > c1), and the eight possible alternatives (2^{3} = 8) as ordered triples of the categories on the criteria (abc): 222, 221, 212, 122, 211, 121, 112, 111. These eight alternatives and their total score equations – derived by simply adding up the variables corresponding to the point values (which are as yet unknown: to be determined by the method being demonstrated here) – are listed in Table 2.
Undominated pairs are represented as ‘221 vs (versus) 212’ or, in terms of the total score equations, as ‘a2 + b2 + c1 vs a2 + b1 + c2’, etc. [Recall, as explained earlier, an ‘undominated pair’ is a pair of alternatives where one is characterised by a higher ranked category for at least one criterion and a lower ranked category for at least one other criterion than the other alternative, and hence a judgement is required for the alternatives to be pairwise ranked. Conversely, the alternatives in a ‘dominated pair’ (e.g. 121 vs 111 – corresponding to a1 + b2 + c1 vs a1 + b1 + c1) are inherently pairwise ranked due to one having a higher category for at least one criterion and none lower for the other criteria (and no matter what the point values are, given a2 > a1, b2 > b1 and c2 > c1, the pairwise ranking will always be the same).]
‘Scoring’ this model involves determining the values of the six point value variables (a1, a2, b1, b2, c1, c2) so that he decisionmaker’s preferred ranking of the eight alternatives is realised.
For many readers, this simple value model can perhaps be made more concrete by considering an example to which most people can probably relate: a model for ranking job candidates consisting of the three criteria (for example) (a) education, (b) experience, and (c) references, each with two ‘performance’ categories, (1) poor or (2) good. (This is a simplified version of the illustrative value model in Table 1 earlier in the article.)
Accordingly, each of this model’s eight possible alternatives can be thought of as being a ‘type’ (or profile) of candidate who might ever, hypothetically, apply. For example, ‘222’ denotes a candidate who is good on all three criteria; ‘221’ is a candidate who is good on education and experience but poor on references; ‘212’ a third who is good on education, poor on experience, and good on references; etc.
Finally, with respect to undominated pairs, 221 vs 212, for example, represents candidate 221 who has good experience and poor references whereas 212 has the opposite characteristics (and they both have good education). Thus, which is the better candidate ultimately depends on the decisionmaker’s preferences with respect to the relative importance of experience visàvis references.
Table 2: The eight possible alternatives and their totalscore equations
Alternative Totalscore equation 222 a2 + b2 + c2 221 a2 + b2 + c1 212 a2 + b1 + c2 122 a1 + b2 + c2 211 a2 + b1 + c1 121 a1 + b2 + c1 112 a1 + b1 + c2 111 a1 + b1 + c1 Identifying undominated pairs
PAPRIKA’s first step is to identify the undominated pairs. With just eight alternatives this can be done by pairwise comparing all of them visàvis each other and discarding dominated pairs.
This simple approach can be represented by the matrix in Figure 1, where the eight possible alternatives (in bold) are listed down the lefthand side and also along the top. Each alternative on the lefthand side is pairwise compared with each alternative along the top with respect to which of the two alternatives is higher ranked (i.e. in the present example, which candidate is more desirable for the job). The cells with hats (^) denote dominated pairs (where no judgement is required) and the empty cells are either the central diagonal (each alternative pairwise ranked against itself) or the inverse of the nonempty cells containing the undominated pairs (where a judgement is required).
Figure 1: Undominated pairs identified by pairwise comparing the eight possible alternatives (emboldened)
vs 222 221 212 122 112 121 211 111 222 ^ ^ ^ ^ ^ ^ ^ 221 (i) b2 + c1 vs b1 + c2 (ii) a2 + c1 vs a1 + c2 (iv) a2 + b2 + c1 vs a1 + b1 + c2 ^ ^ ^ 212 (iii) a2 + b1 vs a1 + b2 ^ (v) a2 + b1 + c2 vs a1 + b2 + c1 ^ ^ 122 ^ ^ (vi) a1 + b2 + c2 vs a2 + b1 + c1 ^ 112 (*i) b1 + c2 vs b2 + c1 (*ii) a1 + c2 vs a2 + c1 ^ 121 (*iii) a1 + b2 vs a2 + b1 ^ 211 ^ 111 Figure 1 notes: ^ denotes dominated pairs. The undominated pairs are labelled with Roman numerals; the three with asterisks are duplicates of pairs (i)(iii).
As summarised in Figure 1, there are nine undominated pairs (labelled with Roman numerals). However, three are duplicates after any variables common to a pair are ‘cancelled’ (e.g. pair *i is a duplicate of pair i, etc). Thus, there are six unique undominated pairs (without asterisks in Figure 2, and listed later below).
The cancellation of variables common to undominated pairs can be illustrated as follows. When comparing alternatives 121 and 112, for example, a1 can be subtracted from both sides of a1 + b2 + c1 vs a1 + b1 + c2. Similarly, when comparing 221 and 212, a2 can be subtracted from both sides of a2 + b2 + c1 vs a2 + b1 + c2. For both pairs this leaves the same ‘cancelled’ form: b2 + c1 vs b1 + c2.
Formally, these subtractions reflect the ‘jointfactor’ independence property of additive value models:^{[21]} the ranking of undominated pairs (in uncancelled form) is independent of their tied rankings on one or more criteria. Notationally, undominated pairs in their cancelled forms, like b2 + c1 vs b1 + c2, are also representable as _21 ‘‘vs’’ _12 – i.e. where ‘_’ signifies identical categories for the identified criterion.
In summary, here are the six undominated pairs for the value model:
 (i) b2 + c1 vs b1 + c2
 (ii) a2 + c1 vs a1 + c2
 (iii) a2 + b1 vs a1 + b2
 (iv) a2 + b2 + c1 vs a1 + b1 + c2
 (v) a2 + b1 + c2 vs a1 + b2 + c1
 (vi) a1 + b2 + c2 vs a2 + b1 + c1
The task is to pairwise rank these six undominated pairs, with the objective that the decisionmaker is required to perform the fewest pairwise rankings possible (thereby minimising the elicitation burden).
Ranking undominated pairs and identifying implicitly ranked pairs
Undominated pairs with just two criteria are intrinsically the least cognitively difficult for the decisionmaker to pairwise rank relative to pairs with more criteria. Thus, arbitrarily beginning here with pair (i) b2 + c1 vs b1 + c2, the decisionmaker is asked: “Which alternative do you prefer, _21 or _12 (i.e. given they’re identical on criterion a), or are you indifferent between them?” This choice, in other words, is between a candidate with good experience and poor references and another with poor experience and good references, all else the same.
Suppose the decisionmaker answers: “I prefer _21 to _12” (i.e. good experience and poor references is preferred to poor experience and good references). This preference can be represented by ‘_21 ≻_12’, which corresponds, in terms of total score equations, to b2 + c1 > b1 + c2 [where ‘≻’ and ‘~’ (used later) denote strict preference and indifference respectively, corresponding to the usual relations ‘>’ and ‘=’ for the total score equations].
Central to the PAPRIKA method is the identification of all undominated pairs implicitly ranked as corollaries of the explicitly ranked pairs. Thus, given a2 > a1 (i.e. good education ≻ poor education), it is clear that (i) b2 + c1 > b1 + c2 (as above) implies pair (iv) (see Figure 2) is ranked as a2 + b2 + c1 > a1 + b1 + c2. This reflects the transitivity property of (additive) value models. Specifically, 221≻121 (by dominance) and 121≻112 (i.e. pair i _21≻_12, as above) implies (iv) 221≻112; equivalently, 212≻112 and 221≻212 implies 221≻112.
Next, corresponding to pair (ii) a2 + c1 vs a1 + c2, suppose the decisionmaker is asked: “Which alternative do you prefer, 1_2 or 2_1 (given they’re identical on criterion b), or are you indifferent between them?” This choice, in other words, is between a candidate with poor education and good references and another with good education and poor references, all else the same.
Suppose the decisionmaker answers: “I prefer 1_2 to 2_1” (i.e. poor education and good references is preferred to good education and poor references). This corresponds to a1 + c2 > a2 + c1. Also, given b2 > b1 (good experience ≻ poor experience), this implies pair (vi) is ranked as a1 + b2 + c2 > a2 + b1 + c1.
Furthermore, the two explicitly ranked pairs (i) b2 + c1 > b1 + c2 and (ii) a1 + c2 > a2 + c1 imply pair (iii) is ranked as a1 + b2 > a2 + b1. This can easily be seen by adding the corresponding sides of the inequalities for pairs (i) and (ii) and cancelling common variables. Again, this reflects the transitivity property: (i) 121≻112 and (ii) 112≻211 implies (iii) 121≻211; equivalently, 122≻221 and 221≻212 implies 122≻212.
As a result of two explicit pairwise comparisons – i.e. explicitly performed by the decisionmaker – five of the six undominated pairs have been ranked. The decisionmaker may cease ranking whenever she likes (before all undominated pairs are ranked), but let’s suppose she continues and ranks the remaining pair (v) as a2 + b1 + c2 > a1 + b2 + c1 (i.e. in response to an analogous question to the two spelled out above).
Thus, all six undominated pairs have been ranked as a result of the decisionmaker explicitly ranking just three:
 (i) b2 + c1 > b1 + c2
 (ii) a1 + c2 > a2 + c1
 (v) a2 + b1 + c2 > a1 + b2 + c1
The overall ranking of alternatives and point values
Because the three pairwise rankings above are consistent – and all n (n−1)/2 = 28 pairwise rankings (n = 8) for this simple value model are known – a complete overall ranking of all eight possible alternatives is defined (1^{st} to 8^{th}): 222, 122, 221, 212, 121, 112, 211, 111.
Simultaneously solving the three inequalities above (i, ii, v), subject to a2 > a1, b2 > b1 and c2 > c1, gives the point values (i.e. the ‘points system’), reflecting the relative importance of the criteria to the decisionmaker. For example, one solution is: a1 = 0, a2 = 2, b1 = 0, b2 = 4, c1 = 0 and c2 = 3 (or normalised so the ‘best’ alternative, 222, scores 100 points: a1 = 0, a2 = 22.2, b1 = 0, b2 = 44.4, c1 = 0 and c2 = 33.3).
Thus, in the context of the example of a value model for ranking candidates for a job, the most important criterion is revealed to be (good) experience (b, 4 points) followed by references (c, 3 points) and, least important, education (a, 2 points). Although multiple solutions to the three inequalities are possible, the resulting point values all reproduce the same overall ranking of alternatives as listed above and reproduced here with their total scores:
 1^{st} 222: 2 + 4 + 3 = 9 points (or 22.2 + 44.4 + 33.3 = 100 points normalised) – i.e. total score from adding the point values above.
 2^{nd} 122: 0 + 4 + 3 = 7 points (or 0 + 44.4 + 33.3 = 77.8 points normalised)
 3^{rd} 221: 2 + 4 + 0 = 6 points (or 22.2 + 44.4 + 0 = 66.7 points normalised)
 4^{th} 212: 2 + 0 + 3 = 5 points (or 22.2 + 0 + 33.3 = 55.6 points normalised)
 5^{th} 121: 0 + 4 + 0 = 4 points (or 0 + 44.4 + 0 = 44.4 points normalised)
 6^{th} 112: 0 + 0 + 3 = 3 points (or 0 + 0 + 33.3 = 33.3 points normalised)
 7^{th} 211: 2 + 0 + 0 = 2 points (or 22.2 + 0 + 0 = 22.2 points normalised)
 8^{th} 111: 0 + 0 + 0 = 0 points (or 0 + 0 + 0 = 0 points normalised)
Other things worthwhile noting
First, the decisionmaker may decline to explicitly rank any given undominated pair (thereby excluding it) on the grounds that at least one of the alternatives considered corresponds to an impossible combination of the categories on the criteria. Also, if the decisionmaker cannot decide how to explicitly rank a given pair, she may skip it – and the pair may eventually be implicitly ranked as a corollary of other explicitly ranked pairs (via transitivity).
Second, in order for all undominated pairs to be ranked, the decisionmaker will usually be required to perform fewer pairwise ranking if some indicate indifference rather than strict preference. For example, if the decisionmaker had ranked pair (i) above as _21~_12 (i.e. indifference) instead of _21≻_12 (as above), then she would have needed to rank only one more pair rather than two (i.e. just two explicitly ranked pairs in total). On the whole, indifferently ranked pairs generate more corollaries with respect to implicitly ranked pairs than pairs that are strictly ranked.
Finally, the order in which the decisionmaker ranks the undominated pairs affects the number of rankings required. For example, if the decisionmaker had ranked pair (iii) before pairs (i) and (ii) then it is easy to show that all three would have had to be explicitly ranked, as well as pair (v) (i.e. four explicitly ranked pairs in total). However, determining the optimal order is problematical as it depends on the rankings themselves, which are unknown beforehand.
Applying PAPRIKA to ‘larger’ value models
Of course, most realworld value models have more criteria and categories than the simple example above, which means they have many more undominated pairs. For example, the value model referred to earlier with eight criteria and four categories within each criterion (and 4^{8} = 65,536 possible alternatives) has 2,047,516,416 undominated pairs in total (analogous to the nine identified in Figure 1), of which, excluding replicas, 402,100,560 are unique (analogous to the six in the example above).^{[1]} (As mentioned earlier, for a model of this size the decisionmaker is required to explicitly rank approximately 95 pairs defined on two criteria atatime, which most decisionmakers are likely to be comfortable with.)
For such realworld value models, the simple pairwisecomparisons approach to identifying undominated pairs used in the previous subsection (represented in Figure 1) is highly impractical. Likewise, identifying all pairs implicitly ranked as corollaries of the explicitly ranked pairs becomes increasingly intractable as the numbers of criteria and categories increase. The PAPRIKA method, therefore, relies on computationally efficient processes for identifying unique undominated pairs and implicitly ranked pairs respectively. The details of these processes are beyond the scope of this article, but are available elsewhere.^{[1]}
How does PAPRIKA compare with traditional scoring methods?
PAPRIKA entails a greater number of judgments (but typically fewer than 100 and often fewer than 50^{[1]}) than most ‘traditional’ scoring methods, such as direct rating,^{[22]} SMART,^{[23]} SMARTER^{[24]} and the Analytic Hierarchy Process.^{[25]} Clearly, though, different types of judgments are involved. For PAPRIKA, the judgements entail pairwise comparisons of undominated pairs (usually defined on just two criteria atatime), whereas most traditional methods involve interval scale or ratio scale measurements of the decisionmaker’s preferences with respect to the relative importance of criteria and categories respectively. Arguably, the judgments for PAPRIKA are simpler and more natural, and therefore they might reasonably be expected to reflect decisionmakers’ preferences more accurately.
See also
 Decision making
 Multicriteria decision analysis
 Conjoint analysis (marketing)
 Decisionmaking software
References
 ^ ^{a} ^{b} ^{c} ^{d} ^{e} ^{f} ^{g} ^{h} Hansen, P and Ombler, F (2009) “A new method for scoring multiattribute value models using pairwise rankings of alternatives”, Journal of MultiCriteria Decision Analysis, 15: 87107.
 ^ Wagstaff, A, “Asian Innovation Awards: Contenders stress different ways of thinking – entries vary from software for narrowing preferences to an imaginative auto”, The Asian Wall Street Journal, 21 September 2005, p. A15.
 ^ Taylor, W and Laking, G (2010), “Value for money – recasting the problem in terms of dynamic access prioritisation”, Disability & Rehabilitation, 32: 102027
 ^ Barber, A, Hansen, P, Naden, R, Ombler, F and Stewart, R (2011), “Who’s next? A new process for creating points systems for prioritising patients for elective health services”, Economics Discussion Papers No. 1104, University of Otago.
 ^ Fitzgerald, A, Conner Spady, B, De Coster, C, Naden, R, Hawker, GA and Noseworthy, T (2009), “WCWL Rheumatology Priority Referral Score reliability and validity testing”, abstract, The 2009 ACR/ARHP Annual Scientific Meeting, Arthritis & Rheumatology, 60 Suppl 10: 54.
 ^ Fitzgerald, A, De Coster, C, McMillan, S et al. (2011), “Relative urgency for referral from primary care to rheumatologists: The priority referral score”, Arthritis Care & Research, 63: 23139.
 ^ Noseworthy, T, De Coster, C and Naden, R (2009), “Prioritysetting tools for improving access to medical specialists”, poster presentation, 6th Health Technology Assessment International Annual Meeting, Singapore, 2009, Annals, Academy of Medicine, Singapore, 38: S78.
 ^ Neogi, T, Aletaha D, Silman AJ et al. (2010), “The 2010 American College of Rheumatology / European League Against Rheumatism classification criteria for rheumatoid arthritis: Phase 2 methodological report”, Arthritis & Rheumatism, 62: 258291.
 ^ Golan, O, Hansen, P, Kaplan, G and Tal, O (2011), “Health technology prioritization: Which criteria for prioritizing new technologies and what are their relative weights?”, Health Policy, 102: 12635.
 ^ Taylor, W, Singh, JA, Saag, KG et al. (2011), “Bringing it all together: A novel approach to the development of response criteria for chronic gout clinical trials”, The Journal of Rheumatology 38: 146770.
 ^ Ruhland, J (2006), “Strategic mobilization: What strategic management can learn from social movement research”, Management, 11: 2331.
 ^ Smith, C (2009), “Revealing monetary policy preferences”, Reserve Bank of New Zealand Discussion Paper Series, DP2009/01;
 ^ Smith, K and Fennessy, P (2011), “The use of conjoint analysis to determine the relative importance of specific traits as selection criteria for the improvement of perennial pasture species in Australia”, Crop & Pasture Science, 62: 35565.
 ^ Byrne, T, Amer, P, Fennessy, P, Hansen, P and Wickham, B (2011), “A preferencebased approach to deriving breeding objectives – applied to sheep breeding”, Animal (in press).
 ^ Boyd, P, Law, C and Doney, S (2011), “Commentary: A climate change atlas for the ocean” , Oceanography, 24: 1316.
 ^ Belton, V and Stewart, TJ, Multiple Criteria Decision Analysis: An Integrated Approach, Kluwer: Boston, 2002.
 ^ Johnson, RM (1976), “Beyond conjoint measurement: A method of pairwise tradeoff analysis”, Advances in Consumer Research 3: 35358.
 ^ Green, PE, Krieger, AB and Wind, Y (2001), “Thirty years of conjoint analysis: reflections and prospects”, Interfaces 31: S56S73.
 ^ Larichev, OI and Moshkovich, HM (1995), “ZAPROSLM – A method and system for ordering multiattribute alternatives”, European Journal of Operational Research 82: 50321.
 ^ Moshkovich, HM, Mechitov, AI and Olson, DL (2002), “Ordinal judgments in multiattribute decision analysis”, European Journal of Operational Research 137: 635.
 ^ Krantz, DH (1972), “Measurement structures and psychological laws”, Science, 175: 142735.
 ^ Von Winterfeldt, D and Edwards, W, Decision Analysis and Behavioral Research, Cambridge University Press: New York, 1986.
 ^ Edwards, W (1977), “How to use multiattribute utility measurement for social decision making”, IEEE Transactions in Systems, Man & Cybernetics 7: 32640.
 ^ Edwards, W and Barron, FH (1994), “SMARTS and SMARTER: Improved simple methods for multiattribute utility measurement”, Organizational Behavior & Human Decision Processes 60: 30625.
 ^ Saaty, TL, The Analytic Hierarchy Process, McGrawHill: New York, 1980.
Categories: Decision theory
 Operations research
 Clinical trials

Wikimedia Foundation. 2010.
Look at other dictionaries:
Pairwise comparison — This article is about pairwise comparisons in psychology. For statistical analysis of paired comparisons, see paired difference test. Pairwise comparison generally refers to any process of comparing entities in pairs to judge which of each entity … Wikipedia
Decisionmaking software — (DMS) is a term integrating decision analysis tools to facilitate a person s decision making process, which results in a choice of a course of action or a variant among several alternatives. DMS belongs to the class of decision support systems… … Wikipedia
Multicriteria decision analysis — Multiple criteria decision making or multiple criteria decision analysis is a sub discipline of operations research that explicitly considers multiple criteria in decision making environments. Whether in our daily lives or in professional… … Wikipedia
Decision making — For Decision making in groups, see Group decision making. Sample flowchart representing the decision process to add a new article to Wikipedia. Decision making can be regarded as the mental processes (cognitive process) resulting in the selection … Wikipedia
Decisionmaking paradox — The word paradox (parádoxon (παράδοξον) in Greek) comes from the Greek words para (meaning against, contrary to) and doksa or doxa (meaning belief, understanding). A paradox is a seemingly true statement or group of statements that lead to a… … Wikipedia
Voting system — For other uses, see Voting system (disambiguation). Part of the Politics series Electoral methods … Wikipedia