Next Article in Journal
A Machine Learning Framework for Condition-Based Maintenance of Marine Diesel Engines: A Case Study
Next Article in Special Issue
Meshfree Variational-Physics-Informed Neural Networks (MF-VPINN): An Adaptive Training Strategy
Previous Article in Journal
Explainable Machine Learning Model to Accurately Predict Protein-Binding Peptides
Previous Article in Special Issue
Algorithms for Various Trigonometric Power Sums
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Algorithm for Option Number Selection in Stochastic Paired Comparison Models

by
László Gyarmati
,
Csaba Mihálykó
*,† and
Éva Orbán-Mihálykó
Department of Mathematics, University of Pannonia, Egyetem u. 10, 8200 Veszprém, Hungary
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Algorithms 2024, 17(9), 410; https://doi.org/10.3390/a17090410
Submission received: 23 August 2024 / Revised: 6 September 2024 / Accepted: 9 September 2024 / Published: 14 September 2024
(This article belongs to the Special Issue Numerical Optimization and Algorithms: 2nd Edition)

Abstract

:
In this paper, paired comparison models with a stochastic background are investigated and compared from the perspective of the option numbers allowed. As two-option and three-option models are the ones most frequently used, we mainly focus on the relationships between two-option and four-option models and three-option and five-option models, and then we turn to the general s- and ( s + 2 ) -option models. We compare them from both theoretical and practical perspectives; the latter are based on computer simulations. We examine, when it is possible, mandatory, or advisable how to convert four-, five-, and ( s + 2 ) -option models into two-, three-, and s-option models, respectively. The problem also exists in reverse: when is it advisable to use four-, five-, and ( s + 2 ) -option models instead of two-, three-, and s-option models? As a result of these investigations, we set up an algorithm to perform the decision process.

1. Introduction

Comparisons in pairs are frequently used in a broad spectrum of problems, across all areas of life. Without claiming to be exhaustive, we mention education [1], management [2], psychology [3], marketing [4], construction [5], social science [6], group decisions [7], engineering [8], medicine [9], economics [10], resource allocation [11], and sports [12] as areas of their application. A paper by Vaidya and Kumar [13] presents more than 150 of the applications that they had been used for before 2003.
Paired comparisons play a significant role in decision-making [14,15]. This method allows for a detailed analysis of the differences and similarities between objects, properties, and aspects, aiding in the selection of the best choice. It is particularly useful in situations where multiple factors must be considered and a clear ranking is necessary [13,16]. By applying paired comparisons, complex problems can be simplified, leading to more objective decisions.
Paired comparisons means that, instead of using a scale, the objects to be evaluated are compared directly to each other, and the decision of the observer (evaluator) is based on the relative merits of the two objects. These opinions are generally more reliable than evaluations based on scales, as it is easier to determine which of two objects, properties, aspects, etc., is ‘better’, ‘more appropriate’, ‘more important’, etc., than to describe them with a single number. This is especially true for evaluations based on subjective criteria. In the case of paired comparisons, their data represent the relations between the observed objects. These types of data require methods that are different from conventional statistical procedures.
The most popular method based on comparisons in pairs is the Analytic Hierarchy Process (AHP) associated with Saaty [17]. This single paper has more than 10 thousand citations; some of these articles provide criticism, see for example [18]. The starting point of the model is a pairwise comparison matrix (PCM), the elements of which show how many times one object is stronger than another. The matrix A = ( a i , j ) is reciprocal–symmetric, that is a i , j = 1 a i , j . According to Saaty’s original model [17], a i , j { 1 / 9 , 1 / 8 , , 1 / 2 , 1 , 2 , , 8 , 9 } . The observers usually express their decisions verbally by voting for ‘equally preferred’, ‘moderately preferred’, ‘strongly preferred’, ’very strongly preferred’, or ‘extremely preferred’. These verbal expressions help decision-makers articulate their preferences in a qualitative manner, which are converted to quantitative values; 1, 3, 5, 7, or 9, respectively. Intermediate values are used for comparisons that fall between these levels. We mention that the number of options may be reduced in light of the problem under consideration. Their quantitative results are analyzed using different methods. Usually, every object is compared to every other object; that is, the PCM does not contain missing elements. In this case, the number of pairs under comparison increases quadratically with the number of objects examined. If the number of objects is 10, then the number of pairs equals 10 · 9 2 = 45 , which can be quite overwhelming for an observer. The reliability of their opinions and the measure of the contradictions between their decisions are checked by the inconsistency index of the PCM. The definition of consistency is given in [17] for complete cases; it was recently generalized to incomplete cases in [19]. There are numerous ways to quantify the degree of inconsistency seen [20,21]. Some research papers present procedures for reducing the level of inconsistency [22,23,24,25]. Most measurements are for complete cases, but they can also be generalized to incomplete cases [26]. The most frequently used measurement is related to the eigenvector method [17,24], which is the most common evaluation method, but the logarithmic least squares method (LLSM) is also readily used for evaluations, although mainly in incomplete cases [19,27,28].
Not only can the evaluation method change, but so can the way the PCM is constructed. Maintaining its reciprocal–symmetric property, the matrix can also be formed in another way than how Saaty constructed it. In [29], the authors define the elements of the PCM as the ratio of the number of wins to losses. In a socio-economic study [30], the author used the ratio of certain economic characteristics as the elements of the matrix. In this case, the number of possible options that require a decision is infinitely many. Both the eigenvector method (with some modification) and LLSM work for these PCM cases under the same conditions: the PCM matrix has to be irreducible, i.e., the representing graph must be connected [27,28].
We note that other methods, such as the Preference Ranking Organization Method for Enrichments Evaluations (PROMETHEE) and Elimination and Choice Translating Reality (ELECTREE, ELECTRE) are often applied, mainly when focusing on multi-criteria decision making [31,32,33].
Another branch of the models connected to comparisons in pairs are the Thurstone motivated models, which have stochastic background. Based on Thurstone’s brilliant idea, these models envision the latent probabilistic variables behind each object [34]. The starting points were psychological problems but, based on Thurstone motivated models, it is also possible to solve the ranking and rating of objects related to other research questions [35,36,37,38]. Thurstone assumed the latent variables had a Gaussian distribution, while Bradley and Terry assumed a logistic distribution [39]. In [34,39], two options were allowed for the decisions of the observers, but the models were generalized to three options, allowing ties [40,41]. Models allowing more than three options are presented in [42]. The ranking and rating of these objects are determined by the expectation of latent random variables. There are numerous possibilities for estimating their parameters. In the case of a Gaussian distribution, least squares methods are usually applied, but in cases of logistic distribution, maximum likelihood estimations are widespread. One publication [42] also applied least squares estimations based on relative frequencies and a cumulative distribution function. In [43], a general model for applying a Gaussian distribution and allowing more than two options is presented. The parameter estimation method is the maximum likelihood method, as it contains many inherent possibilities.
The maximum likelihood estimation (MLE) method is based on the maximization of a multi-variate function, and it requires the clarification of the conditions under which the maximum exists and is unique. In [43], the authors provide a sufficient condition for its existence and uniqueness in the case of a Gaussian distribution, and in [44] they generalize this for the case of general strictly log-concave distributions. Previously known conditions are used in [45], assuming a logistic distribution in the case of the two-option model, and in [46] they are used to investigate a modified model in the case of three choices. This condition, in the case of two options, is a necessary and sufficient condition. The condition given by Davidson is a sufficient condition but not necessary. It has been significantly generalized in [47] for the case of two and three options within choices, and the new system of conditions is a generalization of the set of conditions in [44]. The general conditions are met much more frequently [47], making them more useful for both theoretical research and practical applications. This set of conditions and its generalization for more than three options are used in this paper during comparisons of option numbers from a theoretical perspective.
During this research, the applied number of options is usually two or four, but there are cases when further numbers of options are allowed: four options are used in [43] for evaluating women tennis players, five options are used in [48], six options are used in [38], and seven options are used in [49]. In these studies, the numbers of options were fixed before data collection, based on preliminary empirical observations. If we fix the number of options, then the data will be available to us in a corresponding structure. It seems natural to use a fixed number of options during an evaluation. But data can also be transformed. In the past, ties were transformed into half a win and half a loss. Which option number is the best choice for an evaluation? Is this method better than an evaluation using a three-option model? Is it possible that by using a different number of options and transforming the data, we could obtain a model that fits the data better? In which cases is this possible? What can be said about the choice of the number of options? After studying the literature, we did not find an answer to this question. We intend this paper to fill this gap. As a result, we develop an algorithm by which the appropriate number of options can be selected. This algorithm relies on both theoretical and experimental aspects. Theoretical aspects concern evaluability considerations, while experimental aspects concern more accurate estimations and are supported by a large number of computer simulations.
This paper is structured as follows: In Section 2 we introduce a general model allowing for s options with an MLE parameter estimation. Then, we turn to the theoretical aspects of evaluability. First we investigate the relationship between two- and four-option models, and then the general case of s and s + 2 follows. The preliminary results for the case s = 3 and s + 2 = 5 are contained in the conference paper [50]. We can see that if data are evaluable in the s-option model and the data’s transformation satisfies a natural condition, then the data can also be evaluated in the ( s + 2 ) -option model. After these theoretical considerations, in Section 4 we present the results of a large number of computer simulations. The runs investigate how often the s-option model ( s = 2 , 3 ) and the ( s + 2 ) -option model ( s + 2 = 4 , 5 ) are evaluable. If both are evaluable, we investigate which one fits better with the data, and that is the one worth using for evaluations. We define a measurement for fitting which helps us decide on the choice of the number of options. In the subsequent section, we elaborate on and present an algorithm for performing the decision process, and the paper then concludes with our final remarks.

2. The Investigated Models

Let the objects to be rated and ranked be denoted by the numbers 1 , 2 , , n . The models with a stochastic background assume that the current performances of the objects are random variables denoted by ξ i , i = 1 , 2 , , n . Let the expectations be E( ξ i ) = m i , which expresses the expected strengths of the object i. Decisions comparing two objects i and j are related to the difference between these latent random variables, i.e.,  ξ i ξ j . We can separate the expectations as follows:
ξ i ξ j = m i m j + η i , j
where η i , j are independent identically distributed random variables with the cumulative distribution function F. F is supposed to be three times continuously differentiable, 0 < F < 1 , its probability density function is symmetric to zero, i.e.,  f ( x ) = f ( x ) , x R , and its logarithm is strictly concave. If  η i , j are normally distributed, then we use the nomination of the Thurstone model, if their distribution is logistic, we use the Bradley–Terry model. These models were originally defined for two-option choices, ‘worse’ or ’better’, and then were generalized for three options: ‘worse’, ‘better’, or ‘equal’. In the case of the two-option model, the differences are compared to zero: the decision ‘better’/‘worse’ indicates whether the difference is positive/negative, respectively. In the three-option model, the value of the difference is compared to the parameter 0 < d ; if it is in [ d , d ], we can consider it a tie. More than three, (s), options can be allowed too, as follows: The options within choices are denoted by C k , k = 1 , 2 , , s ( s = 5 : ‘much worse’, ‘slightly worse’, ‘equal’, ‘slightly better’, ‘much better’). The real line R is divided into disjoint sub-intervals I 1 , I 2 , …, I s . I 1 = ( , d 1 ) , I 2 = [ d 1 , d 2 ) , …, I s = [ d s 1 , ) . Decisions have a kind of symmetry, i.e., if i is ‘better’/‘much better’/…than j, then j is ‘worse’/‘much worse’/…than i. This means that d s 1 = d 1 and, in general, d s i = d i , i = 1 , 2 , , [ s / 2 ] . The figure belonging to the model allowing s options can be seen in Figure 1.
If s = 2 , then no parameter is needed to assign the endpoints of the intervals. If  s = 3 , 4 , then we need one positive parameter; we will denote it using 0 < d = d s 1 . If  s = 5 , 6 , we need two parameters, denoted by 0 < d = d s 2 < D = d s 1 , to assign the intervals, etc. These parameters will be estimated on the basis of the comparison results, which are called data.
The data are included into a three-dimensional ( n x n x s ) data matrix A ( s ) . Its elements A i , j , k ( s ) are the number of comparisons in which decision C k is the result when we compare object i and j. Of course, A i , i , k ( s ) = 0 , i = 1 , , n , k = 1 , , s and A i , j , k ( s ) = A j , i , s + 1 k ( s ) , i = 1 , , n , j = 1 , , n , k = 1 , , s . The upper index (s) indicates the number of options.
The probability that the difference between the random variables belonging to the objects i and j is in the interval I k can be expressed as follows, in the case of 2 < s :
p i , j , 1 = P ( ξ i ξ j I 1 ) = F ( d 1 ( m i m j ) )
p i , j , k = P ( ξ i ξ j I k ) = F ( d k ( m i m j ) ) F ( d k 1 ( m i m j ) ) , k = 2 , , s 1
and
p i , j , s = P ( ξ i ξ j I s ) = 1 F ( d s 1 ( m i m j ) ) .
If s = 2 , d 1 = 0 . In this case, instead of (2) and (4), the appropriate probabilities can be expressed by
p i , j , 1 = P ( ξ i ξ j I 1 ) = F ( 0 ( m i m j ) )
and
p i , j , 2 = P ( ξ i ξ j I 2 ) = 1 F ( 0 ( m i m j ) ) .
Assuming independent decisions, the probability of the data, i.e., the likelihood function, is
L ( A ( s ) | m 1 , , m n , d 1 , , d [ ( s 1 ) / 2 ] ) = k = 1 s i = 1 n 1 j = i + 1 n p i , j , k A i , j , k ( s ) .
We estimate the parameters by maximizing the likelihood function (7), or, equivalently, its logarithm:
l o g L ( A ( s ) | m 1 , m 2 , , m n , d 1 , , d [ ( s 1 ) / 2 ] ) = k = 1 s i = 1 n 1 j = i + 1 n A i , j , k ( s ) · l o g ( p i , j , k ) .
The maximum likelihood estimation of the parameters m ̲ = ( m 1 , , m n ) and d v ̲ = ( d 1 , , d [ ( s 1 ) / 2 ] ) is the argument for the maximal value of (8); that is
( m ̲ ^ ( s ) , d v ^ ̲ ( s ) ) = arg max m ̲ R n , 0 < d [ s 1 2 ] < < d 1 L ( A ( s ) | m ̲ , d v ̲ ) .
We note that the probabilities (2), (3), (4), (5), and (6) depend only on the differences between their expectations; therefore, one coordinate of the parameter vector m ̲ , for example m 1 , can be fixed at zero or the constraint i = 1 n m i = 0 can be assumed. Naturally, the maximum value is not necessarily always attained, but some conditions for the data can guarantee the existence and uniqueness of a maximizer. These conditions are different in the case of different options. Hereinafter, data will be referred to as evaluable in an s-option model if the maximum likelihood estimate of the parameters (9) exists and is unique.

3. Comparison of Models Allowing Different Options in Choices from a Theoretical Aspect

3.1. The Case of s = 2 and s + 2 = 4

The early models of paired comparisons allowed for only two options: ‘worse’ or ‘better’ [34,39]. Thurstone himself argued that, based on experience, people are able to make decisions between these two options and there is not necessarily a need for more. In some sports, such as tennis or basketball, matches can only end in a win or a loss. However, knowing the result, we can sometimes add nuance: for example, the victory was decisive, or one side barely defeated the other. In other words, there may be justification for additional decision options, such as a narrow victory (‘slightly better’/‘slightly defeated’) or a decisive victory (‘much better’/‘much worse’). These models, which allow for and process more information, could yield more nuanced results.
The data’s structure is contained in the data matrices but the data matrices A ( 4 ) and A ( 2 ) can be transformed into each other. We follow the following natural transformation rules:
A i , j , 1 ( 2 ) = A i , j , 1 ( 4 ) + A i , j , 2 ( 4 )
A i , j , 2 ( 2 ) = A i , j , 3 ( 4 ) + A i , j , 4 ( 4 )
This means that if we want to transform the four-option model into a two-option model, we merge the options ‘slightly better’ and ‘much better’ into the ‘better’ option. In the opposite case, we split the option ‘better’ into two parts, and the numbers of decisions will be divided into two parts. We require that 0 < A i 1 , j 1 , 3 ( 4 ) and 0 < A i 2 , j 2 , 4 ( 4 ) for at least one pair of the indices ( i 1 , j 1 ) and ( i 2 , j 2 ) , respectively. This requirement ensures that the data A i , j , k ( 4 ) actually belong to the four-option model.
Now, we compare these models based on their evaluability. First, we track which theoretical cases are possible:
(A)
Data are evaluable only in the two-option model.
(B)
Data are evaluable only in the four-option model.
(C)
Data are evaluable in both models.
(D)
Data are not evaluable in either model.
(Case A) We can prove that this theoretical case may not happen. For this, we summarize the conditions for evaluability in the case of the two-option and the four-option models. For this, we need two graph definitions.
Definition 1
(The graph belonging to A ( 2 ) ). The nodes are the objects. There is a directed edge from i to j (denoted by →) if  0 < A i , j , 2 ( 2 ) , i.e., there is a decision according to which object i is ‘better’ than object j. This graph will be denoted by G C ( A ( 2 ) ) .
Definition 2
(The graph belonging to A ( 4 ) ). The nodes are the objects. Two types of directed edges are contained in the graph. There is a directed edge from i to j (denoted by → and called the ‘slightly better’ edge) if  0 < A i , j , 3 ( 4 ) , i.e., there is a decision according to which object i is ‘slightly better’ than object j. The other type of directed edge is defined as follows: there is a double directed edge from i to j, denoted by ↠ and called the ‘much better edge’, if  0 < A i , j , 4 ( 4 ) . In other words, there exists a decision according to which object i is ‘much better’ than object j. This graph will be denoted by G C ( A ( 4 ) ) .
In the case of the two-option model, the necessary and sufficient condition for evaluability, given by Ford, for logistic distributions and generalized for strictly log-concave distributions in [47] is the following:
  • C.2.1. For every nonempty partition of S and S ¯ of the set {1, 2, …, n} ( S S ¯ = { 1 , 2 , , n } , S S ¯ = ), there are two (not necessarily different) index pairs ( i 1 , j 1 ) and ( i 2 , j 2 ) , i 1 , i 2 S , j 1 , j 2 S ¯ , for which 0 < A i 1 , j 1 , 1 ( 2 ) and 0 < A i 2 , j 2 , 2 ( 2 ) . In other words, these subsets are connected in both directions.
These types of graphs are known as strongly connected graphs in graph theory. This property is equivalent to the existence of a directed path between any two nodes.
In the case of the four-option model, the following set of conditions has been proven sufficient for evaluability:
  • C.4.1. Both types of edges are included in the graph G C ( A ( 4 ) ) .
  • C.4.2. For every nonempty partition of S and S ¯ ( S S ¯ = { 1 , 2 , , n } , S S ¯ = ), there are two (not necessarily different) index pairs ( i 1 , j 1 ) and ( i 2 , j 2 ) , i 1 , i 2 S , j 1 , j 2 S ¯ , for which
    0 < A i 1 , j 1 , 4 ( 4 ) and 0 < A i 2 , j 2 , 1 ( 4 )
    are satisfied or there exists such an index pair ( i 3 , j 3 ) for which
    0 < A i 3 , j 3 , 2 ( 4 ) or 0 < A i 3 , j 3 , 3 ( 4 )
    is satisfied.
  • C.4.3. There exists a cycle in the graph G C ( A ( 4 ) ) along the ‘slightly better’ and ‘much better’ edges which contains at least one ‘much better’ edge. In other words, there exists a cycle ( i 1 , i 2 , , i k , i 1 ), i l , l = 1 , 2 , , k for some k = 2 , , n , where i 1 , i 2 , , i k are distinct, for which 0 < A i l , i l + 1 , 3 ( 4 ) or 0 < A i l , i l + 1 , 4 ( 4 ) , and there exists at least one index pair in the cycle for which 0 < A i l , i l + 1 , 4 ( 4 ) . We note that this cycle might contain only two different nodes.
Using the above set of conditions for evaluability, C.2.1., as well as C.4.1., C.4.2., and C.4.3., we can prove the following theorem:
Theorem 1.
Suppose that there exists at least one index pair ( i 1 , j 1 ) for which 0 < A i 1 , j 1 , 3 ( 4 ) is satisfied and there exists an index pair ( i 2 , j 2 ) for which 0 < A i 2 , j 2 , 4 ( 4 ) holds. Redefine the data in the two-option model using (10) and (11). If condition C.2.1.is satisfied, then A ( 4 ) is evaluable in the four-option model.
Proof. 
We check the conditions C.4.1., C.4.2., and C.4.3.
C.4.1. is already required in the assumptions.
C.4.2. Consider an arbitrary partition S, S ¯ . As there exists an index pair ( i 1 , j 1 ) i 1 S j 1 S ¯ for which 0 < A i 1 , j 1 , 2 ( 2 ) , there exists an index pair for which 0 < A i 1 , j 1 , 3 ( 4 ) or 0 < A i 1 , j 1 , 4 ( 4 ) . Similarly, as there exists an index pair ( i 2 , j 2 ) i 2 S j 2 S ¯ for which 0 < A i 2 , j 2 , 1 ( 2 ) , there exists an index pair for which 0 < A i 2 , j 2 , 2 ( 4 ) or 0 < A i 2 , j 2 , 1 ( 4 ) . These mean that (12) or (13) is satisfied.
C.4.3. Let us choose a node from which a ‘much better’ edge starts ( i 1 ). It goes to the object i 2 . C.2.1. implies that there is a path along the directed edges from i 2 to i 1 in G C ( A ( 2 ) ) . We denote it by saying that ( i 2 , i 3 , , i 1 ) and i j are different. These edges are also included as ‘slightly better’ or ‘much better’ edges in G C ( A ( 4 ) ) . Therefore, the cycle ( i 1 , i 2 , , i 1 ) forms a cycle along the ‘slightly better’ and ‘much better’ edges in G C ( A ( 4 ) ) and it contains at least one ‘much better’ edge.
We proved that all three conditions, C.4.1., C.4.2., and C.4.3., hold; consequently, data matrix A ( 4 ) is evaluable.    □
(Case B) This case may occur. A simple example can be seen in Figure 2. The justification is as follows:
Data are evaluable in the four-option model as
C.4.1. is obviously satisfied; A 1 , 2 , 4 ( 4 ) = 1 and A 2 , 3 , 2 ( 4 ) = 1 .
C.4.2. In the case of every partition, the subsets are connected either by ‘much better’ edges, both forward and backward, or a ‘slightly better’ edge. More exactly, if  S = { 1 } , S ¯ = { 2 , 3 } , or  S = { 1 , 3 } , S ¯ = { 2 } , then A 1 , 2 , 4 ( 4 ) = 1 , A 2 , 1 , 4 ( 4 ) = A 1 , 2 , 1 ( 4 ) = 1 . If  S = { 1 , 2 } and S ¯ = { 3 } , then A 3 , 2 , 3 ( 4 ) = A 2 , 3 , 1 ( 4 ) = 1 . When S and S ¯ are changed, the situation remains the same.
C.4.3. The required cycle is (1,2,1).
Data cannot be evaluated using the two-option model as C.2.1. does not hold. See S = { 1 , 2 } and S ¯ = { 3 } ; moreover, take into the consideration that A 1 , 2 , 1 ( 2 ) = A 1 , 2 , 2 ( 2 ) = 2 , A 2 , 3 , 1 ( 2 ) = A 3 , 2 , 2 ( 2 ) = 1 , A 2 , 3 , 2 ( 2 ) = A 3 , 2 , 1 ( 2 ) = 0 , A 1 , 3 , 1 ( 2 ) = A 3 , 1 , 2 ( 2 ) = A 1 , 3 , 2 ( 2 ) = A 3 , 1 , 1 ( 2 ) = 0 .
In this case, the situation is clear; there is no choice.
(Case C) The data given by A ( 4 ) and its transformed version A ( 2 ) can be evaluated in the four- and two-option model, respectively; see Figure 3. This is a very common case. This simple example is a modification of the data presented in Figure 2. The comparison results in Figure 2 are supplemented by a ‘slightly better’ edge from object 2 to object 3; see Figure 3. The data are obviously evaluable in the four-option model. Moreover, C.2.1. is also satisfied; hence, data can be evaluated in the two-option model too.
In this case, we can decide which model will be used for the data evaluation. In Section 4, we investigate these cases’ practical aspects using computer simulations.
(Case D) Data cannot be evaluated in either the two-option or the four-option model.
Let A 1 , 2 , 1 ( 4 ) = A 1 , 2 , 2 ( 4 ) = A 1 , 2 , 3 ( 4 ) = A 1 , 2 , 4 ( 4 ) = 1 , A 2 , 3 , 1 ( 4 ) = A 2 , 3 , 2 ( 4 ) = A 2 , 3 , 3 ( 4 ) = 0 , A 2 , 3 , 4 ( 4 ) = 1 , A 1 , 3 , k ( 4 ) = 0 , and k = 1 , 2 , 3 , 4 . The graph of the results belonging to the data matrix A ( 4 ) can be seen in Figure 4.
The reason why the data are not evaluable is clear: object 3 is ‘worse’ than objects 2 and 1, but we do not know by how much they differ. Formally, conditions C.2.1. and C.4.2. do not hold: they take the partition S = { 1 , 2 } and S ¯ = { 3 } in both cases.

3.2. Comparison of Models Allowing Different Options within Choices: The General Case

The comparison of three-option and five-option cases is included in the conference paper [50]; therefore, we do not provide details here. Instead, in this section, we turn to the general case, which includes the analysis in Section 3.1 and the analysis of the cases s = 3 and s + 2 = 5.
Data will be denoted by A i , j , k ( s ) , ( k = 1 , 2 , , s ) and A i , j , k ( s + 2 ) , ( k = 1 , 2 , , s + 2 ) in the s and (s + 2)-option models, respectively. Data conversion is performed using the following formulas:
A i , j , 1 ( s ) = A i , j , 1 ( s + 2 ) + A i , j , 2 ( s + 2 )
A i , j , s ( s ) = A i , j , s + 1 ( s + 2 ) + A i , j , s + 2 ( s + 2 )
and
A i , j , k ( s ) = A i , j , k + 1 ( s + 2 ) , k = 2 , 3 , , s 1
This means that the two outermost decisions are merged and the ‘internal’ decisions remain unchanged. The appropriate graph definition is the following:
Definition 3
(The graph belonging to A ( s * ) ). The nodes are the objects. Next to the edges belonging to ties are two types of directed edge: The first one is a ‘simple’ directed edge from i to j (denoted by → and called the ‘better to some extent’ edge) that exists if  0 < A i , j , k ( s * ) k = [ s * + 1 2 ] + 1 , , s * 1 ; i.e., there is a decision according to which object i is ‘better to some extent’ than object j. The other is an ‘extremely better’ directed edge: it exists from i towards j if 0 < A i , j , s * ( s * ) and it is denoted by ↠. Expressed in words, there is an extreme decision according to which i exceeds j. This graph will be denoted by G C ( A ( s * ) ) . We substitute s * = s and s * = s + 2 .
The set of conditions that guarantees data evaluability in the model allowing s * options is as follows:
  • C.s*.1. For every value of k = 1 , 2 , , s * there exists an index pair i k , j k for which 0 < A i k , j k , k ( s * ) .
    We note that this condition is necessary. If it does not hold, the number of options has to be reduced.
  • C.s*.2. For every nonempty partition of S and S ¯ ( S S ¯ = { 1 , 2 , , n } , S S ¯ = ), there are two (not necessarily different) index pairs ( i 1 , j 1 ) and ( i 2 , j 2 ) i 1 , i 2 S , j 1 , j 2 S ¯ for which
    0 < A i 1 , j 1 , 1 ( s * ) and 0 < A i 2 , j 2 , s * ( s * )
    are satisfied or there exists such an index pair ( i 3 , j 3 ) for which
    0 < A i 3 , j 3 , k ( s * ) , for some k = 2 , 3 , , s * 1
    is satisfied. Expressed in words, the S and S ¯ are connected either by a non-extreme decision or by two extreme decisions made back and forth.
    We note that this condition is also a necessary condition, not just sufficient.
  • C.s*.3. There exists a cycle in the graph G C ( A ( s * ) ) along the ‘better to some extent’ and ‘extremely better’ edges which contains at least one ‘extremely better’ edge. In other words, there exists a cycle ( i 1 , i 2 , , i h , i 1 ) for which i 1 , i 2 , , i h are different, and 0 < A i l , i l + 1 , k l ( s * ) , where k l { [ s * + 1 2 ] + 1 , , s * } , and there exists at least one index pair in the cycle for which k l = s * . We note that it may be the case that this cycle contains only two different nodes.
As a generalization of Theorem 4 in [47], one can prove that conditions C.s*.1., C.s*.2., and C.s*.3. guarantee the evaluability of the data matrix A ( s * ) in the s * -option model. We will apply this statement for s * = s and s * = s + 2.

Comparison Based on Evaluability

(A)
Data are evaluable only in the s-option model.
(B)
Data are evaluable only in the (s+2)-option model.
(C)
Data are evaluable in both models.
(D)
Data are not evaluable in either model.
(Case A) This case cannot occur. We can state the following Theorem:
Theorem 2.
Suppose that there exists at least one index pair ( i 1 , j 1 ) for which 0 < A i 1 , j 1 , s + 1 ( s + 2 ) is satisfied and there exists an index pair ( i 2 , j 2 ) for which 0 < A i 2 , j 2 , s + 2 ( s + 2 ) holds. Redefine the data in the s-option model using (14), (15), and (16). If conditions C.s.1., C.s.2, and C.s.3. are satisfied, then A ( s + 2 ) is evaluable in the ( s + 2 ) -option model.
The proof of this can be easily carried out by following the steps of s = 2 cases step by step in Section 3.1.
(Case B) This case can happen; for example, take the following data matrix A ( s + 2 ) , which is a modification of the comparison results presented in Figure 2: A 1 , 2 , k ( s + 2 ) = A 2 , 1 , k ( s + 2 ) = 1 , k = 1 , 2 , , s + 2 , A 2 , 3 , 2 ( s + 2 ) = A 3 , 2 , s + 1 ( s + 2 ) = 1 , and the other values are zero. This implies that 0 < A 1 , 2 , k ( s ) , k = 1 , 2 , , s and A 2 , 3 , 1 ( s ) = A 3 , 2 , s ( s ) = 1 . Similarly to the case s = 2, in the general s-option model, object 3 exceeds object 2 but we do not know by how much. By increasing its expectation, its likelihood function increases. C.s.2. does not hold if  S = { 1 , 2 } and S ¯ = { 3 } .
In this case, it is clear which model should be used for the evaluation; there is no choice between them.
(Case C) Similar to the case where s = 2, it may be the case that the data can be evaluated using both numbers of options. As an example, consider the previous data matrix A, and supplement it with an ‘extremely better’ edge between the objects 2 and 3. That is, let A 1 , 2 , k ( s + 2 ) = A 2 , 1 , k ( s + 2 ) = 1 , k = 1 , 2 , , s + 2 , A 2 , 3 , 2 ( s + 2 ) = A 3 , 2 , s + 1 ( s + 2 ) = A 2 , 3 , s + 2 ( s + 2 ) = A 3 , 2 , 1 ( s + 2 ) = 1 , and the other values be zero. One can easily check that C.s.1, C.s.2., C.s,3., C.s + 2.1., C.s + 2.2., and C.s + 2.3. are all satisfied. Consequently, the data can be evaluated by both models.
In this case, we can choose which model to use for evaluating the data.
(Case D) The data cannot be evaluated using either the s-option or the ( s + 2 ) -option model.
An easily understandable example is a modification of the data presented in Figure 4. Let A 1 , 2 , k ( s + 2 ) = A 2 , 1 , k ( s + 2 ) = 1 , k = 1 , 2 , , s + 2 , and  A 2 , 3 , s + 2 ( s + 2 ) = 1 , while the other values are zero. Now, 0 < A 1 , 2 , k ( s ) = A 2 , 1 , s + 1 k ( s ) k = 1 , 2 , , s and  A 2 , 3 , s ( s ) = A 3 , 2 , 1 ( s ) = 1 , while the other values are zero. In this case, neither C.s.2. nor C.s + 2.2. holds, as S = { 1 , 2 } and S ¯ = { 3 } .
In this case, further data have to be collected; that is, new comparisons have to be made.

4. Simulation Results

In this chapter, we deal with the two most commonly used option numbers: s = 2 and s = 3. First, we present the relationship between the two- and four-option models, and then we turn to the relationship between the three- and five-option models. We analyze these relationships by applying computer simulations. We point out when it is more appropriate to use which model in order to generate the best possible retrieval of the strength of the objects.
In these cases, we predefined the strength (expectation) of each object within certain bounds and the separation parameters for each object, including 0 < d for the three-, four- and five-option models and d < D for the five-option model. We then generated comparison results for different comparison numbers using these fixed strengths ( m i ) and values d, D. We checked whether the data set could be evaluated using the s- and ( s + 2 ) -option models. If it was possible, we evaluated the data sets using both the s- and ( s + 2 ) -option models and compared the estimated strengths obtained by the models with the original strengths m i , i = 1 , 2 , , n . We used a large number of simulations ( 10 5 ) in our work, applying logistic distribution. The simulation program was implemented in C#.

4.1. The Cases s = 2 and s + 2 = 4

The two-option and four-option models are compared in this subsection. Table 1 shows an example for the simulation results of the two models. We present a case in Table 1 where the probabilities of an object falling into the options available are approximately equal. In the presented situation, there are eight objects to compare, their expectations are between 0 and 2, and d = 1.35 . In the case of the four-option model, the ‘slightly better’ and ‘much better’ options can be distinguished: a ‘slightly worse’ or ‘much worse’ decision is a ‘slightly better’ or ‘much better’ decision from the perspective of the other object. In this case, the ratio of ‘much better’ decisions to all decisions was 0.5186, while the ratio of ‘slightly better’ was 0.4814.
The first column in Table 1 shows the number of comparisons denoted by | A | . The second column shows the number of cases in which we were able to use both models. The third column shows the number of cases for which we were only able to evaluate objects using the four-option model and the fourth column shows the number of cases for which we were not able to make an evaluation using either model. The fifth and sixth columns contain the average sum of the squared differences between the exact and the estimated expectations found by the two-option model and the four-option model, respectively. The quantities
e ( s * ) = i = 1 n ( m ^ i ( s * ) m i ) 2
measure how close the estimated expected values are to the exact expected values in the model allowing s * options. They are zero if, and only if, the estimated parameters coincide with the exact parameters. If we take the average of these values, we get a measure of the model’s accuracy, denoted by e ( s * ) ¯ .
The seventh (last) column shows their relative relation, which is defined as follows:
R ( 2 ) , ( 4 ) = e ( 2 ) ¯ e ( 4 ) ¯ e ( 4 ) ¯ .
In these formulas, as previously, the upper index refers to the option number of the model used for the evaluation. As the quantities (19) are non-negative, the sign of (20) characterizes which model is better: if (20) is positive, then the four-option model is better; if it is negative, then the two-option model is. It can clearly be seen that in the case of few comparisons—i.e., when the number of comparisons conducted is less than 21—the two-option model is able to approximate the exact expectations more accurately. Meanwhile, in the case of more comparisons, the more complex four-option model performs better. It can also clearly be seen that as more comparisons are included, we can make better evaluations with both models and the average errors e ( 2 ) ¯ and e ( 4 ) ¯ become smaller and smaller. Due to space limitations, we have only presented the case m i : 0 2 , d = 1.35 , but in other cases the results are similar: if the number of comparisons reaches 21, the more complex four-option model performs better than for 8 compared objects. We conducted simulations with different numbers of objects to be evaluated and we observed a linear relationship between the number of objects evaluated and the number of comparisons when the model with more options performs better, on average. The critical number is the twice of the number of the objects plus 5 in the case of s = 2. This means that if 2 · n + 5 | A | , then the complicated model is advisable.
This phenomenon can be explained as follows: if there are few comparisons, randomness plays a significant role. The more options there are, the greater the variability. As the number of comparisons increases, the generated data increasingly follow the probabilities derived by the model. The parameters become easier to estimate accurately. Models with more options contain more information, allowing for more precise estimates.

4.2. The Cases s = 3 and s + 2 = 5

Now we compare the three-option and five-option models. The reason for choosing s = 3 as the general case is that it is often used, and particularly in sports applications. In numerous sports, the possible outcomes of matches are a draw, a victory, or a defeat. Examples of such sports include football and handball. However, if we want to evaluate the results using a three-option model, we cannot do so. One possible reason for this could be that one team defeats every other team. This happened in the 2024 UEFA European Championship: Spain won all of its matches. If we evaluate these results using the five-option model, the evaluation can be carried out. Many other examples can be given of this phenomenon. These also highlight the importance of examining the problem carefully.
We conducted computer simulations. In the presented example, the number of compared objects was eight. The expectations were chosen from the interval [ 0 , 2 ] and d and D were fixed: 0 < d < D , d = 0.1 , D = 1.35 . In this case, we wanted to show an extreme scenario in which there are almost no ‘equal’ decisions, while the number of ‘slightly better’ and ‘much better’ decisions is approximately equal. The ratios of the decisions ‘equal’, ‘slightly better’, and ‘much better’ are 0.0540/0.4768/0.4692, respectively. The simulation results are shown in Table 2.
The structure of Table 2 follows the structure of Table 1; the columns’ names and reports change according to the option numbers of the models. Here, as in the two-option model, the accuracy of estimating expected values is better for the simpler model if the numbers of comparisons are small, while, for larger numbers of comparisons, the more complex model performs better. As more and more comparisons are made between random objects, the errors become smaller and smaller. Based on lots of further experiments, we can ascertain that, in the case of 8 objects and at least 20 comparisons, the five-option model estimates the expectations m i ( i = 1 , 2 , , n ) better, on average; therefore, it is recommended.
Of course, the number of comparisons depends on the number of objects to be evaluated when the model that allows more options is able to use its additional information. By conducting a large number of simulations with different object numbers, we can conclude that five-option model performs better than the three-option model if  2 · n + 4 | A | .

5. Algorithm for Model Choice

In this section, based on the results of the previous sections, we present an algorithm that allows us to select which model to use; see Algorithm 1.
The function E v ( A , o m ) specifies, for a given A (data set), whether A can be evaluated with an o m -option model. As previously, we denote the number of comparisons as | A | . The number of ‘equal’ choices is denoted by A e q u a l . The auxiliary variable k that we use takes only positive integer values. The function C a l c ( A , o m ) evaluates the data set A in an o m -option model. M d c ( ) represents additional data collection. o c denotes the number of objects in the comparisons, while C m ( o m ) is the constant multiplier in each case where o m = s + 2 . For  s + 2 = 4 and s + 2 = 5 , the C m ( o m ) is 2. The  o d is a constant value for each model; it equals 5 for s + 2 = 4 and it equals 4 for s + 2 = 5 .
We present the algorithm described below in the flowchart in Figure 5.
Algorithm 1 Option Number Selection
1:
if  s 2 = 2 k + 1   and  A e q u a l = 0   then
2:
     s 2 : = s 2 1
3:
end if
4:
while not  E v ( A , s 2 )  and  s 2 > 3 do
5:
     s 2 : = s 2 2
6:
end while
7:
if  s 2 > 3  then
8:
     s : = s 2 2
9:
    if  E v ( A , s )  then
10:
        if  | A | C m ( s 2 ) · o c + o d  then
11:
            C a l c ( A , s 2 )
12:
        else
13:
            C a l c ( A , s )
14:
        end if
15:
    else
16:
         C a l c ( A , s 2 )
17:
    end if
18:
else
19:
    if  E v ( A , s 2 )  then
20:
         C a l c ( A , s 2 )
21:
    else
22:
         M d c ( )
23:
    end if
24:
end if

6. Conclusions

The selection of the number of allowed options was studied to improve evaluability and provide a better approximation of the strength of the objects compared. An algorithm was created to decide which option number is best to use for data evaluations. The algorithm includes theoretical results for evaluability and empirical observations based on simulations. Based on theoretical considerations, we can assert that if the data can be evaluated in an s-option model, then it can also be evaluated in an ( s + 2 ) -option model, provided that there are comparison results for every option. However, the statement is not true in the reverse scenario: there are situations where only the model with a higher number of options is suitable for the evaluation. If both models are suitable for the evaluation, we have provided a metric to determine which model allows for a more accurate estimation of the objects’ strengths. We can conclude that, in the case of a small number of comparisons, the simpler model is more suitable. However, beyond a certain number of comparisons, the model that allows for more options is better, as it can utilize its additional information. This critical number is approximately twice the number of objects to be evaluated plus five for the s + 2 = 4-option model and four for the s + 2 = 5-option model. Further investigations are needed for cases where s + 2 = 6, 7, and beyond.

Author Contributions

Conceptualization, L.G., C.M. and É.O.-M.; methodology, L.G., C.M. and É.O.-M.; software, L.G.; validation, L.G. and C.M.; writing, É.O.-M. and C.M.; review and editing, C.M., É.O.-M. and L.G.; visualization, L.G.; supervision, C.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Most of the data are computer-generated. Little public data were used.

Acknowledgments

This research was partly supported by the ÚNKP-23-3 New National Excellence Program of the Ministry for Culture and Innovation from the source of the National Research, Development and Innovation Fund. L. Gyarmati thanks them for the support. This work has been implemented by the TKP2021-NVA-10 project with the support provided by the Ministry of Culture and Innovation of Hungary from the National Research, Development and Innovation Fund, financed under the 2021 Thematic Excellence Programme funding scheme.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Erdil, Ü.D.; Tümer, M.; Nadiri, H.; Aghaei, I. Prioritizing information sources and requirements in students’ choice of higher education destination: Using AHP analysis. Sage Open 2021, 11. [Google Scholar] [CrossRef]
  2. Esangbedo, M.O.; Bai, S.; Mirjalili, S.; Wang, Z. Evaluation of human resource information systems using grey ordinal pairwise comparison MCDM methods. Expert Syst. Appl. 2021, 182, 115151. [Google Scholar] [CrossRef]
  3. Kulak, A.; Selvi, H. Scaling the Psychological Variables that Affect Performance of Athletes by Means of Pairwise Comparison Method. Int. J. Recreat. Sports Sci. 2019, 3, 16–24. [Google Scholar] [CrossRef]
  4. Krivulin, N.; Prinkov, A.; Gladkikh, I. Using pairwise comparisons to determine consumer preferences in hotel selection. Mathematics 2022, 10, 730. [Google Scholar] [CrossRef]
  5. Darko, A.; Chan, A.P.C.; Ameyaw, E.E.; Owusu, E.K.; Pärn, E.; Edwards, D.J. Review of application of Analytic Hierarchy Process (AHP) in construction. Int. J. Constr. Manag. 2019, 19, 436–452. [Google Scholar] [CrossRef]
  6. Lee, S.E.; Choi, M.; Kim, S. How and what to study about IoT: Research trends and future directions from the perspective of social science. Telecommun. Policy 2017, 41, 1056–1067. [Google Scholar] [CrossRef]
  7. Sasaki, Y. Strategic manipulation in group decisions with pairwise comparisons: A game theoretical perspective. Eur. J. Oper. Res. 2023, 304, 1133–1139. [Google Scholar] [CrossRef]
  8. Dym, C.L.; Wood, W.H.; Scott, M.J. Rank ordering engineering designs: Pairwise comparison charts and Borda counts. Res. Eng. Des. 2002, 13, 236–242. [Google Scholar] [CrossRef]
  9. Amlani, A.M.; Schafer, E.C. Application of paired-comparison methods to hearing aids. Trends Amplif. 2009, 13, 241–259. [Google Scholar] [CrossRef]
  10. Rosenberger, R.S.; Peterson, G.L.; Loomis, J.B. Applying a method of paired comparisons to measure economic values for multiple goods sets. J. Agric. Appl. Econ. 2002, 34, 215–229. [Google Scholar] [CrossRef]
  11. Huang, J.J.; Chen, C.Y. Resource Allocation of Cooperative Alternatives Using the Analytic Hierarchy Process and Analytic Network Process with Shapley Values. Algorithms 2024, 17, 152. [Google Scholar] [CrossRef]
  12. Temesi, J.; Szádoczki, Z.; Bozóki, S. Incomplete pairwise comparison matrices: Ranking top women tennis players. J. Oper. Res. Soc. 2024, 75, 145–157. [Google Scholar] [CrossRef]
  13. Vaidya, O.S.; Kumar, S. Analytic hierarchy process: An overview of applications. Eur. J. Oper. Res. 2006, 169, 1–29. [Google Scholar] [CrossRef]
  14. Weernink, M.G.; Janus, S.I.; van Til, J.A.; Raisch, D.W.; van Manen, J.G.; IJzerman, M.J. A systematic review to identify the use of preference elicitation methods in healthcare decision making. Pharm. Med. 2014, 28, 175–185. [Google Scholar] [CrossRef]
  15. Cheng, K.E.; McHugh, J.A.; Deek, F.P. On the use of paired comparisons to construct group preference scales for decision making. Group Decis. Negot. 2013, 22, 519–540. [Google Scholar] [CrossRef]
  16. Zhang, G.; Lu, J.; Gao, Y. Multi-Level Decision Making, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2015; ISBN 978–3-662-46059-7. [Google Scholar]
  17. Saaty, T.L. A scaling method for priorities in hierarchical structures. J. Math. Psychol. 1977, 15, 234–281. [Google Scholar] [CrossRef]
  18. Dyer, J.S. Remarks on the Analytic Hierarchy Process. Manag. Sci. 1990, 36, 249–258. [Google Scholar] [CrossRef]
  19. Bozóki, S.; Tsyganok, V. The (logarithmic) least squares optimality of the arithmetic (geometric) mean of weight vectors calculated from all spanning trees for incomplete additive (multiplicative) pairwise comparison matrices. Int. J. Gen. Syst. 2019, 48, 362–381. [Google Scholar] [CrossRef]
  20. Brunelli, M. A survey of inconsistency indices for pairwise comparisons. Int. J. Gen. Syst. 2018, 47, 751–771. [Google Scholar] [CrossRef]
  21. Brunelli, M.; Fedrizzi, M. Inconsistency indices for pairwise comparisons and the Pareto dominance principle. Eur. J. Oper. Res. 2024, 312, 273–282. [Google Scholar] [CrossRef]
  22. Sato, Y.; Tan, K.H. Inconsistency indices in pairwise comparisons: An improvement of the consistency index. Ann. Oper. Res. 2023, 326, 809–830. [Google Scholar] [CrossRef]
  23. Ágoston, K.C.; Csató, L. A lexicographically optimal completion for pairwise comparison matrices with missing entries. Eur. J. Oper. Res. 2024, 314, 1078–1086. [Google Scholar] [CrossRef]
  24. Tekile, H.A.; Fedrizzi, M.; Brunelli, M. Constrained eigenvalue minimization of incomplete pairwise comparison matrices by Nelder-Mead algorithm. Algorithms 2021, 14, 222. [Google Scholar] [CrossRef]
  25. Pascoe, S. A simplified algorithm for dealing with inconsistencies using the Analytic Hierarchy Process. Algorithms 2022, 15, 442. [Google Scholar] [CrossRef]
  26. Ágoston, K.C.; Csató, L. Inconsistency thresholds for incomplete pairwise comparison matrices. Omega 2022, 108, 102576. [Google Scholar] [CrossRef]
  27. Harker, P.T. Incomplete pairwise comparisons in the analytic hierarchy process. Math. Modell. 1987, 9, 837–848. [Google Scholar] [CrossRef]
  28. Bozóki, S.; Fülöp, J.; Rónyai, L. On optimal completion of incomplete pairwise comparison matrices. Math. Comput. Modell. 2010, 52, 318–333. [Google Scholar] [CrossRef]
  29. Bozóki, S.; Csató, L.; Temesi, J. An application of incomplete pairwise comparison matrices for ranking top tennis players. Eur. J. Oper. Res. 2016, 248, 211–218. [Google Scholar] [CrossRef]
  30. Petróczy, D.G. An alternative quality of life ranking on the basis of remittances. Socio-Econ. Plann. Sci. 2021, 78, 101042. [Google Scholar] [CrossRef]
  31. Brans, J.P.; Vincke, P.; Mareschal, B. How to select and how to rank projects: The PROMETHEE method. Eur. J. Oper. Res. 1986, 24, 228–238. [Google Scholar] [CrossRef]
  32. Behzadian, M.; Kazemzadeh, R.B.; Albadvi, A.; Aghdasi, M. PROMETHEE: A comprehensive literature review on methodologies and applications. Eur. J. Oper. Res. 2010, 200, 198–215. [Google Scholar] [CrossRef]
  33. Aherdoost, H.; Madanchian, M. A Comprehensive Overview of the ELECTRE Method in Multi-Criteria Decision-Making. J. Manag. Sci. Eng. Res. 2023, 6, 5–16. [Google Scholar] [CrossRef]
  34. Thurstone, L.L. A law of comparative judgment. Psychol. Rev. 1927, 34, 273–286. [Google Scholar] [CrossRef]
  35. Tutz, G.; Schauberger, G. Extended ordered paired comparison models with application to football data from German Bundesliga. AStA Adv. Stat. Anal. 2015, 99, 209–227. [Google Scholar] [CrossRef]
  36. Jeon, J.J.; Kim, Y. Revisiting the Bradley-Terry model and its application to information retrieval. J. Korean Data Inf. Sci. Soc. 2013, 24, 1089–1099. [Google Scholar] [CrossRef]
  37. Stein, A.; Aryal, J.; Gort, G. Use of the Bradley-Terry model to quantify association in remotely sensed images. IEEE Trans. Geosci. Remote Sens. 2005, 43, 852–856. [Google Scholar] [CrossRef]
  38. Kosztyán, Z.T.; Orbán-Mihálykó, É.; Mihálykó, C.; Csányi, V.V.; Telcs, A. Analyzing and clustering students’ application preferences in higher education. J. Appl. Stat. 2020, 47, 2961–2983. [Google Scholar] [CrossRef]
  39. Bradley, R.A.; Terry, M.E. Rank analysis of incomplete block designs: I. The method of paired comparisons. Biometrika 1952, 39, 324–345. [Google Scholar] [CrossRef]
  40. Glenn, W.A.; David, H.A. Ties in paired-comparison experiments using a modified Thurstone-Mosteller model. Biometrics 1960, 16, 86–109. [Google Scholar] [CrossRef]
  41. Rao, P.V.; Kupper, L.L. Ties in paired-comparison experiments: A generalization of the Bradley-Terry model. J. Am. Stat. Assoc. 1967, 62, 194–204. [Google Scholar] [CrossRef]
  42. Agresti, A. Analysis of ordinal paired comparison data. J. R. Stat. Soc. Ser. C Appl. Stat. 1992, 41, 287–297. [Google Scholar] [CrossRef]
  43. Orbán-Mihálykó, É.; Mihálykó, C.; Koltay, L. A generalization of the Thurstone method for multiple choice and incomplete paired comparisons. Cent. Eur. J. Oper. Res. 2019, 27, 133–159. [Google Scholar] [CrossRef]
  44. Orbán-Mihálykó, É.; Mihálykó, C.; Koltay, L. Incomplete paired comparisons in case of multiple choice and general log-concave probability density functions. Cent. Eur. J. Oper. Res. 2019, 27, 515–532. [Google Scholar] [CrossRef]
  45. Ford, L.R., Jr. Solution of a ranking problem from binary comparisons. Am. Math. Mon. 1957, 64, 28–33. [Google Scholar] [CrossRef]
  46. Davidson, R.R. On Extending the Bradley-Terry Model to Accommodate Ties in Paired Comparison Experiments. J. Am. Stat. Assoc. 1970, 65, 317–328. [Google Scholar] [CrossRef]
  47. Gyarmati, L.; Orbán-Mihálykó, É.; Mihálykó, C. Comparative analysis of the existence and uniqueness conditions of parameter estimation in paired comparison models. Axioms 2023, 12, 575. [Google Scholar] [CrossRef]
  48. Orbán-Mihálykó, É.; Koltay, L.; Szabó, F.; Csuti, P.; Kéri, R.; Schanda, J. A new statistical method for ranking of light sources based on subjective points of view. Acta Polytech. Hung. 2015, 12, 195–214. [Google Scholar]
  49. Gyarmati, L.; Edvy, L.; Mihálykó, C.; Orbán-Mihálykó, É. Decision support for sports selection based on performance measurement applying the generalized Thurstone method. Int. J. Sports Sci. Coach. 2024, 17479541241240609. [Google Scholar] [CrossRef]
  50. Gyarmati, L.; Mihálykó, C.; Orbán-Mihálykó, É. Application of different option numbers in Thurstone motivated models. In Proceedings of the 3rd Conference on Information Technology and Data Science, Debrecen, Hungary, 26–28 August 2024. [Google Scholar]
Figure 1. The options and the intervals belonging to them in an s-option model.
Figure 1. The options and the intervals belonging to them in an s-option model.
Algorithms 17 00410 g001
Figure 2. Comparison results which can be evaluated using the four-option model but not in the two-option model (the objects to be compared are denoted by 1, 2, and 3).
Figure 2. Comparison results which can be evaluated using the four-option model but not in the two-option model (the objects to be compared are denoted by 1, 2, and 3).
Algorithms 17 00410 g002
Figure 3. Comparison results which can be evaluated in both models (objects to be compared are denoted by 1, 2, and 3).
Figure 3. Comparison results which can be evaluated in both models (objects to be compared are denoted by 1, 2, and 3).
Algorithms 17 00410 g003
Figure 4. Comparison results which can be evaluated in neither model (objects to be compared are denoted by 1, 2, and 3).
Figure 4. Comparison results which can be evaluated in neither model (objects to be compared are denoted by 1, 2, and 3).
Algorithms 17 00410 g004
Figure 5. Flowchart of the algorithm for the choice of option numbers.
Figure 5. Flowchart of the algorithm for the choice of option numbers.
Algorithms 17 00410 g005
Table 1. Comparison of the simulation results of the two- and four-option models using the parameters m i : 0 2 , d = 1.35 .
Table 1. Comparison of the simulation results of the two- and four-option models using the parameters m i : 0 2 , d = 1.35 .
| A | 2 and 4Only 4Neither e ( 2 ) ¯ e ( 4 ) ¯ R ( 2 ) , ( 4 )
1023400695,9713.8876.059−0.359
1185861191,3044.5666.405−0.287
1227914,81984,9024.8806.823−0.285
1370821,87277,4205.1687.231−0.285
14147329,08669,4415.3136.917−0.232
15264635,96261,3925.4216.736−0.195
16419542,12153,6845.5766.655−0.162
17640147,08046,5195.7096.465−0.117
18898150,95940,0605.8396.332−0.078
1911,99953,68334,3185.9246.179−0.041
2015,49955,24429,2575.9465.970−0.004
2119,28455,87824,8385.9715.8110.027
2223,22055,72021,0605.9395.6430.052
2327,24754,96217,7915.9215.4560.085
2431,45353,50215,0455.8825.2840.113
2535,55251,61212,8365.8085.0900.141
3054,84539,61355425.4244.2960.263
4079,97718,94910744.4413.1510.410
5091,26585142213.5692.4170.477
6096,1423797612.8981.9290.503
7098,2241763132.3881.5900.502
8099,16783122.0111.3430.497
9099,58741211.7291.1630.487
10099,78121901.5101.0230.476
20099,999100.6500.4600.412
300100,000000.4140.2970.394
400100,000000.3040.2190.387
500100,000000.2400.1740.382
600100,000000.1990.1440.381
700100,000000.1690.1230.378
800100,000000.1480.1070.376
900100,000000.1310.0950.374
1000100,000000.1170.0850.373
Table 2. Comparison of the simulation results of the three- and five-option models using the parameters m i : 0 2 , d = 0.1 , D = 1.35 .
Table 2. Comparison of the simulation results of the three- and five-option models using the parameters m i : 0 2 , d = 0.1 , D = 1.35 .
| A | 3 and 5Only 5Neither e ( 3 ) ¯ e ( 5 ) ¯ R ( 3 ) , ( 5 )
1039228397,6785.7157.765−0.264
11168489994,9336.2039.178−0.324
12454831391,2336.5088.666−0.249
1393612,43786,6276.8388.495−0.195
14178816,61781,5956.8708.342−0.177
15304120,51376,4466.9598.112−0.142
16473823,83771,4257.0917.820−0.093
17676626,47066,7647.0747.563−0.065
18917128,70462,1257.0457.260−0.030
1911,90130,14657,9536.9136.949−0.005
2014,93130,91254,1576.8476.7060.021
3046,29521,86031,8455.5834.4660.250
4066,62110,44122,9384.3103.1810.355
5077,156459418,2503.3702.4040.402
6082,770196415,2662.6881.8960.417
7086,06088713,0532.2061.5590.415
8088,26341211,3251.8571.3160.411
9089,78419110,0251.5931.1370.401
10090,8929290161.3911.0010.390
20095,497045030.6070.4520.341
30096,976030240.3870.2920.327
40097,762022380.2850.2150.323
50098,232017680.2250.1710.319
60098,530014700.1860.1420.317
70098,714012860.1590.1210.315
80098,867011330.1380.1050.314
90098,991010090.1220.0930.314
100099,10508950.1100.0840.314
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gyarmati, L.; Mihálykó, C.; Orbán-Mihálykó, É. Algorithm for Option Number Selection in Stochastic Paired Comparison Models. Algorithms 2024, 17, 410. https://doi.org/10.3390/a17090410

AMA Style

Gyarmati L, Mihálykó C, Orbán-Mihálykó É. Algorithm for Option Number Selection in Stochastic Paired Comparison Models. Algorithms. 2024; 17(9):410. https://doi.org/10.3390/a17090410

Chicago/Turabian Style

Gyarmati, László, Csaba Mihálykó, and Éva Orbán-Mihálykó. 2024. "Algorithm for Option Number Selection in Stochastic Paired Comparison Models" Algorithms 17, no. 9: 410. https://doi.org/10.3390/a17090410

APA Style

Gyarmati, L., Mihálykó, C., & Orbán-Mihálykó, É. (2024). Algorithm for Option Number Selection in Stochastic Paired Comparison Models. Algorithms, 17(9), 410. https://doi.org/10.3390/a17090410

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop