1. Introduction
Group decision making (GDM) is a widely researched topic with a plethora of items in the literature on decision-making processes in groups of various sizes. The analytic hierarchy process (AHP) [
1] is a commonly used multi-criteria method to support GDM processes in practice. On a contextual side, the group decision can be questionable if it does not include all relevant items and/or if the methodology for deriving the decision is inadequate. If the procedure did not include all relevant participants, the decision can be criticized by another group(s).
This paper focuses on developing a method for assessing the quality of experts who act as decision makers. One of the major challenges is determining how many experts should be involved in making group decisions and which instruments to use to evaluate their consistency while judging decision elements. The other important issue is to measure the deviations of individual decisions from the group decision, once the latter is derived by some method of aggregation, consensual or non-consensual. Consistency and compatibility measures may help to determine the optimal size of the group of experts and make the process of decision making efficient. Worth mentioning is that conducting a thorough analysis of consistency and compatibility measures represented by specific indicators is partly contingent on the specific problem and contributes to efficiently clustering decision makers. The clustering process, based on selected distance measures as parts of the fitness function in the genetic algorithm, can only indicate the optimal number and size of clusters, subsequently guiding further research in group decision-making processes. The literature review did not uncover research findings regarding the identification of the ideal group size. Any methodological approach to grouping may face skepticism and criticism. For instance, assigning weights to decision makers within aggregation schemes could pose challenges for several reasons, rendering it an inadvisable element of grouping methodology.
The proposed method to rank experts based on their consistency and compatibility with the group decision is based on the combined use of the analytic hierarchy process (AHP) [
1] and the technique for order of preference by similarity to the ideal solution (TOPSIS) [
2]. To derive the group decision, first, the AHP has to be used in its standard, crisp, version to obtain individual priority vectors by experts and then to geometrically aggregate these vectors into the group vector. Computations of consistency and deviation parameters based on crisp AHP results enable the assessment and ranking of the experts by their quality of decision-making performance using the TOPSIS method. The approach involves creating a decision matrix that contains specific performance indicators used as the criteria for evaluating experts. The CRITIC method [
3] is proposed to determine objective weights for these performance indicators resulting from the crisp AHP application. Subsequently, the TOPSIS method enables the identification of possible subgroups of experts, determining outliers, and deriving optimal solutions. Finally, the use of a genetic algorithm may allow for mapping experts in clusters according to demonstrated performance.
It is important to note that, after the group members have established their priority vectors using the AHP, the organizer of the decision-making process has several options for determining the outcome—the group priority vector. One approach is to use the priority vectors derived by each individual and aggregate them to create a group priority vector. Another approach is to perform an adjacency assessment of the individual vectors, grouping members based on the distances among their vectors, and deriving sub-group priority vectors. This approach may result in a different group vector than if all individuals belong to a single group. Additionally, this assessment can identify members who significantly deviate from the group decision, leading to their exclusion from the process or assigning low weight to their individual decisions. Managing the process with outliers in the group is a complex issue that requires careful consideration of both subjective and objective factors.
Moreover, group decision making poses significant challenges due to the diverse backgrounds, attitudes, communication skills, and adaptability of its members. The behavior of decision makers throughout the process is crucial, as it can be prolonged and repetitive, requiring constant adjustments to judgments based on new information. During this process, decision makers must demonstrate their ability to assess causality, importance, preferences, and goals while considering the available data and any limitations or constraints.
Measuring the quality of decision-making performance within a group is a delicate task, involving subjective characteristics such as cognition, reasoning, inference, and deduction. Aggregating these aspects into a reliable judgment outcome can be difficult, leading to errors and complications in an already complex decision-making process.
Given the aforementioned challenges, the use of the rough numbers theory [
4,
5] and the application of a rough version of the AHP are proposed to address individual judgments directly. This involves creating and manipulating a sequence of judgments for each instance of comparing two decision elements by importance, using rough rules instead of the traditional aggregation methods like geometric or additive averages.
By utilizing both crisp and rough AHP vectors, valuable insights are gained into the objective evaluation of the experts’ performance. This combined approach proves to be a useful tool for assessing their decision-making capabilities. It allows them to consider not only the precise well-defined aspects of their judgments (standard AHP) but also the inherent uncertainty and imprecision in their assessments (rough AHP).
By incorporating rough numbers theory, accounting for the vagueness and ambiguity that often exist in decision-making scenarios is acknowledged. This empowers handling individual judgments more realistically and flexibly, considering the nuances and complexities that can arise during the decision-making process.
The use of rough AHP in handling individual judgments directly, rather than relying solely on traditional aggregation techniques, enables a more nuanced and adaptive analysis. This approach ensures that each decision element’s importance is properly assessed with consideration of the available data, leading to a comprehensive evaluation of the experts’ performance.
In summary, the integration of crisp and rough AHP offers a comprehensive framework for evaluating decision-making performance. By acknowledging uncertainty and employing a more flexible analysis, this complementary approach provides a richer understanding of experts’ abilities, making it a valuable tool in the assessment process.
This study presents a novel approach for restructuring decision-making groups by clustering experts using a combination of genetic algorithm and minimum distance rule based on their consistency and agreement with the group. The proposed method utilizes performance indicators that assess individual consistency and agreement with the group decision, enabling the identification of sub-groups of experts and potential outliers that may be excluded from future decision-making processes. For an illustration of how GDM challenges can be addressed in practice, an example method application is presented for a group of twelve decision makers from diverse sectors, educational backgrounds, and professional expertise who participated in evaluating measures on how to reduce drought risks in managing wetland in floodplains along the Danube River in Serbia.
2. Related Work on Group Decision-Making
2.1. Approaches and Methods
According to [
6,
7], multi-factor decision-making frameworks can be classified as group decision making, negotiated decision making, and systemic decision making. In recent decades other multi-factor settings also received the attention of researchers and professionals in business, industry, academia, etc. For instance, consensual decision making enables the integration of opinions from involved experts, enabling, thus, a better understanding of decision-related problems and a more efficient collection of individual viewpoints in the search for agreed solutions by consensus.
The literature sources offer a comprehensive understanding of existing approaches, methods, and methodologies across all the aforementioned group decision-making settings. It is noteworthy to mention several published research results in this context. Du and Shan [
8] discuss the issue of possible large differences in the knowledge and educational background of participants in large groups when an accurate evaluation of the criteria set is unlikely to be expected. An evaluating system is proposed to assess input criteria and detect the critical set of so-called output criteria based on a probability assignment function as an information extraction method aimed at capturing and accurately reflecting the authenticity of experts’ judgments. The authors claim that the method can effectively help in the real-time updating of ideas and the screening of best ideas in an intelligence recommendation framework. Aggregation or consensus models for combining individual ideas into group ideas and measures on how to validate such models’ quality are discussed to only a limited extent.
Interesting research is presented in [
9] related to assessing the team roles of participants in the group. An AI approach is developed to enable the creation of efficient and consistent teammates by considering their roles in the group. Exploratory and confirmatory factor analysis is involved to assess the effects of changes in team composition when deriving joint decisions. Based on expert interviews and pre-defined four team roles, it was shown that the consistent team roles are identified as coordinator, creator, perfectionist, and doer. This study is interesting because it relates to our research in part on grouping experts based on their consistency and deviation (from the group) performance indicators.
Dayeh and Morrison [
10] claim the importance of properly using the hidden-profile paradigm, which is commonly understood as a research design in which team members have information that should be shared to arrive at an accurate final (group) decision. Namely, the authors rely on various research and conclusions that group members usually fail to exploit their information in group settings. Their study explored different situations in an attempt to determine how each team member’s perception of competence, relative to other team members, influences information sharing and decision accuracy in hidden profiles. In conclusion, it is indicated that individuals (within a group) who perceive themselves as less competent are willing to share more information than others. Also, regarding decision accuracy, the conclusion is that it is better in a cooperative environment.
Escobar and Moreno-Jimenez [
6] presented a novel approach for the aggregation of individual decisions in the group, named aggregation of individual preference structures (AIPS). An approach is strictly oriented to the AHP method, incorporates ideas from the social choice theory method and the Borda count and, in a way, is a continuation of well-known aggregations in AHP-group decision making known as AIP and AIJ. The AIPS method captures the uncertainties inherent in human beings; the interdependencies between the alternatives and the preferences are associated by each decision maker to these interdependencies. The approach is proposed to be situated in the initial phases of the decision-making process and as a tool to support the negotiation process within the decision-making group.
Zorluoğlu and Kabak [
11] structured a specific project portfolio-selection problem as a hierarchical group decision-making problem and applied it to the automotive sector. An idea behind the proposed information-technology-business-oriented approach is to involve, as much as possible, information from a large number of employees in the organization and, in a way, suppress the influence or dominance of one decision maker, whoever he or she is. Also, the aim is to reduce biases and the irrelevant evaluation of decision elements, including ‘noise’ information while judging decision elements (in this case, criteria). A weighted cumulative belief degree approach is proposed for aggregating the evaluations and the final criteria weights determined by participants in the group decision-making process. This concept is, in a way, similar to the one known in consensus-reaching methodologies (e.g., [
12,
13,
14]).
2.2. Analytic Hierarchy Process
In the standard version of AHP, at a given level of the hierarchy, each individual creates square matrix
A by comparing decision elements. The pair-wise comparison principle is applied by using Saaty’s scale with nine levels of preferences or 17 levels of comparison (when both parts of the scale are taken into account, linear 1–9 and nonlinear 1/9–1/2),
Table 1.
When a group consists of
M individuals, each member of the group creates a pairwise comparison matrix of type (1). From each matrix, a weight vector
w needs to be extracted using one of the existing methods. Weights provide cardinal information about the significance of compared elements, and based on them, the ranks of elements are determined as ordinal information. There are matrix-based and optimization-based methods commonly used to calculate the weight vector
w from the matrix (1), collectively referred to as AHP prioritization methods. The most used are the eigenvector method, additive normalization method, weighted least squares method, logarithmic least squares method, logarithmic goal programming, fuzzy preference programming method, and cosine maximization method [
15]. Ongoing scientific discussions exist regarding which method is superior, and various studies have compared the methods mentioned to determine their effectiveness.
The consensus is that no method can be given a priori preference. In this particular paper, the eigenvector method was employed as a key component of the standard AHP method. The prioritization of elements employed in a matrix (1) by the eigenvector method produces the weight vector
= (
, …,
) after the linear system (2) is solved. In relation (2),
λ represents the principal eigenvalue of the matrix, and
e is the unit vector
= (1, 1, …, 1) of dimension
n.
If an individual is completely consistent, meaning the transitive rule holds for every i, j, and k from the set of values (1, 2, …, n), then λ = n. Otherwise, λ > n. The maximum eigenvalue (λmax) for an inconsistent matrix can be estimated by successive squaring matrix A. When squaring is performed, the elements are summed by rows, and the sums are normalized to be equal in sum. Thus, approximations of the desired vector are obtained, and the procedure is terminated when the difference between two successive vectors is less than a defined value.
In group applications of crisp AHP, the most commonly used aggregation procedure is the AIP, the aggregation of individual priorities. Another option for aggregation is to apply procedure AIJ, the aggregation of individual judgments at each entry of matrix
A, and then, continue with prioritization, for instance, by the eigenvector method. More details on AIP and AIJ can be found in [
16].
To check the consistency of pairwise comparisons and the quality of the results obtained by the standard, crisp, version of AHP, the consistency ratio (CR) and Euclidean distance (ED) are commonly used as measures of judgment quality demonstrated by an individual.
The consistency ratio (
CR) is calculated during the standard AHP procedure. First, the consistency index (
CI) is calculated by relation (3)
where
is the maximum eigenvalue of the matrix
A. Then, using this index and the random index (
RI), which depends on the order of the matrix, the consistency ratio is obtained by relation (4).
The value of
RI for various matrix orders is determined statistically using samples of 500 and 1000 randomly generated comparison matrices [
1]. Researchers have shown that, for small deviations from consistency ratios
wi/wj, the eigenvector method provides a satisfactory approximation of the priority vector
w. The tolerance value for the consistency degree is 0.10. If
CR is less than 0.10, the decision-making process presents acceptable inconsistency. In some cases, this statement holds even if
CR is greater than 0.10. However, adjustments may need to be made to the evaluations until a small enough value of
CR is achieved.
The Euclidean distance (
ED), also referred to as the total deviation, represents the distance measured between all elements in a comparison matrix (1) and the related ratios of the weights of the derived priority vector, as given in relation (5).
This consistency measure is a universal error measure and is, therefore, invariant to the prioritization method (including the eigenvector method) used for deriving vector w.
2.3. Rough Analytic Hierarchy Process
Rough AHP is the result of applying the theory of rough sets to the original AHP, similar to how it was done with the application of fuzzy theory to AHP. The theory of rough sets was presented in the early 1980s [
4], and numerous authors have contributed to its further development and elaboration through applications (e.g., [
17,
18]). Like fuzzy theory, the rough set theory deals with uncertainties that exist in a real environment. The idea is to establish approximate boundaries for given sets of numbers. There are no parameters that indirectly treat vagueness, and the given data structure speaks for itself [
18].
In the rough version of AHP, the formation of the group comparison matrix is done by first creating a sequence of ratings on each position of matrix
A obtained from
M members of the group:
aijg = {
aij1,
aij2, …,
aijM}. Then, all elements of the sequence are translated into rough numbers
RN(
aijm) = [
aijmL,
aijmU] (
m = 1,…,
M) using the method defined in [
19,
20]. The superscript notations
L and
U indicate the lower and upper bounds of the rough number
RN.
The rough sequence at the given position (
i,
j) in the matrix is given as:
By conversion, the sequence is transformed into its average rough number:
where:
The rough group weights are calculated using rough numbers (8) and (9) according to relation (10).
Averaging the values for the lower (
L) and upper (
U) bounds of the rough group weights gives:
The final normalization of values calculated using relation (11) yields the weights (12) that can be compared with the weights from the crisp version of the AHP method.
The basic mathematical operators on two rough numbers RN(a) = and RN(b) = are as follows:
Addition (+): RN(a) + RN(b) =
Subtraction (−): RN(a) RN(b) =
Multiplication (∙): RN(a) ∙ RN(b) =
Division (/): RN(a)/RN(b) = (assuming bL and bU are non-zero).
Scalar multiplication is also used, m∙RN(a) = , where m is a scalar and a is a rough number. More complex operators are performed using the above-mentioned rules based on the theory and principles of rough numbers.
Notice that rough weights (12), obtained by the rough version of AHP, assume that individual judgments at each position of the joint matrix (1) created the sequences from which the corresponding lower and upper limits are computed. This way, the data (judgments) ‘spoke by themselves’ instead of any aggregation, such as, for instance, by the AIJ procedure [
16]. It is worth mentioning that rough AHP does not handle issues of consistency or deviations of individuals’ judgments.
2.4. TOPSIS—Technique for Order Preference by Similarity to Ideal Solution
The TOPSIS method was originally developed by Hwang and Yoon [
2] and further improved by Yoon [
21] and Hwang et al. [
22]. The method is based on the concept that the chosen alternative should have the shortest geometric distance from the ideal solution and the longest geometric distance from the anti-ideal solution. In other words, the method treats the distances between alternatives and non-existent, so-called ‘ideal’ alternatives. It first determines the ‘ideal’ and ‘anti-ideal’ alternatives and calculates the Euclidean distances of actual alternatives from these ‘ideal points’. The best alternative is the one with the smallest distance from the ideal point, and the method guarantees that its distance from the anti-ideal point is the maximum.
The multi-criteria analysis with TOPSIS is characterized by a decision matrix, commonly called a rating matrix
R. The rows of the matrix correspond to the alternatives, while the columns represent the criteria. Each element of the matrix represents the rating of a particular alternative concerning a particular criterion. For
m criteria (
C1,
C2,…,
Cm) and
n alternatives (
A1,
A2,…,
An), the matrix
R has the form (13). The values (
w1,
w2,…,
wm) written above the columns represent the weights of the criteria defined by the decision maker or determined in some other way; the sum of these weights is one.
TOPSIS calculates a score for each alternative by comparing it to both the ideal and anti-ideal solutions, represented by the best and worst values (rijb, rijw) for each criterion in the matrix (13). The method evaluates the distance between each alternative and these two solutions using a chosen distance metric, often the Euclidean distance (rarely the Manhattan distance). The relative closeness of each alternative to the ideal and anti-ideal solution is determined by the ratio of the distance to the anti-ideal solution to the sum of distances to both the ideal and anti-ideal solutions. A higher relative closeness indicates a better performance relative to the other alternatives.
The method provides a systematic and effective way to handle complex decision-making scenarios with multiple criteria, allowing decision makers and/or stakeholders to make informed choices and enhance their decision-making processes [
23].
According to the review [
24], the TOPSIS method is a simple process to implement, use, and program. The algorithm has six basic steps, and the number of steps remains the same regardless of the number of decision elements, such as attributes. A disadvantage of the method is that it uses Euclidean distance, which does not consider the correlation of attributes; moreover, it is difficult to weigh and maintain the consistency of judgment. Many applications of this method are recorded in the literature related to environmental modeling and management, water resources management, engineering, business, and marketing (e.g., [
25,
26]).
2.5. CRITIC—Method for Objective Weighting Criteria
In decision-making problems, evaluating criteria before evaluating alternatives can be difficult due to subjective, incompetent, or inconsistent human judgment. Additionally, criteria can be quantitative (price and profit), qualitative (appearance and impression), or ‘gray’ (average value and availability) in nature, which further complicates the decision-making process. As proposed for the first time in the work of Doyle [
27], the decision-maker can be eliminated to a certain extent, allowing the “alternatives to decide for themselves” on the importance of criteria. The idea is to directly analyze the decision matrix and, by exploiting information on the performance of all alternatives relative to all criteria, to derive the weights of criteria.
To address these issues, the CRITIC (criteria importance through inter-criteria correlation) method was proposed in [
3]. This method uses inter-criteria correlation analysis to determine the importance of criteria and involves calculating the correlation between each pair of criteria to derive weights for each criterion. Once the criteria weights are determined, alternatives can be ranked based on their performance concerning the criteria.
The method is particularly useful in decision-making problems where there are multiple criteria, and their relative importance is unclear or subjective. The method’s objective nature is based on the statistical processing of information contained in the decision matrix. The method tends to “smooth” the overall statistics of the ratings, even when they differ significantly from each other.
The paper by Androulakis and Chatzidimitriou [
28] provides a detailed description of the CRITIC method and includes examples of its application in various decision-making problems. Examples of the application of the method are also given in [
29,
30].
2.6. Clustering Individuals within a Group by Using Genetic Algorithm
The values of performance indicators related to consistency and deviations of individuals in the group are easy to imagine as coordinates of individuals in a multi-dimensional space. If the problem is to create a given number of clusters of individuals, certain criteria for doing that are required. Such criterion can be the shortest distance of points (individuals) from the central point of the given cluster, and an appropriately structured genetic algorithm can perform clustering for a given number of points.
Figure 1 illustrates the basic computations required to determine the ‘fitness’ of a given cluster with five points in 2D (x-y) space. With known coordinates of points B, C, D, E, and F in a cluster, the sum of distances of these points from central point A is the fitness of this cluster. Once the coordinates of central point A in each generated cluster are determined, all distances of points in all clusters are easy to calculate, and their sums are cluster-related (local) fitness. The genetic algorithm through an iterative generation process searches for the best positioning of points (individuals) in a given number of clusters. The best solution is the one with the lowest value of the sum of local distances of the points positioned in all clusters.
There are other possible ways for clustering points in two-dimensional space, such as the use of the L2 (Euclidean) metric, L1 (Manhattan) metric, or Tchebysheff distances. The L2 metrics work well for evenly distributed points, while L1 metrics might be better suited for points clustered near axes. The Tchebycheff distance might be used within the fitness function in genetic algorithms if solutions should be evaluated in a more spatially oriented way, especially in multi-dimensional spaces. This distance calculation is also known as the ‘maximum metric’ because it measures the distance between two points as the maximum difference over any of their axis values.
The triangles used in our genetic algorithm are suitable for problems where the fitness is assessed based on geometric relationship or similarity. When proportional triangles are used as illustrated in
Figure 1, the coordinates of the central points in the triangles are multiplied by the corresponding areas of triangles (P1, P2, and P3), and by dividing their sum by the total area P, the coordinates of cluster center A are obtained. When there are more points in the cluster, the number of triangles increases, but the concept of proportional triangles created for a given set of points is the same.
In the presented study, two common consistency indicators are used, CR for the eigenvector prioritization method in AHP and ED for total Euclidean distance—universal L2 metrics. The other two indicators (CO and SC) are related to the compatibility of individuals with the group. Normally, the number and type of indicators can be different and problem related.
To illustrate the concept, these four indicators are selected as sufficient to facilitate analysis of experts’ behavior and cluster them in a predetermined number of clusters based on, separately, their consistency and their compatibility. Corresponding two-dimensional diagrams, (CR-ED) and (CO-SC), enable positioning each member of the group as a point in a related scatter diagram. For the 12-member group in our study, considering two or three clusters in either diagram appeared to be a reasonable approach. However, drawing a definitive general conclusion from this methodological step is inherently contentious, as its appropriateness depends significantly on the nuances of the particular decision-making scenario.
4. The Proposed Method
The AHP-based approach serves as the fundamental method for solving individual decision-making problems. As the context shifts towards group decision making, individuals are evaluated on the quality of their decision-making performance, and their consistency is measured using multiple performance indicators. Compatibility indicators represent how closely an individual’s decision aligns with the group’s decision. To determine the group’s consensus, individual outcomes can be aggregated through methods such as additive or geometric aggregation. The TOPSIS method can be used to evaluate and rank each member’s consistency and compatibility within the newly established multi-criteria framework. The CRITIC method can be applied as an objective instrument to determine the weights of the performance indicators necessary for the TOPSIS method. Regarding group settings, in addition to standard AHP, based on using crisp numbers as equivalents to formal linguistic judgments, the rough version of AHP can be added to the analyses to enable a comparison of the aggregation schemes used in the standard and rough AHP versions. Choosing to include this addition, the rough version of AHP, is not mandatory; indeed, the standard AHP can be substituted with its rough version. Additionally, considering alternative variants of AHP, such as hesitant AHP, fuzzy AHP, or employing stochastic approaches and Bayesian analysis, is methodologically feasible. The original and rough versions of AHP used in our study are not only applicable but also sufficiently simple and beneficial for gaining deeper insights into the decision-making process. To summarize, for group decision-making settings where diverse judgments commonly occur, the methodology proposed below complements the evaluation of experts’ performance quality based on only crisp values, with the rough theory that, instead of aggregating priorities of decision elements derived by standard AHP, uses sequences of original judgments in a special way to derive group priorities according to rough theory.An approach to the following problem is proposed. At an early stage of the decision-making process, it is important to assess and evaluate the quality of involved experts in making decisions by a formal evaluation of their consistencies and agreement with the rest of the group. A methodology is proposed based on a combination of crisp and rough AHP, TOPSIS, and CRITIC methods. All the methods mentioned are thoroughly established and scientifically validated. In the context of group decision-making scenarios, their implementation is straightforward. In the final stage of the proposed methodology, clustering individuals based on demonstrated consistencies and similarities with the group is suggested to be achieved through the utilization of a genetic algorithm, known for its efficiency as a stochastic search engine. Alternatively, other heuristics could be considered, such as simulated annealing or evolutionary strategy, for instance.
The proposed method is organized in six steps, as shown in
Figure 2, and represents an extended procedure published in [
38]).
Step #1: In this initial step, each member of the group solves the given decision-making problem using the AHP method, creating a hierarchy with the goal on the top and at the level below decision elements to be evaluated concerning a goal, such as sub-goals, criteria, or alternatives. The AHP-based prioritization is used to obtain individual priority vectors of decision elements concerning the goal. During this process, the consistency of the members is measured, and individual indicators of their consistency are recorded.
If a one-level hierarchy is created with a goal at the top and selected decision elements (sub-goals, criteria, or alternatives) at a level below, the term “AHP-based” refers to an incomplete AHP, as there is no synthesis involved. In this case, only prioritization of the decision elements versus the goal takes place, and for each member of the group, consistency can be measured based on the derived weights and original judgments in his/her comparison matrix. The consistency ratio (
CR) and total Euclidean distance (
ED) are commonly used for this purpose.
CR is established exclusively for the AHP and eigenvalue prioritization method, while the
ED is the universal error measure calculated as the total distance between all judgments
aij in the comparison matrix
A and the related ratios of weights (
wi/wj) contained in the vector of weights derived during prioritization. The
ED consistency measure can be used with any prioritization method to assess consistency, as reported by many researchers, e.g., [
39,
40,
41,
42].
Step #2: In this step, individual priority vectors are synthesized into the group priority vector by the AIP method (an acronym for aggregation of individual priorities) [
16]. Then, deviation, correlation, and other indicators of individual agreement and/or disagreements with the rest of the group are computed. These indicators are recorded as group-related information on the performance quality of individuals. In addition, a rough version of AHP manages sequences of judgments in each position in the joint comparison matrix, and determines the priority vector for the group that can be directly compared with the AIP vector obtained by Equation (14)
where
M represents the number of members in the group,
represents the priority of the
i-th alternative for the
j-th member,
αj represents the weight of the
j-th member, and
wiG represents the aggregated group weight. The weights
αj should be additively normalized before being used in Equation (14), and the final additive normalization of priorities
wiG is required.
The group vector
wG (
) can be used as the reference vector for the members of the group to compute deviations of their vectors from the group vector. The term ‘conformity’ or ‘compatibility’ (CO) is commonly used to express individual deviations and is synonymous with the Manhattan distance. It is illustrated by Equation (15)
where the superscript
G represents the reference priority vector obtained by aggregation.
CO indicates the overall similarity between the individual priority vector and the reference group vector. This performance indicator only applies after all computations in AHP are concluded, which is different from using consistency indicators
CR and
ED.
Different statistical measures can also be used to measure individual agreements or disagreements with the rest of the group. One of the more used measures is Spearman’s rank correlation coefficient (
SC) calculated by using Equation (16).
This measure enables comparing the ranks of the corresponding elements of the individual priority vectors and the reference vector for the group (Di stands for individual vector). SC values range from −1 to 1, where a value of −1 indicates an ideal negative correlation, +1 shows an ideal positive correlation, and a value of 0 indicates no correlation.
In practice, SC is a reliable statistical measure, even for small groups. It is a relative measure and not absolute, indicating that two vectors can vary widely in rank preference while remaining relatively close in absolute preference. With larger groups, potential misguidance is typically minimized. In the proposed multi-model framework, SC is elaborated as an ordinal information measure that can be combined with cardinal information (CO). The combination of CO and SC facilitates efficient clustering, allowing the exploration of both cardinal and ordinal information simultaneously.
After the completion of prioritization in AHP for all members in a group, it is possible to calculate SC as analogous to the compatibility indicator CO.
Step #3: In this step, a decision matrix is created with rows representing members of the group and columns representing the performance indicators identified in steps #1 and #2. The matrix entries in each row correspond to given member scores regarding the performance indicators of consistency and deviation. The decision matrix is flexible in size, accommodating any number of columns for performance indicators (criteria) and rows for members (alternatives). Employing cross-referencing and visualization techniques aids in the matrix’s preparation, fostering insights and comprehension of its elements’ impacts.
The entries of the matrix are performance indicators of members as represented by Expression (17). To complete the decision matrix, weights
w1,
w2,
w3, and
w4 must be assigned to the ‘criteria’ (
CR,
ED,
CO, and
SC), with the sum of the weights totaling one.
Step #4: In this step, the TOPSIS method is utilized to assess the decision matrix, ranking group members based on their overall performance concerning consistency and alignment with the group. Before employing the TOPSIS method, it is imperative to define the weights related to the matrix (17) for all performance indicators. The CRITIC method is recommended for this task, which is particularly beneficial for larger groups.
It is worth noting that before applying multi-criteria methods, the analyst may subjectively assign weights to performance indicators or opt for an objective method. One alternative to the statistical method CRITIC is the application of the entropy principle to determine objective weights for indicators. As per Shannon and Weaver [
43], entropy serves as a gauge of information uncertainty, indicating the “strength” of decision elements (such as criteria) in communicating a message to the decision maker. The crux lies in addressing uncertainty within the information framework of the decision matrix, known as Shannon entropy. The concept of entropy has been used in various areas of multi-criteria optimization with good results, e.g., [
15,
23,
29,
44]. In an example provided in the subsequent section, the entropy method could not be applied due to a negative Spearman correlation coefficient identified for one member within a group, rendering logarithmic operations impossible within the entropy method. Given the potential recurrence of such cases in other scenarios, the CRITIC method is employed to ascertain the objective weights of performance indicators.
Step #5: During this stage, a sensitivity analysis can be conducted by assigning varying weights to performance indicators to cluster or regroup individuals based on their consistency and deviation from the group solution. The iterative application of TOPSIS, depending on the weighting scheme for performance indicators, facilitates the assessment of group members’ performance quality, with a focus on demonstrated consistency in judging the initial decision-making problem or compatibility and statistical agreement with the rest of the group. In the provided example, different preference schemes are used to weight groups of performance parameters (consistency, CR and ED; deviation, CO and SC) and to explore opportunities for sub-grouping members. Sensitivity analysis aids in enhancing the quality of the decision-making process by identifying and mitigating potential pitfalls, such as avoiding subpar performances from influencing further decision-making. It serves to validate the performance quality of each group member, thereby improving the overall decision-making process.
Step #6: In this last step, the outcome of the process is analyzed, and the competency of members of the group to further get involved in the decision-making process is decided. If necessary, some previous steps may be repeated by sub-grouping members, eliminating odd members, etc.
This step requires additional consideration of the results of steps #4 and #5 because they may directly determine further steps in group assessments of decision elements. The group can be restructured, with new members involved and/or odd members excluded. If decided upon, new decision elements can be added and/or old elements excluded, for instance, if the initial evaluation process indicated sharp differences between superior and inferior decision elements. However, any change in the structuring decision process must respect restrictions related to the duration of the entire process. Important issues to consider include the number and eligibility of members in the (restructured) group and the number of decision elements to be assessed. The described procedure outlined in steps #1-#6 can be easily generalized by incorporating or replacing performance indicators before the final assessment of the quality of the group members. Different combinations of multi-criteria methods can be employed to obtain individual and group solutions and perform the final evaluation based on demonstrated consistencies and deviations. The results can be visualized using scatter diagrams, such as CR-ED and CO-SC, and a genetic algorithm can be applied to identify decision-maker clusters and the consequences of sub-grouping individuals rather than treating all members as part of a single group.
According to Srdjevic et al. [
38], when using AHP for individual assessments, it is important to note that the consistency parameter
CR should only be utilized if the prioritization of decision elements is carried out using the additive normalization (AN) or eigenvector (EV) method, as noted by Saaty [
1], Srđević [
42] and Kou and Lin [
39].If the LLS prioritization method is being used, parameter
CR should be replaced by
GCI, as suggested in [
45,
46,
47,
48]. Similarly, the FPP method proposed in [
37] requires the use of the parameter
μ, while the CMM method recommended in [
27] uses the
CCI cosine consistency index.