Next Article in Journal
Study on the Influence of Air Tightness of the Building Envelope on Indoor Particle Concentration
Next Article in Special Issue
An ANP-Balanced Scorecard Methodology to Quantify the Impact of TQM Elements on Organisational Strategic Sustainable Development: Application to an Oil Firm
Previous Article in Journal
Thoroughbred Racehorse Welfare through the Lens of ‘Social License to Operate—With an Emphasis on a U.S. Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrating Sustainability in the Quality Assessment of EHEA Institutions: A Hybrid FDEMATEL-ANP-FIS Model

Gijon Polytechnic School of Engineering, Department of Business Administration, University of Oviedo, 33003 Oviedo, Spain
*
Author to whom correspondence should be addressed.
Sustainability 2020, 12(5), 1707; https://doi.org/10.3390/su12051707
Submission received: 23 December 2019 / Revised: 17 February 2020 / Accepted: 19 February 2020 / Published: 25 February 2020
(This article belongs to the Special Issue Synergies between Quality Management and Sustainable Development)

Abstract

:
This paper proposes the design of a conceptual model of quality assessment in European higher education institutions (HEIs) that takes into account some of the critical reflections made by certain authors in the literature regarding standards and guidelines suggested for this purpose by the European Higher Education Area (EHEA). In addition, the evaluation of the conceptual model was carried out by means of the reliable hybrid methodology MCDM-FIS (multicriteria decision making approach–fuzzy inference system) using FDEMATEL and FDANP methods (fuzzy decision-making trial and evaluation laboratory and FDEMATEL-based analytic network process). The choice of these methodologies was justified by the existing interrelationships among the criteria and dimensions of the model and the degree of subjectivity inherent in its evaluation processes. Finally, it is suggested to include sustainability as a determining factor in the university context due to its great relevance in the training of future professionals.

1. Introduction

The concept of quality in European higher education is not new. Since the 1980s, national initiatives have been adopted to systematise evaluations and promote quality improvements [1,2]. The Sorbonne Declaration (1998), signed by the education ministers of four countries (Italy, France, Great Britain and Germany), was the prelude to the Bologna Declaration (1999), initially signed by representatives of 29 countries and subsequently extended to include 47 participants.
The Bologna Declaration laid the foundation for the construction and implementation of the so-called ‘European Higher Education Area (EHEA)’, organised around four principles: quality, mobility, diversity and competitiveness. The European convergence in education aimed to promote the mobility of students, teachers and researchers while ensuring the quality of higher education, in order to ultimately increase employment in the European Union and the socioeconomic progress of the territory (http://www.eees.es/).
The objectives set by the EHEA were fundamentally articulated around the document ‘Standards and guidelines for quality assurance in the EHEA’ (ESG), with their monitoring and improvement discussed in subsequent meetings (Prague (2001), Berlin (2003), Bergen (2005), London (2007), Lovaina (2009), Budapest and Vienna (2010), Bucharest (2012) and Yerevan (2015)). It is important to highlight that these criteria are indicative of areas considered critical for high-quality and successful learning environments for higher education.
The criteria outlined in the ESG are classified under the following three headings: (a) internal quality assurance in higher education institutions (where 10 criteria are listed); (b) external quality assurance in higher education (including 7 criteria); and (c) quality assurance of external quality assurance agencies (where another 7 are listed). All of the criteria are, in some way, interrelated.
Before going deeper into our study, it is appropriate to point out some reflections drawn from the literature regarding these criteria. Several authors [3,4] have pointed out three basic pillars in a university quality system: education, research and social impact. The objectives proposed for an educational system at any level (Bachelor, Master and/or PhD) will try to improve its quality and to demonstrate compliance and excellence to stakeholders [5]. This purpose requires evaluating and, if necessary, modifying those three key aspects at the unit, institutional and social level. However, as Brdulak [4] has already noted, most of the ESG criteria are widely oriented to assess and guarantee quality in the educational area, only one of the three missions assigned to the institution.
The importance of research is mentioned by the ESG standards (i.e., “The focus of the ESG is on quality assurance related to learning and teaching in higher education, including the learning environment and the relevant links for research and innovation” [6]. However, explicit guidelines are not articulated in this regard. Something similar occurs in relation to the third pillar—social impact—where the stakeholders are entities that affect or are affected by the activities of an organisation and are interested in their proper performance [7,8]. At the university level, stakeholders include a wide variety of actors from the academic community, government, students, industry, business, politics, the public sector and the general public [9]. The document mentions higher education institutions’ (HEIs’) commitment to community and industry stakeholders to recognise the role of these institutions “in supporting social cohesion, economic growth and global competitiveness”. However, in this case, the criteria also lack the emphasis and detail reflected in the educational mission.
Apart from reflecting on the suitability of the ESG criteria to describe the three pillars of university institutions, it is necessary to consider other important issues that hinder the task of evaluating educational systems. On one hand, as indicated above, the ESG criteria are merely indicative. In fact, each country can adopt different local policies for accreditation and opt to focus on different aspects. As evidence, the criteria analysed by 15 agencies were consulted and compared. Ten of these agencies belong to the European Network for Accreditation of Engineering Education (ENAEE)—responsible for evaluating quality assurance and accreditation agencies in the EHEA in respect to their standards and procedures when accrediting engineering degrees—and five of the agencies belong to the European Registry of Quality Assurance for Higher Education (EQAR). From the comparative analysis performed, a clear consensus was detected on only 4 of the 20 factors collected to carry out the quality assessment (curriculum content; student performance and training results; material, human and financial resources; monitoring and dissemination of quality improvement measures). The other factors were generally used in a single country. On the other hand, there are different types of audits and evaluations carried out in institutions of higher education: internal audits, carried out by personnel of the assessed institutions, and external audits, carried out by teams of external experts, peers or inspectors. This variability in the formats, quality frameworks or models can generate a certain degree of scepticism about the validity of the principles of quality assessment when applied in HEIs [10]. In addition, the evaluation criteria do not focus on isolated and perfectly defined aspects; therefore, it is necessary to analyse the influence that some criteria may have on others.
Another aspect to consider is the different perception of the concept of quality by professionals, academic staff and students [11], which introduces a component of relevant subjectivity in the evaluation. This fact could be due to the elusive and multi-edge character of quality in education, as different authors have already observed, describing it as a multidimensional, multilevel and dynamic concept [12,13,14,15]. This multidimensional character includes all of the functions and activities that can be evaluated, i.e., teaching and academic programmes, research and scholarship, staff, students, facilities, equipment and services for the community and the university world, reconciling the interests of different stakeholders, national cultures and political contexts—too many dimensions to easily obtain a standardised concept of quality in HEIs.
The above reflections give an indication of the complex nature inherent in the design and implementation of a system of evaluation, control and continuous improvement of quality, not only in the teaching process but also in the whole university system, and point to the need to use a methodology capable of covering, as far as possible, the multiple facets that must be considered in the search for a solution. In this paper, two contributions are proposed: (a) the design of a conceptual model for quality assurance in European HEIs; and (b) its evaluation by means of a hybrid methodology, MCDM-FIS (multi-criterion decision-making approach–fuzzy inference system), for which no prior reference could be found in the literature.
Finally, there are many voices and agendas [16,17,18,19] articulating the need to instrumentalise higher education to train future professionals to be aware of the sustainability values. Sustainable development should be an integral and systemic part of the university institutional framework. The demand is for not only the inclusion of environmental content in the curricula of the different subjects [20], but also in the criteria for evaluation of university quality, organisation and management [21], as well as in the teacher evaluation process. The aim is to impregnate the entire university universe with the philosophy of sustainable development and the promotion of environmental education. However, it is worth noting that, despite the increasing number of universities involved with sustainability, explicit references to this issue have not been appreciated in the drafting of ESGs, a topic of great importance currently demanded for both academic and social worlds. Therefore, this work proposes as a third contribution a conceptual model to evaluate the quality of HEIs that includes an explicit dimension for sustainability.
Section 2 of this document identifies the dimensions and criteria used to define the conceptual evaluation model. In Section 3, the methodology for the development of the model through MCDM tools and FIS is detailed. Section 4 includes the empirical evaluation of the proposed model, and presents and discusses the obtained results. Finally, Section 5 presents the conclusions of the study.

2. Dimensions and Criteria for Quality Assurance in HEIs

In order to evaluate the proposed conceptual model, it was first necessary to identify the different dimensions to be considered in the study. This task was of critical importance to the extent that the results may be highly dependent on and sensitive to the successful selection of the most appropriate dimensions. Given the imbalance identified in linking ESGs to the three basic pillars mentioned in the previous section, it was imperative to find alternative dimensions under which to encompass the ESG.
To this aim, several possible tools to use in the continuous quality improvement field were considered (i.e., the plan–do–check–act (PDCA) tool (also referred to as the Deming cycle), Six Sigma techniques, the SERVQUAL model and the EFQM model, among others [22]). From these, the PDCA was chosen. Several reasons supported this decision. First, PDCA (plan–do–check–act) is a well-known and universally accepted tool for systematically achieving continuous quality improvement in a variety of sectors, education being one of them [23,24]. It also provides a simple but effective framework, with dimensions (shown in Figure 1) that encompass the main ESG criteria. Note that dimension ‘C’ brings together both check actions and their corresponding improvement activities.
The next step involved the gathering of the ESG criteria under each of the three dimensions. As previously stated, the ESG are divided into three parts (internal quality assurance, external quality assurance and quality assurance agencies). In this paper, only the criteria included under the internal quality assurance label were considered, as self-evaluation is often carried out as a previous condition for accomplishing the external assessment [14]. In addition, the aim pursued by the authors was to demonstrate the suitability of the proposed model for quality evaluation in the higher education context; for this purpose, extending the number of criteria in order to introduce all the standards contemplated in the ESG2015 was not considered relevant.
With the focus on the first rubric, ESGs contain the following 10 criteria:
ESG01.
Policy for quality assurance
ESG02.
Design and approval of programmes
ESG03.
Student-centred learning, teaching and assessment
ESG04.
Student admission, progression, recognition and certification
ESG05.
Teaching staff
ESG06.
Learning resources and student support
ESG07.
Information management
ESG08.
Public information
ESG09.
On-going monitoring and periodic review of programs
ESG010.
Cyclical external quality assurance
As shown in Figure 2, ESG02, ESG05 and ESG06 criteria were grouped under the ‘Plan’ dimension, as they outline global objectives to be achieved in HEIs in terms of programmes, processes, methods and resources. Hereinafter, they are identified as A1, A2 and A3, respectively. Under the ‘Do’ dimension, the ESG03, ESG04 and ESG08 criteria were included, as they imply different aspects of the means of executing the previous global objectives (B1, B2, B3). Finally, the ‘Check/Act’ factor encompassed the ESG07, ESG09 and ESG10 criteria, which explicitly refer to the introduction of controls by which to articulate the monitoring and verification of quality assurance (C1, C2, C3). The ESG01 criterion refers to the establishment of a quality assurance policy, which captures the commitment of institutional representatives to the continuous improvement and integrity of the implemented quality systems. This criterion was not explicitly considered in the proposed model as it is a guiding principle that permeates the other PDCA dimensions.
Figure 2 also depicts ‘sustainability’ as the fourth dimension (Dimension D) considered in the proposed conceptual model.
The methodology suggested in this paper proposes the use of the first two MCDM techniques: the Fuzzy Decision-making Trial and Evaluation Laboratory method (FDEMATEL) to identify the priorities to be assigned to the dimensions of the model, and the Analytic Network Process (ANP) to weight the model criteria. Second, the paper proposes the design of a fuzzy inference system (FIS) that, while maintaining the structure of the conceptual model described, integrates the knowledge necessary for its evaluation.
In the next section, the MCDM and FIS methods are explained in deeper detail.

3. Methodology

In this paper, a systematic quality assessment system for European HEIs is proposed. The system is able to prioritise the determining factors involved in such an evaluation and to estimate the overall degree of quality of any institution based on the assessments of those factors. For this purpose, a hybrid methodology, MCDM-FIS, was chosen, for which no prior reference was found in the literature in this ambit.
The assessments were made via face-to-face interviews with eight university members (referred to, from here on, as experts). These interviews were scheduled to gather the information required for the study. The individual interviews were conducted, when possible, at interviewees’ offices, and opinions on the pairwise interrelations among dimensions and criteria were requested and collected using the ad-hoc matrices prepared for that purpose. The profile of the interviewees sought required that they be members belonging to university quality committees with no less than eight years of affiliation in the institution, in order to guarantee enough knowledge and experience on the issue.
Regarding MCDM, there are many multi-criterion techniques used to obtain a criterion’s weight in a decision system and to establish a ranking of its potential alternatives (Analytic Hierarchy Process –AHP-, Analytic Network Process –ANP-, Preference Ranking Organization Method for the Enrichment of Evaluations –PROMETHEE-, Decision Making Trial and Evaluation Laboratory DEMATEL, Technique for Order of Preference by Similarity to Ideal Solution –TOPSIS- and VIseKriterijumska Optimizacija I Kompromisno Resenje –VIKOR-, among others; [25,26,27,28,29,30,31]). Since none is clearly superior to the others [28], an appropriate combination of them should be chosen depending on the context in which they are to be applied. Fuzzy inference systems (FIS) have shown a high degree of effectiveness in the evaluation of systems when the definition of the factors on which the systems depend is subject to high degrees of uncertainty [32].
Based on the above-mentioned proposal of the conceptual evaluation model, the use of the FDEMATEL-ANP (FDANP) technique was proposed to identify the priority of the dimensions/criteria according to the influence interrelations provided by the interviewees.
Relying on the experts’ opinions, the design of an FIS was then established that, while maintaining the structure of the conceptual evaluation model previously described, integrated the weights of its factors in the determination of its knowledge base.
In accordance with the above-mentioned techniques, the collaboration elicited from these university experts allowed (a) assessment of the interrelationship influences between each of the paired dimensions proposed in the model (from which the FDEMATEL method was developed); (b) assessment of the interrelationship influences among each of the paired criteria proposed in the model (from which the FDANP method will be developed); (c) proposal of the partitions of the fuzzy variables of the model inference subsystems (once the partitions were agreed, the FIS could be further designed).
The above methodologies are explained step-by-step in Section 3.1 and 3.2, and are graphically summarised in Figure 3.

3.1. FDEMATEL-ANP (FDANP)

3.1.1. Fuzzy-DEMATEL

The decision-making trial and evaluation laboratory method (DEMATEL) is a multi-criterion decision technique developed in the mid-1970s by the Geneva Research Center of the Battelle Memorial Institute. DEMATEL makes it possible to convert reciprocal influence relationships between the factors (whether dimensions (D) or criteria (C)) on which a decision system depends into an intelligible structural model capable of identifying the typology of these factors (cause or effect) and assessing their impact on the system [31,33]. The representation of these impacts on maps (influence relation maps—IRM), allows the strategic factors of the decision system to be prioritised, and guides further improvement approaches. In recent years, the extension of this technique to the fuzzy field has allowed the vague assessments of experts to be processed using fuzzy numbers without modifying the general procedure of the DEMATEL technique [34]. The steps of this F-DEMATEL methodology are analytically detailed below.
Step 1. Establishment of the agreed fuzzy direct-relation matrix of pairwise influence comparisons (among ‘n’ factors) [ A ˜ ] (Equation (1)).
[ A ˜ ]   =   [ a ˜ 1 , 1 a ˜ 1 , n a ˜ n , 1 a ˜ n , n ]
Every element of this matrix a ˜ i , j   = ( l i , j   , m i , j   , u i , j   ) is a fuzzy number (in our case, a triangular fuzzy number—TFN) which aggregates the influence degrees—given by the group of experts—that a factor ‘i’ exerts on factor ‘j’ (i.e., through its fuzzy arithmetic mean [35]). Here, experts usually assessed the influence degrees from a fuzzy scale previously defined according to the following:
{ 0 ˜ n o   i n f l u e n c e ,   1 ˜ ,   2 ˜ ,   3   ˜ ,   4 ˜ m a x   i n f l u e n c e   }   =   { ( 0   0   1 ) ,   ( 0   1   2 ) ,   ( 1   2   3 ) , ( 2   3   4 ) , ( 3   4   4 ) }
Step 2. Calculation of the normalised fuzzy direct-relation matrix [ X ˜ ]
The matrix obtained from the previous step was normalised by parameter ‘s’ (Equations (2) and (3)):
s   =   M i n   ( 1 M a x 1 i n ( j   =   1 n u i , j ) ,   1 M a x 1 j n ( i   =   1 n u i , j ) )
[ X ˜ ]   =   s · [ A ˜ ]
Step 3. Calculation of the fuzzy total relation matrix [ T ˜ ]
This matrix can be calculated by Equation (4), [ I ˜ ] being the identity matrix.
[ T ˜ ]   =   lim k ( [ X ˜ ]   +   [ X ˜ ] 2   +     +   [ X ˜ ] k )   =   [ X ˜ ] · [ I ˜ X ˜ ] 1  
During this process, [ X ˜ ] must be decomposed into three submatrices associated with the vertexes of their fuzzy elements:   [ X ˜ l ] ,   [ X ˜ m ] ,   [ X ˜ u ] (Equation (5)).
[ T ˜ l ]   =   [ X ˜ l ] · [ I ˜ l X ˜ l ] 1       [ T ˜ m ]   =   [ X ˜ m ] · [ I ˜ m X ˜ m ] 1     [ T ˜ u ]   =   [ X ˜ u ] · [ I ˜ l X ˜ u ] 1
Subsequently, upon recomposing these three matrices, [ T ˜ ] was obtained.
Step 4. Construction of the influential relation map (IRM)
Upon defuzzifying the matrix [ T ˜ ] (e.g., through the graded mean integration representation method:   a i , j     =   ( l i , j     +   4 · m i , j     +   u i , j   ) / 6 [36], the total relation matrix [ T ] was obtained. The sum of the rows and columns of this matrix allowed two vectors to be built: a column vector [ R ] n x 1 —sum of rows—which shows the total influence effect emanated by a factor over the rest; and a row vector [ C ] 1 xn —sum of columns—which shows the total effect received by a factor from the rest. This row vector can be transposed to a column vector [ C ] n x 1 T (Equation (6)).
[ R ]   =   [   j   =   1 n t i , j   ] n x 1 ;   [ C ]   =   [   i   =   1 n t i , j   ] 1 x n   [ C ] T   =   [   i   =   1 n t i , j   ] T n x 1
By adding and subtracting the homologous components of both column vectors, two parameters—associated to each factor of the system—are obtained: prominence ( R   +   C ) and relation ( R C ) . These parameters can be plotted in a two-axis chart called the IRM of the system. The prominence of a factor ‘i’, (R + C)i —IRM abscissa axis—shows its strength of global influence (emanated and received); the higher the prominence, the higher the relevancy in the system. The relation of a factor ‘i’, (R − C)i—IRM ordinate axis—shows its net effect of influence on the system. Thus, if (R − C)I > 0, it will be considered a ‘cause factor’ (given its influence on the rest). If instead (R − C)I < 0, it will be considered an ‘effect factor’ (influenced by the rest). In this way, IRM helps to solve the decision-making problem by categorising all factors in the system (Figure 4).
If necessary, from the crisp total relation matrix [ T D ] , the importance weights of these factors can be calculated by normalising that matrix according to the sum of its rows, then transposing it and, finally, raising it to high powers till convergence is reached (the latter matrix will contain by columns of identical value the weights of the factors. These dimension weights are used below in both Section 3.1.2 (to obtain the criterion weights within the ANP supermatrix) and Section 3.2 (to build the FIS rule bases).

3.1.2. ANP and FDANP

The analytic network process (ANP) proposed by Saaty [37] extends the analytic hierarchy process (AHP) by allowing the consideration of a network of interdependences among the different elements and levels of a decision system (objectives, criteria, sub-criteria and alternatives). Thus, although ANP is also based on pairwise comparisons, it takes into account both the dependence between the elements of a cluster (inner dependence) and between clusters themselves (outer dependence), thereby allowing more complex and realistic systems to be modelled [38]. The ANP method starts from the definition of a network of influence interdependencies (agreed a priori by the decision-makers). Subsequently, an unweighted supermatrix is built based on the priority vectors obtained from the pairwise comparison matrices between the nodes of the network (also given by the decision-makers) using some prioritisation method (usually the eigenvalue method [39]). From this matrix, a weighted supermatrix is obtained through a normalisation process (unit sum columns). Finally, by means of successive powers of this matrix, the limit supermatrix (with identical columns) is obtained, which supplies the definitive priorities of each element of the network. However, ANP usually presents problems both in defining the structural interdependencies of the network and in obtaining consistency in the judgments given by experts in matrices of pairwise comparisons. This is due to the cognitive limitations of the decision-makers and the use of nine point Likert rating scales in high-order matrices [40]. In order to avoid these problems, the total relation matrix extracted from the DEMATEL technique can be considered an unweighted supermatrix, configuring a hybrid technique called DEMATEL-based ANP (DANP), which was used in this paper. Below, the steps of this methodology in its fuzzy version (FDANP) are analytically detailed [25,41,42].
Step 1. Consideration of the total relation matrix of criteria [ T C ]
Matrix [ T C ] can be obtained by defuzzifying the fuzzy total relation matrix of criteria obtained by FDEMATEL (Equation (7)), where Di denotes cluster ‘i’, Cij represents criterion ‘j’ within cluster ‘i’ and [Tcij] symbolises the relation submatrix between every two criteria from clusters ‘i’ and ‘j’.
Sustainability 12 01707 i001
Step 2. Calculation of the normalised total relation matrix [ T C α ]
[ T C α ] is obtained by normalising each of the submatrices [ T C i j ]   from   [ T C   ] depending on the sum of their rows. Equations (8)–(10) illustrate this process for the submatrix between Clusters 1 and 2, [ T C 12 ] .
Sustainability 12 01707 i002
Sustainability 12 01707 i003
Sustainability 12 01707 i004
Step 3. Obtaining the unweighted supermatrix [ W ]
[ W ] is obtained by transposing the matrix resulting from the previous step, [ T C α ] (Equation (11)).
Sustainability 12 01707 i005
Step 4. Obtaining the weighted supermatrix [ W W ]
Once the total relation matrix of dimensions is obtained [ T D   ] , it must be normalised (by sum of rows) to give [ T D α ] and transposed to give [ T D α ] . This last matrix will act as the weighting matrix over the unweighted matrix of the previous step to obtain the weighted supermatrix [ W W ] . This whole process is described analytically in Equations (12)–(15).
Sustainability 12 01707 i006
Sustainability 12 01707 i007
Sustainability 12 01707 i008
Sustainability 12 01707 i009
Finally, by raising [ W W ] to high powers, a limit supermatrix of identical columns is obtained. The values in those columns correspond to the definitive weights of all criteria. These criterion weights are used (in conjunction with dimension criteria obtained in Section 3.1.1) to build the FIS rule bases, as explained in Section 3.2.

3.2. Fuzzy Inference Systems

Fuzzy inference systems (FIS) allow the incorporation of expert decision-support knowledge in the form of fuzzy rules that relate system variables subject to subjective or uncertain assessments (which are difficult to accurately describe using traditional numerical variables). To design an FIS, its knowledge base must initially be defined as follows: (a) linguistic labels that determine the partitions of its fuzzy input and output variables; and (b) ‘if–then’ type rules reflecting the label to be given to each output fuzzy variable according to the label set of the input fuzzy variables (e.g., IF [Var1 is LOW & Var2 is MED] THEN [Var3 is MED]). In the processing of these fuzzy variables, the use of trapezoidal fuzzy numbers has been demonstrated to have sufficient efficiency and utility [43].
Subsequently, the process of inference (Mamdani type) takes place in five phases: (1) crisp input fuzzification of each of the cases to infer; (2) application of fuzzy operators in the antecedent of each rule; (3) implication to the consequent of each rule; (4) aggregation of all the implicated consequents of the entire rule base; and (5) defuzzification of the final aggregate. A detailed description of each of these phases can be found in [44]).
In general terms, the definition of the knowledge base is built upon the judgement of a team of experts who agree on: (a) the typology of the labels to be used in all variables as well as their corresponding partitions; and (b) the labels to be assigned to the output of each rule, which will subsequently trigger the inference process.
In this paper, to partition the variables, the expert team relied on a linguistic representation method based on 2-tuples that allowed its systematic design by means of fuzzy numbers. To this aim, and based on a linguistic preferences scale previously defined, the parameterisation of all partitions for all system variables was agreed [45,46].
Regarding the definition of the system rule based on the model, it was decided to identify the label to be assigned to each rule by the minimum distance from the weighted fuzzy sum of its input labels (using as local weights those obtained using the FDANP methodology) to all labels in the agreed partition of the output variable. The use of a knowledge elicitation method based on the extension principle [47] according to the expression given in Equation (16) was proposed to be used in the calculation of these distances, assuming trapezoidal fuzzy numbers:
D j = D ( [ a o , b o , c o , d o ] [ a j , b j , c j , d j ] ) =                       P a ( a o a j ) 2   +   P b ( b o b j ) 2   +   P c ( c o c j ) 2   +   P d ( d o d j ) 2
where [ a o , b o , c o , d o ] represents the weighted fuzzy output of the rule for which the output is to be estimated, [ a j , b j , c j , d j ] is the j-th fuzzy label of the output variable partition, and Pa, Pb, Pc and Pd are the priorities given to the vertices of fuzzy trapezoids considered (i.e., Pa + Pb + Pc + Pd = 1).
Finally, the label of the partition to which the previous distance was minimal was chosen as the output label. This methodological proposal is feasible if the input and output variables have identical ranges and label progression (lower values–worse/higher values–better). In this way, the procedure exempts the expert team from agreeing to the outputs of the rules, which results both in the prevention of judgement inconsistency with respect to influence valuations issued on FDANP and in a substantial resource saving in the rule extraction process.
In this paper, the design of the model knowledge base was implemented using the Matlab Fuzzy Logic Toolbox ©. The inference maps available in this tool allow the analysis, in a simple and intuitive manner, of the evaluations obtained in any HEI. Thus, each map evaluated an output variable for any values of two of its input variables—those shown on the map—and fixed values of the rest.

4. Empirical Evaluation of the Proposed Quality Assessment Model: Results and Discussion

In this section, the above-explained methodologies were applied to the proposed quality assessment model.

4.1. F-DEMATEL

Initially, the eight expert team members provided pairwise evaluations on the interrelationship influences among the four dimensions of the proposed model (A—‘plan’, B—‘do’, C—‘check/act’ and D—‘sustainability’). They used a five level linguistic preference scale (triangular fuzzy numbers—TFN).
{ 0 ˜ n o   i n f l u e n c e ,   1 ˜ ,   2 ˜ ,   3   ˜ ,   4 ˜ m a x   i n f l u e n c e   } = { ( 0   0   1 ) ,   ( 0   1   2 ) ,   ( 1   2   3 ) , ( 2   3   4 ) , ( 3   4   4 ) }
The agreed fuzzy direct-relation matrix of pairwise influence comparisons [ A ] ˜ (Equation (17)) was obtained by using the fuzzy arithmetic mean [35] of those evaluations; from it, the steps outlined in Section 3.1.1 were followed.
Sustainability 12 01707 i010
After normalising the [ A ] ˜ matrix with s = 9,88 (Equations (2)–(5)), the fuzzy total relation matrix of the four dimensions [ T D ] ˜ (Equation (18)) was estimated.
Sustainability 12 01707 i011
This matrix was then defuzzified by using the graded mean integration representation method to compute the crisp total relation matrix [ T D ] (Equation (19)). From this, once the dimensions’ prominences (R + C) and relations (R − C) (according to Equation (6)) were calculated, the influential relation map (Figure 5) of the model dimensions was plotted.
Sustainability 12 01707 i012
The chart shows that the dimensions A—‘plan’ and B—‘do’ had the most important prominences and were cause-type dimensions, whereas dimensions C—‘check/act’ and D—‘sustainability’ were effect-type dimensions. This result strongly suggests, in line with the real world, that the sustainability dimension is influenced by the planning and implementation dimensions of the model, and that measures and procedures taken in the two early phases of the quality assessment system will translate into clear impacts upon it.
From the matrix [ T D ] , the importance weights of each dimension [WD4] (see the shadow column in Equation 20) were calculated by first normalising the matrix according to the sum of its rows to give [ T D 4 α ] , then transposing it to give [ T D 4 α ] and , finally, raising it to high powers until convergence was reached at ( [ T D 4 α ] ) n .
Sustainability 12 01707 i013
The results showed that sustainability (Dimension D) obtained a substantial weighting in the system (26%). This result endorsed some statements from the literature about the potential responsibility of HEIs in bringing about necessary changes not only through research, but also through better training of their graduates, those who are becoming professionals capable of responding more accurately to the challenges posed by the sustainability paradigm [48,49]. This significant figure of 26% suggests, among other things, the need to make efforts to reach a consensus, for instance, on the standardisation of certain sustainability criteria in order to incorporate them into assessment models.
Based on the same evaluations provided by the expert team, but exclusively considering the traditional three dimensions (A, B and C), the weights [WD3] would have been those given in Equation (21). In this alternative scenario, the preponderance of dimension C—‘check/act’ stood out over A—‘plan’ and B—‘do’. These were the data used in the FDANP methodology to determine the overall weighting of the criteria on which these three dimensions depend.
Sustainability 12 01707 i014

4.2. FDANP

After averaging the pairwise fuzzy allocations given by experts to the influence interrelationships among the model criteria, a 9 × 9 fuzzy matrix was obtained. After defuzzifying it and scaling it in the range of scores 0–4 [33], the total relation matrix of criteria [ T C ] (Equation (22)) was obtained. From this matrix, and once the corresponding prominences (R + C) and relations (R − C) were calculated, the influential relation maps (Figure 6) of the model criteria were determined.
Sustainability 12 01707 i015
Figure 6 shows that all criteria had considerable prominence, with the notable exception of B2 (related to academic management), in comparison to the other two in the same Dimension B, and to the rest in the remaining dimensions.
It was also found that the three ‘check/act’ criteria were more relevant globally, especially those relating to information management (C1). On the other hand, and consistent with the IRM dimensions, all planning criteria (A1: Design, A2: Personal, and A3: Resources) were depicted as causing influencers, along with one ‘do’ criterion—public information (B3) —and one ‘check/act’ criterion—information management (C1). In contrast, the criteria B1—‘teaching–learning’ and B2—academic management (related to the ‘do’ dimension) and, in smaller amounts, C2—‘monitoring and evaluation of programmes’ (related to the ‘check/act’ dimension) were clearly perceived as influenced criteria.
From matrix [ T C ] , and following the steps in Equations (8)–(11), the unweighted supermatrix [W] (Equation (23)) was estimated.
Sustainability 12 01707 i016
Subsequently, following the steps of Equations (12)–(15), the weighted supermatrix [ W W ] (Equation (24)) was determined by weighting the previous matrix [ W ] with the transposed total relation matrix [ T D 3 α ] of the three dimensions obtained in Equation (21).
Sustainability 12 01707 i017
Finally, by raising [WW] to high powers, the so-called ‘global weights’ of all criteria [WC] were obtained (Equation (25)); from these, the ‘local weights’ [WC*] were calculated by dividing the global weights into those of their corresponding three dimensions [WD3].
Sustainability 12 01707 i018
Both the local weights obtained for the criteria and those of the four dimensions (shaded in Equation (25)) were used to calculate the outputs of the rules in the FIS, as explained in the next section.

4.3. Fuzzy Inference System

The FIS proposed for evaluation was created in the likeness of the conceptual model and consisted of the following four subsystems. Three criteria subsystems (to evaluate the dimensions A—‘plan’, B—‘do’ and C—‘check/act’ from the valuations [0–10] of their corresponding criteria and a one-dimension subsystem (which collected the previous output evaluations, A, B and C, to be processed now as inputs, together with the valuation [0–10] given to the fourth dimension, D—‘sustainability’)
To design the inference subsystems of the model, the expert team agreed to use three labels for all input variables in the model (low—L, medium—M, high—H), five for the output variables of the criteria subsystems (very low—VL, low—L, medium—M, high—H, very high—VH) and seven for the output of the final dimension subsystem (extremely low—EL, very low—VL, low—L, medium—M, high—H, very high—VH, extremely high—EH), thus allowing a greater degree of discrimination in building the knowledge base. The process followed by the experts to determine a partition started with estimating their preferences regarding the potential location of each label core. Later, once the cores were agreed upon using a 2-tuple method [46], the complete labels were obtained by symmetry [45] in order to achieve strong partitions [50], which have better comprehensibility properties and satisfy important semantic constraints [51].
Figure 7 shows the agreed partitions of all model variables, the parameters of which are shown in Figure 8.
Subsequently, by applying the methodology described in Section 3.2, and using the local weights of criteria and dimensions obtained in Section 4.2, the rule bases of the different subsystems were reached. Figure 9 shows an excerpt of the calculation of the first nine output rules of Inference Subsystem A (see last column). First, the output-weighted trapeze of each rule was calculated as the weighted sum of its input labels in A1, A2 and A3 (according to the weights WA1 = 0.34, WA2 = 0.30, WA3 = 0.35). Subsequently, the distances of each weighted trapeze to all potential output labels of a rule (VL, L, M, H, VH) were calculated. The final output label of each rule was the one corresponding to the minimum distance—shaded in this figure. Figure 10 shows the obtained rule bases of the model in the criteria subsystems.
Once the knowledge base for all subsystems was designed, it was implemented in the MATLAB Toolbox Fuzzy Logic ©. Thus, it is possible to make inferences of the overall quality assessment of any institution in the EHEA according to the crisp values [0–10] granted to its nine criteria and to the D—‘sustainability’ dimension. Note that these crisp values should be quantified for each factor from their corresponding key performance indicators (KPIs), the ad hoc design of which is beyond the scope of this paper. In any case, the surface maps resulting from the different inference subsystems allow analysis of the evolution of both the assessments obtained in the different dimensions of the model and the overall quality assessment of any educational institution of the EHEA.
When interpreting the inference maps of any subsystem, it should be noted that each graph illustrates the output evolution of that subsystem for any value of two of its input factors (and for constant values of the factors not depicted in the chart). As an example, the maps obtained from the designed model are shown in Figure 11 and Figure 12.
In all surface maps, a consistent evolution of the quality assessment of any institution can be observed. For increasing values of the variables illustrated in the base of the graph, progressively higher final evaluations can be obtained according to the surface shown, although with different nuances according to the analysed graph. On the one hand, the knowledge inserted in the rules of each subsystem (Figure 10) can prioritise some variables in the final evaluation—for example, in Figure 11 (right), quality is maximised for necessarily high values in A1, although not so high in A2. On the other hand, the different fixed values arranged for the variables not illustrated in the graph can substantially modify the shape of the surface maps (in general terms, the higher the fixed value, the greater the height reached by the surface map, as can be seen in the successive maps of Figure 12). The results of all these maps are analysed in more detail below.
Figure 11 shows the valuation maps of the A—‘plan’ dimension according to criteria A1 and A2. The map on the left corresponds to a ‘low’ value of A3 and the one on the right corresponds to a ‘High’ value.
In the left chart, the A—‘plan’ dimension started from a very low score of 1 for low A1 and low–medium A2 values. However, low A1 and high A2 values scored up to a 4 point valuation for A (as for medium–high A2 together with low A2 values). A total 5.5 points in A were reached for medium–high A1 and medium A2 values (as well as for medium A1 and medium–high A2 values). Finally, by using high valuations in both A1 and A2, Dimension A gained the maximum score of 7 points.
In the right chart, Dimension A started from a low score of 4 for low values of A1 and A2. For low A1 and medium A2 values—and vice versa—Dimension A went up to 5.5 points. For mean values of A1 and A2, 7 points were reached (as well as for low–moderate A1 and high A2 values or high A1 and low A2 values). Finally, the maximum ratings of 9 was reached when A1 was high and A2 was medium–high.
The top-left map (Figure 12) was obtained for low values of A and B. When D and C were low, the global quality assessments were very low (1). When D was kept at low values but C was increased (and vice versa), the global quality value grew progressively in two steps until it reached a maximum of 4 points (which was also reached for intermediate values of D and C). Only with high values of D and medium–high values of C (and vice versa) was it possible to achieve medium global quality assessments of around 5.5 points.
The top-right map (Figure 12) was obtained for low values of A and high values of B. It started from low global quality assessments (4) for low values of the D and C dimensions. When D was kept at low values and C was increased (and vice versa) the global quality assessment grew progressively in only one step until it reached a maximum of 5.5 points. Only for high values of D and medium–high values of C (and vice versa) was it possible to achieve notable global quality values of around 6,5 points.
The bottom-left map (Figure 12) was obtained for high values of A and low values of B. The main difference with the previous map was the initial assessment of around 3 points.
The bottom-right map (Figure 12) was obtained for high values of A and B. It started from medium global quality assessments of 5 points when D and C were low. When D was kept at low values but C was increased (and vice versa) the global quality assessment grew progressively in only one step until it reached a maximum of 6,5 points (which was also reached for intermediate values of D and C). Only for high values of D and medium–high values of C (and vice versa) were excellent global quality assessments progressively achieved in two steps at around 8 and 9 points.
Nonetheless, in this latter subsystem, a rather symmetrical behaviour was noted in the four maps. This was due to the low dispersion of the model dimension weights (WA-0.22; WB-0.26; WC-0.27; WD-0.26) obtained according to the FDEMATEL methodology. These results highlight the similar importance of Dimension D dimension when compared to the other dimensions generally considered in such evaluations, and highlight the need to incorporate it into the overall quality assessment of EHEA.
The proposed method can be considered reliable since the results can be reproduced in an easy and traceable way by following the proposed methodology. To that end, a detailed explanation of the steps performed provided. With respect to the validity, the literature was reviewed in order to build a model that considers the existing theory and knowledge of the concept of quality to be measured and to detect other aspects, such us sustainability, that could be applied to enrich the model. However, given the novelty of both the model and the proposed methodology, comparison tests could not be carried out. In addition, to implement them, the development of KPIs would be required to initially quantify the crisp inputs of the model which, as already noted, exceeded the scope of this paper.

5. Conclusions

The assurance of quality and the achievement of excellence in higher education institutions has been and still is a global demand in current society. Accreditation as a mechanism to ensure better quality levels has emerged, however, as a continuous challenge, since its effectiveness has been questioned for different reasons that have been explained in the paper. This paper aimed to contribute to the quality assurance issue by first proposing a new conceptual framework based on the PDCA tool, wherein sustainability is considered a momentous dimension to be considered in the evaluation. Next, two MCDM techniques were used to establish priorities to be assigned to the model dimensions from the opinions provided by a group of experts, and to determine weights for the criteria encompassed by those dimensions. Finally, a fuzzy inference system is built to make evaluations of educational institution quality based on the values of the model dimensions.
The proposed FIS allowed the building of the rule bases of the evaluation model by weighting the labels of their antecedents (from the weights obtained with the MCDM methodologies previously applied). This exempted the team of experts from reaching consensus on the output labels of the rules, avoiding possible inconsistencies with their influence assessments and saving resources in their determination.
The results showed a prominent role of by the sustainability dimension when explicitly included in the analysis, suggesting the need for a comprehensive revision of actual standards and guidelines so that sustainability principles are implemented and monitored in higher education quality evaluation procedures.
However, the results obtained in the design of the proposed model were limited by the knowledge provided by the expert team participating in the study. Therefore, a possible extension of this study would be to expand the number of interviewees on the panel of experts.
Other suggestions for future research include (a) the analysis and design of specific indicators for input crisp valuations (0–10) of both criteria inference subsystems and the sustainability dimension of the model; and (b) to identify, analyse and classify the key sustainability aspects in the evaluation of the institutions under study (based on the relevant literature in this area and the assessments that could be gathered from a panel of experts similar to the one proposed). In this way, these concepts could be incorporated into the determination of the indicators necessary to quantify the crisp valuations of the nine classic criteria set out in our model. In such a case, a model similar to the one proposed in this paper could be developed based exclusively on the nine criteria (already incorporating sustainability concepts) and the three initial dimensions: A—‘plan’, B—‘do’ and C—‘check/act’.

Author Contributions

All authors have collaborated in the design of the model as well as in the collection of information, the analysis of the results and the establishment of conclusions. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The author declare no conflict of interest.

References

  1. Bornmann, L.; Mittag, S.; Danie, H.-D. Quality Assurance in Higher Education – Meta-Evaluation of Multi-Stage Evaluation Procedures in Germany. High. Educ. 2006, 52, 687–709. [Google Scholar] [CrossRef]
  2. Neave, G. On the Cultivation of Quality, Efficiency and Enterprise: An Overview of Recent Trends in Higher Education in Western Europe, 1986-1988. Eur. J. Educ. 1988, 23, 7–23. [Google Scholar] [CrossRef]
  3. Astin, A.W.; Antonio, A.L. Assessment for Excellence: The Philosophy and Practice of Assessment and Evaluation in Higher Education, 2nd ed.; Rowman & Littlefield Publishers: Lanham, MD, USA, 2012. [Google Scholar]
  4. Brdulak, J. Two Approaches to Quality Assurance. In Drivers and Barriers to Achieving Quality in Higher Education; Eggins, H., Ed.; SensePublishers: Rotterdam, The Netherlands, 2014; pp. 55–66. [Google Scholar] [CrossRef]
  5. Brink, C. Quality and Standards: Clarity, Comparability and Responsibility. Qual. High. Educ. 2010, 16, 139–152. [Google Scholar] [CrossRef]
  6. ENQA. Available online: Https://Enqa.Eu/Index.Php/Home/Esg/ (accessed on 21 January 2020).
  7. Ackoff, R.L. Creating the Corporate Future: Plan or Be Planned For; Wiley: New York, NY, USA, 1981. [Google Scholar]
  8. Allen, M. The Goals of Universities; Open University Press: Milton Keynes, UK; Philadelphia, PA, USA, 1988. [Google Scholar]
  9. Benneworth, P.; Jongbloed, B.W. Who Matters to Universities? A Stakeholder Perspective on Humanities, Arts and Social Sciences Valorisation. High. Educ. 2010, 59, 567–588. [Google Scholar] [CrossRef] [Green Version]
  10. Asif, M.; Raouf, A. Setting the Course for Quality Assurance in Higher Education. Qual. Quant. 2013, 47, 2009–2024. [Google Scholar] [CrossRef]
  11. Smidt, H. European Quality Assurance—A European Higher Education Area Success Story [Overview Paper]. In The European Higher Education Area: Between Critical Reflections and Future Policies; Curaj, A., Matei, L., Pricopie, R., Salmi, J., Scott, P., Eds.; Springer International Publishing: Cham, Switzerland, 2015. [Google Scholar] [CrossRef] [Green Version]
  12. Green, D. What Is Quality in Higher Education? Taylor & Francis: Bristol, PA, USA, 1994. [Google Scholar]
  13. Ryan, T. Quality Assurance in Higher Education: A Review of Literature. High. Learn. Res. Commun. 2015, 5. [Google Scholar] [CrossRef]
  14. Sarrico, C.S.; Rosa, M.J.; Teixeira, P.N.; Cardoso, M.F. Assessing Quality and Evaluating Performance in Higher Education: Worlds Apart or Complementary Views? Minerva 2010, 48, 35–54. [Google Scholar] [CrossRef]
  15. Vlăsceanu, L.; Grunberg, L.; Pârlea, D. Quality Assurance and Accreditation: A Glossary of Basic Terms and Definitions. Available online: http://lst-iiep.iiep-unesco.org/cgi-bin/wwwi32.exe/[in=epidoc1.in]/?t2000=024133/(100) (accessed on 9 December 2019).
  16. Cullingford, C.; Blewitt, J. The Sustainability Curriculum: The Challenge for Higher Education; Routledge: London, UK, 2013. [Google Scholar]
  17. De Geli, C.A.M.; Leal, F.W. Education for Sustainability in University Studies: Experiences from a Project Involving European and Latin American Universities. Int. J. Sustain. High. Educ. 2006, 7, 81–93. [Google Scholar] [CrossRef]
  18. Junyent, M.; de Ciurana, A.M.G. Education for Sustainability in University Studies: A Model for Reorienting the Curriculum. Br. Educ. Res. J. 2008, 34, 763–782. [Google Scholar] [CrossRef]
  19. Ull, M.Á. Sostenibilidad y educación superior: La formación para la sostenibilidad en los nuevos títulos de grado. Carpeta Inf. del CENEAM 2011, 6, 7–12. [Google Scholar]
  20. Barrón, A.; Navarrete, A.; Ferrer-Balas, D. Sostenibilización curricular en las universidades españolas. ¿ha llegado la hora de actuar? Revista Eureka sobre Enseñanza y Divulgación de las Ciencias 2010, 388–399. [Google Scholar] [CrossRef]
  21. Alba, D.; Barbeitos, R.; Barral, M.T. Estrategias de sostenibilidad y responsabilidad social en las universidades españolas: Una herramienta para su evaluación. Revista de Currículum y Formacióón de Profesorado 2012, 16, 59–75. [Google Scholar]
  22. Sokovic, M.; Pavletic, D.; Pipan, K.K. Quality Improvement Methodologies – PDCA Cycle, RADAR Matrix, DMAIC and DFSS. J. Achiev. Mater. Manuf. Eng. 2010, 43, 8. [Google Scholar]
  23. Chen, S.-H. The Establishment of a Quality Management System for the Higher Education Industry. Qual. Quant. 2012, 46, 1279–1296. [Google Scholar] [CrossRef]
  24. Venkatraman, S. A Framework for Implementing TQM in Higher Education Programs. Qual. Assur. Educ. 2007, 15, 92–112. [Google Scholar] [CrossRef]
  25. Gölcük, İ.; Baykasoğlu, A. An Analysis of DEMATEL Approaches for Criteria Interaction Handling within ANP. Expert Syst. Appl. 2016, 46, 346–366. [Google Scholar] [CrossRef]
  26. Kahraman, C.; Onar, S.C.; Oztaysi, B. Fuzzy Multicriteria Decision-Making: A Literature Review. Int. J. Comput. Intell. Syst. 2015, 8, 637–666. [Google Scholar] [CrossRef] [Green Version]
  27. Mardani, A.; Jusoh, A.; Zavadskas, E.K. Fuzzy Multiple Criteria Decision-Making Techniques and Applications – Two Decades Review from 1994 to 2014. Expert Syst. Appl. 2015, 42, 4126–4148. [Google Scholar] [CrossRef]
  28. Özkan, B.; Özceylan, E.; Korkmaz, I. brahim H.; Çetinkaya, C. A GIS-Based DANP-VIKOR Approach to Evaluate R&D Performance of Turkish Cities. Kybernetes 2019, 48, 2266–2306. [Google Scholar] [CrossRef]
  29. Saaty, T.L. The Modern Science of Multicriteria Decision Making and Its Practical Applications: The AHP/ANP Approach. Oper. Res. 2013, 61, 1101–1118. [Google Scholar] [CrossRef]
  30. Saaty, T.L. How to Make a Decision: The Analytic Hierarchy Process. Eur. J. Oper. Res. 1990, 48, 9–26. [Google Scholar] [CrossRef]
  31. Si, S.-L.; You, X.-Y.; Liu, H.-C.; Zhang, P. DEMATEL Technique: A Systematic Review of the State-of-the-Art Literature on Methodologies and Applications. Math. Probl. Eng. 2018. [Google Scholar] [CrossRef] [Green Version]
  32. Papageorgiou, E.I.; Salmeron, J.L. A Review of Fuzzy Cognitive Maps Research During the Last Decade. IEEE Trans. Fuzzy Syst. 2013, 21, 66–79. [Google Scholar] [CrossRef]
  33. Tzeng, G.; Chiang, C.; Li, C. Evaluating Intertwined Effects in E-Learning Programs: A Novel Hybrid MCDM Model Based on Factor Analysis and DEMATEL. Expert Syst. Appl. 2007, 32, 1028–1044. [Google Scholar] [CrossRef]
  34. Dytczak, M.; Ginda, G. Is Explicit Processing of Fuzzy Direct Influence Evaluations in DEMATEL Indispensable? Expert Syst. Appl. 2013, 40, 5027–5032. [Google Scholar] [CrossRef]
  35. Gao, L.S. The Fuzzy Arithmetic Mean. Fuzzy Sets Syst. 1999, 107, 335–348. [Google Scholar] [CrossRef]
  36. Chen, S.H.; Wang, S.T.; Chang, S.M. Some Properties of Graded Mean Integration Representation of L-R Type Fuzzy Numbers. Available online: https://link.galegroup.com/apps/doc/A155737851/AONE?sid=lms (accessed on 17 December 2019).
  37. Saaty, T.L. Decision Making with Dependence and Feedback: The Analytic Network Process: The Organization and Prioritization of Complexity, 1st ed.; RWS Publications: Pittsburgh, PA, USA, 1996. [Google Scholar]
  38. Hsu, C.-H.; Wang, F.-K.; Tzeng, G.-H. The Best Vendor Selection for Conducting the Recycled Material Based on a Hybrid MCDM Model Combining DANP with VIKOR. Resour. Conserv. Recycl. 2012, 66, 95–111. [Google Scholar] [CrossRef]
  39. Saaty, T.L. A Scaling Method for Priorities in Hierarchical Structures. J. Math. Psychol. 1977, 15, 234–281. [Google Scholar] [CrossRef]
  40. Xu, Z.; Wei, C. A Consistency Improving Method in the Analytic Hierarchy Process. Eur. J. Oper. Res. 1999, 116, 443–449. [Google Scholar] [CrossRef]
  41. Liu, C.-H.; Tzeng, G.-H.; Lee, M.-H. Improving Tourism Policy Implementation—The Use of Hybrid MCDM Models. Tour. Manag. 2012, 33, 413–426. [Google Scholar] [CrossRef]
  42. Chen, F.-H.; Hsu, T.-S.; Tzeng, G.-H. A Balanced Scorecard Approach to Establish a Performance Evaluation and Relationship Model for Hot Spring Hotels Based on a Hybrid MCDM Model Combining DEMATEL and ANP. Int. J. Hosp. Manag. 2011, 30, 908–932. [Google Scholar] [CrossRef]
  43. Lyu, H.-M.; Shen, S.-L.; Zhou, A.; Yang, J. Data in Risk Assessment of Mega-City Infrastructures Related to Land Subsidence Using Improved Trapezoidal FAHP. Data Brief 2020, 28, 105007. [Google Scholar] [CrossRef]
  44. Fuzzy Reasoning and Its Applications; Mamdani, E.H.; Gaines, B.R. (Eds.) Academic Press: London, UK, 1981. [Google Scholar]
  45. Garcia, N.; Puente, J.; Fernandez, I.; Priore, P. Suitability of a Consensual Fuzzy Inference System to Evaluate Suppliers of Strategic Products. Symmetry 2018, 10, 22. [Google Scholar] [CrossRef] [Green Version]
  46. Puente, J.; Gascon, F.; Ponte, B.; de la Fuente, D. On Strategic Choices Faced by Large Pharmaceutical Laboratories and Their Effect on Innovation Risk under Fuzzy Conditions. Artif. Intell. Med. 2019, 100, 101703. [Google Scholar] [CrossRef]
  47. DuBois, D.; Prade, H. Fuzzy Sets and Systems: Theory and Applications; Academic Press: New York, NY, USA, 1980. [Google Scholar]
  48. Davis, S.A.; Edmister, J.H.; Sullivan, K.; West, C.K. Educating Sustainable Societies for the Twenty-first Century. Int. J. Sustain. High. Educ. 2003, 4, 169–179. [Google Scholar] [CrossRef]
  49. Sherren, K. Core Issues: Reflections on Sustainability in Australian University Coursework Programs. Int. J. Sustain. High. Educ. 2006, 7, 400–413. [Google Scholar] [CrossRef]
  50. Casillas, J.; Cordón, O.; Triguero, F.H.; Magdalena, L. Interpretability Issues in Fuzzy Modeling; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  51. Mencar, C.; Fanelli, A.M. Interpretability Constraints for Fuzzy Information Granulation. Inf. Sci. 2008, 178, 4585–4618. [Google Scholar] [CrossRef]
Figure 1. Basic dimensions of the proposed system for higher education quality assessment.
Figure 1. Basic dimensions of the proposed system for higher education quality assessment.
Sustainability 12 01707 g001
Figure 2. Architecture of dimensions and criteria considered in the conceptual model.
Figure 2. Architecture of dimensions and criteria considered in the conceptual model.
Sustainability 12 01707 g002
Figure 3. Methodology flowchart.
Figure 3. Methodology flowchart.
Sustainability 12 01707 g003
Figure 4. Influential relation map (IRM) example (prominence–relation).
Figure 4. Influential relation map (IRM) example (prominence–relation).
Sustainability 12 01707 g004
Figure 5. Influential relation map (four dimensions).
Figure 5. Influential relation map (four dimensions).
Sustainability 12 01707 g005
Figure 6. Influential relation maps (nine criteria).
Figure 6. Influential relation maps (nine criteria).
Sustainability 12 01707 g006
Figure 7. Agreed partitions of all model variables.
Figure 7. Agreed partitions of all model variables.
Sustainability 12 01707 g007
Figure 8. Parameters of the labels in the model partitions.
Figure 8. Parameters of the labels in the model partitions.
Sustainability 12 01707 g008
Figure 9. Calculation of the output labels for Subsystem A rules.
Figure 9. Calculation of the output labels for Subsystem A rules.
Sustainability 12 01707 g009
Figure 10. Rule bases in the criteria subsystems.
Figure 10. Rule bases in the criteria subsystems.
Sustainability 12 01707 g010
Figure 11. Dimension A inference map for low (left) and high (right) values of A3.
Figure 11. Dimension A inference map for low (left) and high (right) values of A3.
Sustainability 12 01707 g011
Figure 12. Assessment of the global quality (labelled as ‘Final’ on the Y-axis) according to Dimensions C—‘check/act’ and D—‘sustainability’ for different combinations of constant values of Dimensions A—‘plan’ and B—‘do’ (Top left: A low and B low; Top right: A low and B High; Bottom left: A High and B low; Bottom right: A High and B High).
Figure 12. Assessment of the global quality (labelled as ‘Final’ on the Y-axis) according to Dimensions C—‘check/act’ and D—‘sustainability’ for different combinations of constant values of Dimensions A—‘plan’ and B—‘do’ (Top left: A low and B low; Top right: A low and B High; Bottom left: A High and B low; Bottom right: A High and B High).
Sustainability 12 01707 g012

Share and Cite

MDPI and ACS Style

Puente, J.; Fernandez, I.; Gomez, A.; Priore, P. Integrating Sustainability in the Quality Assessment of EHEA Institutions: A Hybrid FDEMATEL-ANP-FIS Model. Sustainability 2020, 12, 1707. https://doi.org/10.3390/su12051707

AMA Style

Puente J, Fernandez I, Gomez A, Priore P. Integrating Sustainability in the Quality Assessment of EHEA Institutions: A Hybrid FDEMATEL-ANP-FIS Model. Sustainability. 2020; 12(5):1707. https://doi.org/10.3390/su12051707

Chicago/Turabian Style

Puente, Javier, Isabel Fernandez, Alberto Gomez, and Paolo Priore. 2020. "Integrating Sustainability in the Quality Assessment of EHEA Institutions: A Hybrid FDEMATEL-ANP-FIS Model" Sustainability 12, no. 5: 1707. https://doi.org/10.3390/su12051707

APA Style

Puente, J., Fernandez, I., Gomez, A., & Priore, P. (2020). Integrating Sustainability in the Quality Assessment of EHEA Institutions: A Hybrid FDEMATEL-ANP-FIS Model. Sustainability, 12(5), 1707. https://doi.org/10.3390/su12051707

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop