4.1. Dysfunctions of the MMM and Its Effects against the Public University System’s Sustainability
By extending the comments of [
27], and by taking into account the particular situations of the SUS, below we list the main set of dysfunctions of the metric valuation method, which focus mostly on the university environment. Starting by the distortion of information:
(i) Measuring the most easily measurable, easy versus complex.
It is clear, for metric management followers, that they will not measure the intangibles that are very difficult to detect and quantify. Perhaps they will not even think about their existence at all. They are used to being fooled by rankings and indicators, and do not think about anything different to what they measure.
There is a natural tendency to simplify problems by measuring the most easily measurable elements. Yet what is easily measured is rarely what is more important, and may well not be important at all. This is the first metric dysfunction type.
In Spain, a researcher’s performance, in agreement with the MMM, is measured in terms of the number of papers published in ranked journals, and the number of times an author is cited [
46,
47].
This is not a correct assessment indicator because, without carefully studying the published papers, it is not possible to identify the scientific contribution or the added value of a paper and its applicability and/or effects compared to previous studies. It has also been proved that the citations a paper receives does not correspond to the number of times that a paper has actually been read [
31,
36,
38,
48,
49].
The economic incentive based on pay-for-performance motivates an almost general tendency to maximize productivity by repeating an idea, in a more or less hidden way, to reach another publication that maximizes the economic return. It is a kind of publication greed [
48,
49,
50]. The present Spanish university environment forces overproduction, but hides the poor quality of contributions; see page 122 of [
32].
If anyone wishes to know someone’s research activity, it is necessary to carefully read all his/her produced works and to identify the originality, level of redundancy (which is usually high), and also if the candidate is able to manage and solve different problems. The mechanical procedures involved in the MMM do not capture these fine intangibles.
This “supermarket” research model is what the SUS promulgates and what it should resist. The system enhances the quantity by objectifying research, knowledge and the studied subject; see page 58 of [
32]. This paradigm means that researcher activity lacks interest and intellectual transcendence. As a result, researchers speed up and publish as many papers as they can [
51].
Thus, the system only values research output, understood as published papers, and disregards the scientific and social interest. The consequence of this system is that many researchers exist who essentially perform the same activity throughout their scientific careers, with the inevitable component of repetition and a lack of innovation [
52]. One major problem is not lack of production, but an excessively repeated and irrelevant production.
Candidates should be asked to explain the relevance of their production. No doubt many researchers who are the authors of lots of papers would probably not know what to say. The ranking of a journal does not guarantee the quality of a paper published in it. Moreover, the gap between the prestige and relevance of, and the interest in, lecturers’ teaching and researching activities widens with time, and teaching is increasingly disregarded [
51].
Thus, attention should focus on teaching. Basically, the teaching process is assessed through teaching publications, teaching experience (number of years teaching) and sometimes questionnaires passed to students to assess their opinion about a teacher’s performance. Students’ opinions matter, but it should be measured by avoiding any conflicts of interest that exists when surveys are given during the teaching period. Surveys should be given after the teaching period, so that students develop a critical perspective opinion about the lecturing they have received.
Thus, the teacher who gives students higher scores usually gets better results in assessments. The number of years teaching is no guarantee of knowing how to teach a class. An example of the teaching class needs to be evaluated with an expert committee.
A good teacher can be identified only by checking a class in vivo, and by knowing not only the technical information transmitted, but also how to introduce the subject, how to motivate and how to proceed with the teaching-learning process. These issues are not considered by the MMM.
(ii) Measuring inputs rather than outcomes.
As followers of the MMM need to obtain numbers, the quantification of easy indicators never ends. This even implies the confusion of outputs with outcomes. This misleading behavior is embedded in the Spanish university environment. Hence, those people who have not been qualified for promotion, have research grants and are starting out in their academic life, attribute themselves with quality and excellence. In this context, captured economic resources are the equivalent of excellence. Gaining grants is acknowledged by the university and by government agencies of evaluation.
Excellence is the habit of doing a job well and sustainably, according to Aristotle [
53]. Money implies neither quality nor excellence. Another thing is rectors’ love of money incomes.
This applies to not only at the individual level, but also the organizational level. Individuals argue that they are excellent because they have economic resources gained through grants and competitive projects funded by public authorities.
Nobody complains about the confiscatory behavior of the business manager of a university. Commissioned people must pay everyone from their pockets every time they travel, while rectors spend large amounts of money on nonsensical celebrations, like the university’s anniversary, which is often younger than many of commissioned people are.
Why? This is not because working people at the university agree with this procedure. However, as the university only breathes the air of amateur politics, those who could complain represent the majority, but they prefer to remain silent so as not to risk a project, a position, or a medal. As I wrote in [
13], the democratic election of rectors is an unfortunate error that stems from excessive democracy after Franco’s dictatorship period. Rectors do not make right decisions, but convenient ones, to aid their interests and their public image, and these decisions are far removed from public or patriotic targets.
(iii) Degrading quality through gaming and cheating.
Quantification is seductive because it simplifies and organizes knowledge by offering numerical information that allows easy superficial comparisons to be made between individuals and organizations. In fact, quantitative indicators are used to accredit individuals and to build a ranking of universities [
9]. Yet such simplification may lead to distortions as making things comparable often means they are stripped of their context, history and meaning [
34].
This results in information looking more certain and credible than it actually is: caveats, particularities, ambiguities and uncertainties are peeled away, and nothing does more to create the appearance of being certain and objective than expressing information numerically [
37,
38]. Campbell’s and Goodhart’s above-cited laws warn about the inevitable attempts to game the metric when much is at stake. Gaming comes in a variety of forms.
The main focus of dysfunction is often fraud in some official authors’ multiauthor academic products: paper, book, or PhD dissertation supervision. The standardization metric approach, without the candidate’s personal judgment, is absolutely unable to identify this frequent dysfunction.
A naive glance at this fact may lead us to think that there is no advantage for all the remaining authors if an unrealistic author is included in the publication. In fact, this is naive because no teams operate as a clan. If some cooperators are at the top of their promotion level, they really do not lose anything at all by being a co-author with others, regardless of them being students or colleagues.
In fact, each unrealistic author improves his/her curricula, and remains faithful to some corrupt boss. In addition, unreal authors increase the clan’s dependence and they allow themselves to cooperate. This gaming is systemic, not accidental, in the Spanish university, which is full of mediocre people in all categories as a result of the soft MMM operation.
A second gaming dysfunction level is the repletion of ideas in several publications in some hidden way. If a committee does not invest time in carefully reading all a candidate’s publications, it is impossible to identify this repletion. Credit placed in journals is empty for several reasons. The first is that a referee’s job is not done professionally at a high rate because it is hard work that goes unacknowledged and is, thus, understandably quite disregarded [
49].
Another dysfunction is the market of citations; see [
30,
36,
48,
49]. As the citation index is incorporated into metric evaluations, and as Campbell’s and Goodhart’s laws warn, there are people who merchandize citations to improve their citation index scores. Some simply ask you to include them in the citations of publications. Someone’s lack of scruples with regard to citations can also be seen when they play the referee’s role, as they incorporate the obligatory inclusion of several citations of themselves into their report. I have personal experience of operating as an editor, where a referee claimed a “sine qua non” condition of acceptance; the inclusion of 13 references by himself.
Gaming examples are not exceptional, and even laureated people with honorable prizes are at risk of such practices. Some authors claim their authorship in 30, 40, or even 50 papers, in good journals, per year. In some places, bosses force all their team members to sign him/her as the author of most of the team’s publications [
30,
38,
49].
In other cases, the strategy is more equalitarian for as many members as possible to achieve the top category. For instance, in line with the clan strategy, one helps to first select someone to be promoted. Having met that target, the next person on the clan’s preferences list is promoted, and so on.
The strategy is to artificially choose someone to be the co-author of all the publications, who is selected in order of preference. Thus, the MMM is not only based on personal judgment, but also comes with standardized mechanical criteria accredited to all clan members. The culture of speed and quantity boosts unethical behavior, as Campbell’s and Goodhart’s laws explain.
(iv) Improving numbers by lowering standards.
Spanish public universities are financed partly by the number of graduates in their first year. This quantitative criterion appears, in principle, appropriate insofar as the university’s success should be related to its success, shown by the number of its students gaining a degree. However, this clear example of pay-for-performance suffers the effect of Campbell’s and Goodhart’s laws.
In fact, the indicator measuring the number of graduate students is misleading and generates a cobra effect because rectors spread the idea in all schools or faculties that the number of graduate students should be high because budgets depend on this. Deans of faculties transfer this idea to all lecturers, and if the number of graduate students is lower than expected, they are alerted and advised to improve their success rates.
As student culture is to select the minimal effort itinerary, the conclusion is that, in order to maintain the success level, one has to cut the level of requirement each year by lowering the standards to pass each subject. This conclusion means a permanent slowing down of the standards of requirements. Hence, one experiences a lowering contagion wave in agreement with both the rector and dean’s wills. Implementing lower requirements is consummated yearly. Public universities are institutions that provide increasingly devaluated degrees.
Negative MMM effects do not relate only to the slowing down seen at the academic level that each experienced teacher has experienced but, more importantly, to the involved minimal effort culture that produces a fragility of character, lack of resilience to get over the natural obstacles of real life and weakness of character induced by mollycoddling education, which all contribute to increased mental troubles in young people [
54]. It is worth pointing out that the high unemployment rates in Spain contribute to the labor shock because people eventually graduate, and even those who get a job are very badly paid. It would seem that the distortion of young people minds’ due to the MMM effect and the minimum effort culture also occurs in the USA, as found in [
44].
In addition, as the level of interplay between the university and companies lowers and is not compulsory, graduate students do not know the tasks they must perform in a job and, as a result, jobs on the market pay lower wages. Companies are not willing to pay high wages to individuals who ignore their job. We should also take into account the high number of graduate students because, in Spain, families are influenced by the Veblen effect [
3] and prefer to send their children to a public, but also cheap, university. However, there being more university students in the public sector also depends on the way the economy behaves and on it being steadily maintained in the private university sector [
23].
This is the consequence of a wrong financing criterion that stems from the government, and is stimulated by rectors, deans and most research and teaching staff, who are made prisoners of the model.
The MMM’s dysfunctions and the pay-for-performance criteria are listed above. The teaching and research staff at a university are stressed by more and more demanding metric indicators based on the quantities of production items [
46,
55]. As rectors and vice-rectors of research are fooled by rankings, and are also driven by the same metric model, the demand for numbers increasingly grows. So, researchers become increasingly stressed, and teaching becomes more disregarded, because research is less important than researchers being promoted [
46,
56,
57].
Note that very few people have the freedom to disregard all of the metric criteria because, apart from full professors, the remaining teaching and research staff need to be evaluated by the ministry agency and are, thus, forced to follow the mainstream, unless the paradigm changes.
Thus, lecturers who wish to be promoted become slaves of the metric evaluation process. The only ones who are free from this pressure are full professors. However, as they have students and cooperators who find themselves in the processes of promotion and constant accreditation, even full professors feel forced to stick to these rules.
So almost everybody is forced to follow the wave of metrics. The very few full professors who have the possibility of being free know that if they do not keep quiet, they go against the policy of both the university and government agencies. They know that they may lose a project, a position or a medal. In fact, we are witnessing a “too big to fail” situation.
We are in fact suffering a metric management epidemic whose transmission vectors are rectors and ministerial agencies, and whose victims are not only the majority of teaching and research staff, but also students [
57].
The contagion extends because the younger people are, the higher the influence of the metrics is on them. The worst fact is that the youngest learn the message of quantity, speed and projects having no risk in terms of guaranteeing productivity, which is exactly the opposite of the advice given by Ramón y Cajal [
50]. This kind of supermarket model based on quantities, metrics and rankings does nothing else but deteriorate both the research and education quality, and stresses everybody [
58].
The nonsensical labor stress motivated by this research competition with regard to rankings and quantities discourages excellence, an effect of the quality and slowness habit [
59]. The MMM promotes quantity and speed, and is most unhealthy in both physical and mental terms. Rates of excessive working, depression and separations soar.
To the stress experienced from research competition, we must cover the asphyxiating bureaucracy that is partly motivated by the New Contracts Law [
60], and partly by the iron instructions given by the university’s business manager to the administration staff which, in practice, works more against the teaching and research staff than it does to serve. Actually, the rules imply that anyone commissioned to travel is owed service commissions, but in fact they are made to pay with their money. Thus, many researchers decide not to travel, and are not motivated to work or apply for projects.
The rates set per country to refund the accommodation and subsistence allowances to commissioned people incurred while traveling have not been updated for 14 years in Spain [
61], and commissioned people frequently pay part of the compensation. This issue does not appear in the rankings, everybody silently complains, but nobody does so publicly. Some kind of fear stops critics because, in some cases, a complaint coexists with administrative repression.
Rectors should acknowledge in international G7 meetings that one talks of not only a country’s gross domestic product (GDP), but also the happiness index of countries. They could take this index into account at universities because the deterioration of the welfare inside universities is considerable. The situation would probably be different if lots of people complained, but fear or interest paralyzes them.
Of course, while the business manager’s rules harm commissioned people, the rector spends big money purchasing expensive rankings [
14].
The university environment cuts the excitement of starting new projects short. For the reasons and effects explained above, the actual model is unsustainable. In fact, in the USA, where the model was created, it is no longer applied because it failed in managing the Vietnam War.
However, deterioration goes on in the Spanish public system. Nobody pays attention to Campbell’s and Goodhart’s laws, and the public system fits the MMM in every respect.
The question is, will the Spanish public system survive with this model, which was invented in the USA, and was rejected and has been questioned since the 1980s? We will attempt to explain this in the next section.
4.2. Alternatives to the MMM in the Public Higher Education Spanish System
The MMM is not only applicable to public higher education systems, but also spreads in most management activities, with different degrees of willingness and intensity in many countries. Its impact is due to contagion because it is spread by organizational leaders, and is more or less imposed on organizations throughout incentive policies. A quantitative ideology exists that should not be applied loosely because it involves serious risks wherever it is implemented. For instance, the MMM applied to assess police officers’ performances by measuring the quantity of discovered crimes may easily produce the consideration of innocent citizens as guilty ones. Another example of an MMM application would consist of a surgeon making evaluations based on the number of successful surgeries, which would tempt the surgeon to decline a high-risk surgery. The MMM is not a good alternative to judgment, and using metrics requires judgment about whether to measure, what is important and if it is measurable, and how to evaluate the significance of what has been measured.
The questioning of the MMM in the USA was started by the military staff of the US army [
62]. Robert McNamara, the MMM intellectual creator, was removed from his position as Secretary of Defense in 1968 because of a lack of accuracy in the metric quantitative indicators used to analyze the development of the Vietnam War, which McNamara admitted in his own bibliography [
45]. Then he moved to the World Bank, where he presided until 1981. During this period, he spread the ideas of the MMM around the world. This is why, albeit questioned by the military management and the Pentagon [
62], the MMM was exported abroad, but was also implemented in the USA in the education and economic sectors. This is natural because of the political influence of World Bank policies. Either willingly or by contagion, and more or less consciously imposed, the hegemony of the USA spread the new methods in the western orbit of influence.
The MMM in Spain involves a complete set of nested measures from the national government, others from local governments and, finally, internal ones inside one university where the institution’s presidentship possesses a certain level of autonomy.
Thus, it is not a matter of accepting or rejecting the MMM. Its relevance relies on the critical issues considered and their degrees of intensity in applied metrics and measures.
This means that the MMM is not a plan strictly conceived from the top and does not have a clear origin, but is a process that has accelerated in the last two decades in Spain, and has been spread by a kind of contagion effect through the Council of Rectors of Spanish Universities (CRUE).
Unfortunately, education and research have not ever been priorities for Spanish governments since democracy arrived in 1978. Proof of this is shown by the eight different National Laws of Research and Education in the last 40 years. The two main parties that have governed the country, the left-wing party (PSOE) and right-wing party (PP), have been unable to reach an agreement to build a stable research and education law.
This is one of the reasons why the MMM is a product of the national agencies of evaluation of research and education, mainly of the CRUE. The drivers of the CRUE are much stable than those of the national government, which more or less change every 8 years with some exceptions. In fact, when the national government changes, the names and competencies of the ministry or ministries in charge of research and education also change. Rectors may survive changes in national government composition, as does the CRUE, but are presided by the Spanish Ministry of Education and Research Affairs.
The core damage of the MMM to the SUS lies in three main issues, for which short-term alternatives can be implemented. We go on to explain all of this damage and its consequences (I-III). From these three main causes, other indirect damage spreads to the overall public higher education system, including public universities. Apart from presenting them, we suggest alternatives from the MMM based on the personal face-to-face evaluations of individuals.
We provide mid-term alternatives to the MMM (IV).
- (I)
The university’s financing system, depending on the local government
- (II)
The student evaluation exam to access a public university
- (III)
The National Agency of Quality and Evaluation of Research Criteria
- (IV)
Mid-term alternatives to the MMM
(I) The financing of a given university depends on the local government, and is closely linked with the number of students who graduate every year insofar as the more graduated students, the more financial resources for a university. Thus, according to Campbell’s laws, the financing criterion is corrupted by the political ambitions of rectors, whose management is conducted by monitoring wrong university rankings. The direct consequence of this financing system is that learning processes lack rigor due to the progressive lowering of expected levels, and the removal of the effort-based system in HEIs. This fact deteriorates university students’ knowledge levels every year.
An alternative to the current MMM would consist of changing the criterion for financing universities adopted by the local government by making it independent of the number of students who graduate. There is a plurality of options to avoid the metric “number of graduated students”. Unfortunately, it depends on the local government, which also makes the mistakes of the MMM at the government level. A reasonable alternative financing criterion could be linked with the number of posts offered to students because it is clear that expenses are related to the number of students.
Independently of the metrics used by the government to finance HEIs, rectors have the possibility of not misleading the financing criterion “number of graduates” and of preserving the level of effort required of students, but this does not happen. Indeed, in Spain, rectors are politicians who are elected like professional politicians, and the only difference lies in the electoral body being made up of all university members, including students, instead of citizens.
The effect of the “number of graduated” metric is questionable in line with Campbell’s laws, which essentially regard potential wrong human behavior. Rules and laws should take into account human weakness and attempt to avoid misleading human interpretations. It is understandable that rectors make efforts to increase their university’s budget, which is probably not sufficient, but the financing criterion should avoid rectors feeling tempted to increase income by lowering the quality level expected of students.
Another alternative to the MMM is to delay questionnaires for teacher evaluations until at least 1 year after students were taught by lecturers for the last time. Generally speaking, the teaching performance bias relies on the effort required of lecturers to teach students. Hence, the lecturers who apply an effort-based system are penalized with the poor evaluations that their students score them with. There is a clear conflict of interest in students’ opinions of teaching quality while the course is underway. It is not necessary to carry out an experiment, but only to know human passions [
63].
(II) Another metric indicator to be reviewed in the Spanish HE system is the composition of the grade that allows students to enroll for a bachelor’s degree in a public university. Presently, this degree is based on two components: the mark obtained by students in the last 2 years of high school and the mark obtained by the compulsory exam to gain access to a university.
The exam is a common one for all students in a given political district, independently of the bachelor’s degree that students wish to study. It does not matter if this is civil engineering or foreign languages and literature. This fact, in theory, should influence the type of high school studies, but what actually occurs is the following: to be scored a better historic mark, as the exam to gain access to university is nonspecific, students choose high school studies that allow higher marks. When students select a degree that requires, for instance, mathematics, students frequently do not choose the appropriate high school studies, but easier ones like humanities instead of scientific ones. Once again, Campbell’s laws appear given the use of a wrong metrics indicator.
What would a possible alternative be? There are several: each degree may have a specific serious exam, or one can join several related degrees and apply for the same serious specific exam for several closely related degrees. In these exams, historic scores should have no influence or, if one wishes, historic scores should be appropriate for the high school.
However, it is nonsense to suggest that literature and foreign students sit the same exam as civil engineering students. To avoid student complaints, the exam sat to gain university access is so easy that 95% of students pass it every year.
In addition, the exam to gain access to university is essentially the same from one year to the next. Instead of high school teachers teaching the official program, they train their students to pass the exam to gain access to university. Why? Because this ensures the high school a good position in the ranking of local high schools.
(III) The third alternative concerns lecturers’ evaluations to obtain accreditation at several permanent university levels. This is related to them occupying the tenure position or promoting a full professor level.
For the last two decades, the accreditation system has been based on candidates’ mechanic and impersonal evaluations, with a number of items concerned mainly with the appropriate number of them in ranked journals [
64].
Once the candidate achieves accreditation, this is a sufficiency certificate for a category, and the university where the candidate works selects a committee, which is suggested by the candidate him/herself. Typically, there is only one candidate. So, his/her promotion evidently takes place.
Where do problems arise? As no serious study is available of the papers co-authored by the candidate, there is never any guarantee given by anyone during the process that the candidate is even the real author of papers with three authors or more. A level of repetition of the same idea hidden in several publications is also frequently found, but will not be recognized without serious research, at least of a personal type [
65]. It does not suffice to appeal to a journal’s quality because when one manages a paper, the editorial team does not know if ideas are repeated in another paper submitted elsewhere. So, there is no reliability, not even about the candidate’s real authorship.
The present metric’s automatic process avoids an old problem that we have experienced in Spain for more than 30 years, where the body of full professors is limited, and the influence of some families acting like a “mafia” by means of a national random draw of the candidates who have met the requirements to form part of an evaluation committee. This system today incurs no risk of subjectivity because the size of the potential committee members is uncontrollable.
All systems are imperfect. However, since 1975, Campbell’s laws have emphasised the risks of using quantitative methods to evaluate persons. Negative effects do not finish with a lack of morality or the reliability of a given candidate’s real merits. The issue becomes more complex because the local incentives of universities themselves are addressed to improve the quantity of research production, which causes collateral damage like: not taking the research risk to guarantee productivity; teaching young researchers and PhD students these overproduction habits; disregarding the teaching activity because it is not regarded as a relevant item for promotion.
What is the alternative? This is yet again a matter of degrees and culture. The main issue is to play down rankings, which is
per se a business [
14] of little credibility. Yet it depends on not only leaders’ behavior, but also on both leaders’ and followers’ morality [
66]. When scholars are advised about metric excess, and despite admitting wrongful behavior, they still reply that they are prisoners because their research students will be evaluated by the wrong MMM. This position is not morally right because, if it is true that they cannot change the present situation, the situation will alter if they change their attitude. Collateral damage, such as a reduced risk of research projects, guaranteeing productivity, and setting a bad example for young colleagues, is important enough to deserve a change of attitude.
In an attempt to be constructive, what data-use practices should people cultivate so we can appropriately employ data?
Before answering this question, we should think about what is important to evaluate and then, what is countable, or not, because probably not everything that counts is countable. How should we count? For example, vocation, discipline, generosity, the capacity to generate eagerness, patience, stamina, tolerance, integrity, credibility, self-esteem, awakening illusion, encouraging people, public speaking, the capacity to assume risks and face challenges. Evaluating these features is difficult, but not to the extent that they cannot be taken into account during the evaluation process.
All of these suitable characteristic traits are important for performing the public jobs of lecturer and researcher, and can be identified only by the personal interaction and personal development of the candidate to be evaluated by an evaluation committee [
63,
67].
Experience is not only important, but is also a suitable characteristic that can be evaluated and measured by carefully measuring its reliability; for instance, the real degree of participation in a multiauthor publication, or the same idea being repeated in several publications [
38,
49,
50]. It is possible to measure teaching and research experiences using data. Excellence is the habit of doing a task very well. The search for quality is linked with slowness, while quantity is usually coupled with speed and a low risk. Thus, summarizing data-use practices is profitable to evaluate, but needs to be conducted by taking care of true reliable data. In any case, experience is a minimum requirement.
The stimulus of quantities is damaging. However, personal evaluations are essential to evaluate a candidate to ask about the above-mentioned suitable uncountable characteristics.
In order to avoid today’s number-swamped and audit-crazed academic life, an alternative system should stimulate both the data of countable items and great challenges, even though they are unsuccessful. There is an overdose of irrelevant repetitive production which, apart from all the above-mentioned drawbacks, also leads academics to suffer stress and avoids the sufficient contemplative life needed to make great achievements. The present ideology of quantification and speed drives mediocracy. Quantitative items should be upper-bound and stimulate great challenges.
It is important to point out that the MMM is not something abstract that is applied to a country with a specific cultural factor, which means that many circumstances can be shared by another country, but at different intensities. Some people not being ready to totally accept the criticism of the MMM because they feel they prisoners of the system shows that it all depends on the decisions of political and academic powers to some extent.
The possibility of changing the MMM depends on political power when it draws up public university budgets, or because political power manages the National Agency of Evaluation of Research, even in some countries like Spain, where the university rector is also politically elected by the university community.
Political decisions do not depend only on rationale. The second level of influence stems from the academic power inside universities, and not by stimulating the quantitative issues of the MMM or lowering its degree of intensity. If the two first scenarios cannot be changed, i.e., financing comes from the state or a local government, and from the university’s own policies, then one may resort to individual rebellion where each individual scholar can change his/her behavior with regard to their influence. Individual changes will resist through colleagues’ laziness, interests, vanity or lack of courage, but the contagion effect produced by convinced individuals is not useless at all, mainly when important researchers support the strategy. Each individual scholar’s individual initiative impacts both teaching and research because of the contagion effect. The hardest obstacle is starting the fight against the MMM and diffusing the social damage.
(IV) The following alternatives are addressed to discuss the potential changes in the paradigms of quantification by qualification: the amount of productivity items and their relevance; the speed achieved by slow behavior; pay-for-performance and intrinsic motivation; appreciating multidisciplinary activities without disregarding the quality of monodisciplinary activities. How can these ideas be implemented?
With a change in paradigm, the important issue is not how much or how many, but why and for what. A potential candidate to be promoted at a university should successfully answer in person these questions:
- (i)
When and why did you decide to become a teacher?
- (ii)
What important contribution, and why, do you make in the field?
- (iii)
What are the purposes for you giving a class and how do you motivate students?
Questions (i) and (iii) are driven to identify the intangible suitable character capacities related to vocation, student motivation, courage, stamina, generosity, public communication, etc., that are countable by metric indicators. These issues are related to teaching and, of course, we should add the quantifiable items involved in the MMM, such as the years of teaching experience and degree of satisfaction, as well as relevant teaching publications. Issues related to (i) and (ii) cannot be undervalued because they are essential and have a strong impact on all students and the self-effort culture instead of on protection and coddling.
Question (ii) is extremely important for changing the mindset of quantity and speed to a suitable one for the relevance, immense challenge and interest of research. Appealing to the number of publications in a list of journals, which has been the case to date, is simply not enough.
The mere effect from candidates answering questions about these issues already has a clear impact. The expected potential claim made after any change, and after years of applying the MMM, is expected and deserves an answer. In fact, how objective is the evaluation of these unquantifiable issues?
The answer lies in the evaluation committee’s qualified career and experience that deserve credit within human limits.
Finally, research experience needs to be considered, but also teaching activity, and only as a minimum sufficient threshold, by justifying both achievements and relevance. The minimum threshold may differ for each category, and is properly bound and weighted. The difference with the current MMM approach is that credit is not now deposited in some impact ranking of journals, but relevance must be proved and, of course, the repetition of the same idea in different papers must disappear, which is an essential issue recommended in [
50]. This means that the quality of a paper is not automatically granted by a ranking of the journal where the paper is published.
Once the relevance of a publication is personally justified in the presence of a committee, clearly many current publications would disappear because this would be the way that most of an author’s quoted misconduct would be made transparent. It is very different to show a list of publications where nobody pays attention to justify their relevance and innovation to a committee.
We believe that evaluations in universities need to change urgently and we hope that some ideas in this paper can help to improve them.