Next Article in Journal
Three-Dimensional Assessment of the Effects of Kinesio Taping on Mandibular Condylar Volume with Skeletal Class II Malocclusions
Previous Article in Journal
Optimization Method for Stiffened-Plate Layout in Box Structures Based on Load Paths
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced FMEA Methodology for Evaluating Mobile Learning Platforms Using Grey Relational Analysis and Fuzzy AHP

by
Seren Başaran
1,* and
Odianosen Anthony Ighagbon
2
1
Department of Computer Information Systems, Near East University, Lefkoşa 98010, Turkey
2
Department of Electronic and Electrical Engineering, University of Strathclyde, Glasgow G1 1XW, UK
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(19), 8844; https://doi.org/10.3390/app14198844
Submission received: 20 July 2024 / Revised: 16 September 2024 / Accepted: 25 September 2024 / Published: 1 October 2024
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
This study addresses a significant problem: it is difficult to choose a suitable mobile learning platform effectively since many learning platforms are readily available for users. For this purpose, the study proposes an efficient way to rank and choose the most suitable mobile learning platform by integrating risk analysis and multi-criteria decision-making methods. The selection of a suitable mobile learning platform is challenging due to the vast collection of available platforms. Traditional decision-making approaches often struggle to manage the inherent uncertainty and subjectivity in platform evaluation. To address this, we propose an enhanced methodology that combines grey relational analysis (GRA) and fuzzy analytic hierarchy process (FAHP), leveraging their complementary strengths to provide a robust and adaptive solution. The study employs ISO/IEC 9126 software quality standards to locate the most suitable mobile learning platform. FMEA is based on three risk factors: occurrence, severity, and detection. The fuzzy analytical hierarchy process (FAHP) is used to determine the relative weight of each risk factor to identify the grey risk priority number that can be calculated for each criterion. Mobile learning platforms are then ranked based on their grey risk priority number. The method was applied to five widely used mobile learning platforms with three decision-makers. In addition, the multi-criteria decision-making software was developed to aid users, educators, and administrators in their decision-making processes. The integrated FMEA-GRA-FAHP technique, using ISO/IEC 9126 standards, provides an effective way of locating the most suitable mobile learning platform and ranking them according to their reliability. This research is believed to be the only study applying an integrated FMEA-GRA-FAHP approach to evaluate the risks and quality of mobile learning platforms. The unique approach overcomes certain limitations of the standalone methods such as FMEA and FAHP, making it a valuable tool for identifying the suitability of mobile learning platforms. In addition, the study underscores the importance of inclusivity and equity in ensuring high-quality education and creating an environment conducive to lifelong learning for all.

1. Introduction

Mobile devices have played a significant role in daily life, particularly an inevitable part in improving the quality of education [1]. Students engaged in remote education and open learning have been frequently interacting with mobile learning technologies [2]. Mobile application-based platforms entail an approach of well-organized content, pedagogy, and quality delivery systems, providing the reason why mobile phones are gaining popularity. Mobile learning platforms (MLPs) should make it easier to access learning, reporting, and content management on the available mobile devices. This provides a safe and convenient learning environment. MLP has a lot of advantages in the area of learning in terms of flexibility, mobility, just-in-time learning, retention, tracking and reporting, and continuous learning experience.
Mobile learning platforms, which can also be synonymous with e-learning platforms, provide users with both technical and educational features. Due to the increase in demand and also in the diversity of such systems, the selection of the best platform that meets users’ required priorities is a challenging issue. It was stated that, in 2020, finding mobile learning applications of high quality apparently is a difficult task, considering that there are more than 567,000 educational applications already available (retrieved from https://www.educationalappstore.com/app-lists/apps-for-education (accessed on 20 September 2020)).
According to the research conducted by Google, 80% of the world’s population uses smart mobile phones. Users spend 91% of their time using mobile applications. 30% of the users use their smartphones to access mobile learning application-based platforms. It was reported that mobile learning platforms increase efficiency by 43%. Statistics evidenced the substantial importance and potential of mobile learning platforms (retrieved from https://www.edume.com/blog/mobile-learning-statistics (accessed on 19 July 2024)).
Therefore, quick, feasible solutions are required to choose the most suitable mobile learning platforms. Determining suitable mobile learning platforms can be treated as a multi-criteria decision-making (MCDM) problem. Therefore, there is a need to de-vise an independent way to select and evaluate MLPs. For this purpose, utilizing FMEA and FAHP approaches in evaluating MLPs will produce more reliable and efficient and timeliness results.
The fuzzy analytic hierarchy process (FAHP) is an elaboration of the AHP technique that utilizes numbers, fuzzy logic, and sets. FAHP has also been used in areas like evaluating learning management systems and platforms as well as in e-learning. Based on users’ views, FAHP has been utilized to make e-learning a decisive component in higher education [3]. FAHP was also used by [4] to provide expert evaluation for choosing appropriate learning management systems.
Various software quality models have been designed for the sole purpose of assessing software qualities through ISO standards such as ISO/IEC 9126 and ISO/IEC 25010. Regarding the software quality model review by [5], the ISO/IEC 9126 model is introduced as the optimal and most efficient model used for the evaluation of software qualities. The mobile learning platform is a software-based system designed to facilitate and deliver educational content and learning experiences via mobile devices such as smartphones, tablets, and other portable technologies. ISO/IEC 9126 is an international standard for the evaluation of software quality-in-use. It provides both technical and user-centered characteristics of the system, in this case, the mobile learning platform (MLP). In this research, we focused on the quality-in-use characteristics of the ISO/IEC 9126 standards with the consideration that the evaluated metrics should also be accessible for evaluation by the experts. The scope of the evaluation is limited to the following characteristics. The linkage between metrics and MLPs is given as follows:
Suitability under functionality is the main characteristic of which MLP performs it is required tasks, such as meeting the specified educational needs, such as course delivery, user interaction, and content management.
Understandability, learnability, operability, and attractiveness under the main metric usability indicate the extent to which learners interact with MLP, encompassing these attributes, such as whether users understand MLP effortlessly and how users learn to use MLP easily. This metric focuses on user-centered aspects of the MLPs. This constitutes another rationale for selecting ISO/IEC 9126 standards for the evaluation framework of the study. Since this aspect is believed to have a critical impact on the learning performances. Usability metrics according to ISO 9126 are used for predicting the extent to which the mobile learning platforms in this study can be understood, learned, operated, and attractive.
Fault tolerance under reliability’s main characteristics shows to what extent MLP could perform correctly and consistently with minimized failures and handle the errors well.
Time behavior under efficiency’s main characteristic is to what extent MLP performs smoothly with minimal delay.
For the reasons mentioned above, MLPs can be evaluated using the ISO/IEC 9126 software quality characteristics [6]. Risk management and quality reliability in laboratories require the use of failure modes and effects analysis (FMEA) in its dynamic procedure [7]. FMEA was applicable in two different stages; in the first phase, the potential failure mode was selected, and assessment and analysis were carried out on the probable risks.
In order to identify underlying failures and their effects as risks, risk priority numbers (RPNs) were employed. In the later stage, plans are proposed to implement and quantify those failures’ impact.
While in the series of making choices out of varieties of decision-making, the popular AHP approach is very well known and is also, in fact, the most utilized approach for these types of decision-making processes in which application rankings are based on certain attributes that are expressed with the help of the utilization of distinctive scales.
In this study, we aimed to improve the FMEA risk assessment methodology by combining it with GRA (Grey relational analysis) for making decisions based on incomplete information and FAHP (fuzzy analytic hierarchy process) to assign varying weights to risk factors as occurrence, severity, and detection, which was not possible with conventional FMEA and ranking alternatives. GRA is highly helpful when making decisions in situations when there is uncertainty and inadequate information.
It is expected that some platform performance measures will have incomplete or unclear data when evaluating MLPs. Since GRA compares alternatives’ relative performance across multiple factors, it effectively handles alternatives’ similarity in such settings, making it a great fit for our study. In GRA, results do not need to be normalized.
The FAHP technique is used when dealing with language issues and subjective evaluations. When expert reviews of MLPs incorporate subjective evaluations of usability, usefulness, dependability, and efficiency, this becomes significant. FAHP allows us to generate a more realistic and trustworthy evaluation by capturing the inherent ambiguity and imprecision in these expert opinions.
FAHP is designed to handle complex decision-making scenarios where a large number of factors are graded using expert opinion. Using fuzzy scales and pairwise comparisons, professionals can systematically create weights for each criterion using FAHP, which represents the relative relevance. This strategy ensures that expert knowledge and priorities are accurately represented in our review process, which is crucial for a comprehensive assessment of MLPs.
Additionally, FAHP is used in this work to assign weights to risk factors that cannot be evaluated by FMEA or GRA-FMEA techniques alone, such as occurrence, severity, and detection. Popular MCDM method TOPSIS identifies choices that are closest to the ideal solution; however, it performs best with precise data and may not perform as well with ambiguous or subjective information. In our research, fuzzy data are used to capture the ambiguity and vagueness present in expert assessments, making FAHP a superior choice.
Furthermore, normalizing data is necessary for TOPSIS, which could make comparing criteria with various scales and units more difficult. Another MCDM method that focuses on outranking relationships between options is called ELECTRE. It is frequently more appropriate in situations where there are obvious differences in the performance levels of the available options and can be computationally complex. In contrast, our study addresses platforms that might exhibit similar performance across several criteria, necessitating the use of a method such as GRA that can handle these nuances efficiently and offer a more nuanced ranking based on relative closeness.
MCDM is a mechanism for the determination of the best option from several options, given that multiple alternatives are, in most cases, conflicting decision criteria [8].
This study addresses the essential problem of difficulty in choosing the most suitable mobile learning platform since there are abundant learning platforms readily available. For this purpose, the study proposes an efficient way to rank and choose the best quality mobile learning platforms by integrating risk analysis and multi-criteria decision-making methods. This research is believed to be the only study applying an integrated FMEA-FAHP approach to evaluate the risks and quality of mobile learning platforms. What makes this study unique in comparison to other relevant studies is the fact that it implements failure mode effects analysis (FMEA) in collecting and evaluating data collected from various mobile learning platforms, which are then weighted using the fuzzy inference system (FIS) and finally ranked using the fuzzy analytic hierarchy process (FAHP).
The significant contributions of this study toward selecting a suitable mobile learning platform are given below:
  • The proposed method uses unique, integrated risk analysis and multi-criteria decision-making approaches to overcome the shortcomings of the traditional FMEA approach. An integrated FMEA-GRA-FAHP method is a combination of risk analysis and multi-criteria decision-making methods for the suitability of mobile learning platforms. The initial implementation of Grey relational analysis involved assuming that risk factors carried equal weights. In the subsequent application, distinct weights were assigned to risk factors through the utilization of Fuzzy AHP, as determined by evaluations from experts. This represents a significant advancement in the realm of failure mode and effect analysis (FMEA), as the introduction of varied weights through Fuzzy AHP introduces a novel approach to the FMEA methodology.
  • To researchers’ knowledge, this is the only known study to present the FMEA-GRA-FAHP technique for choosing the top mobile learning platform.
  • The evaluation’s results indicate that the proposed approach is an effective way to find the most suitable mobile learning platform to use.
Therefore, this study aims to employ a multi-criteria method such as FMEA-GRA-FAHP to evaluate and rank the five selected MLPs. Furthermore, the study adopts the technical requirements concerning the ISO/IEC 9126 model to generate a suitable framework for the effective evaluation of the selected MLPs using such techniques.
The research objectives of the study are presented as follows:
The study aims to propose an efficient way to rank and choose the most suitable mobile learning platform by integrating risk analysis and multi-criteria decision-making methods to remedy the difficulty of selecting an adequate mobile learning platform among many that are available.
  • Develop and propose a framework that can be effectively used to assess the potential risks and the quality of the mobile learning platforms based on the ISO/IEC 9126 software external quality characteristics.
  • Incorporating FMEA-FAHP with the use of Grey relational analysis methods to rank the five most widely used mobile learning platforms that were selected based on the Google Play Store ranks.
  • In order to aid decision-makers and educators in effectively selecting suitable mobile learning platforms a standalone software was developed.

2. Related Research

There is a long list of MCDM techniques, such as SMART, FAHP, ANP, ELECTRE, TOPSIS, etc. Normally, different methods are merged to enhance the process of decision-making.
This study aims to employ two multi-criteria methods, which include FMEA and FAHP integrated with grey relational analysis, to evaluate and rank the five selected MLPs. Furthermore, the study adopts the technical requirements concerning the ISO/IEC 9126 model to generate a suitable framework for the effective evaluation of the selected MLPs using a combination of the FMEA and FAHP techniques.
The Fuzzy AHP (FAHP) method is an extension of AHP which was developed by Saaty in 1975 [9]. The FAHP combines the fuzzy set theory and also the analytical hierarchy process, which makes it a more powerful approach for remedying problems that require multi-criteria decision-making (MCDM) approaches. FAHP is generally used as a decision-making method of prioritizing values for different experts in an MDCM [10]. In their study, they employed the Fuzzy AHP method to generate weights for risk factors. Different techniques exist for obtaining the weights from fuzzy pairwise comparison matrices. FAHP leverages a range of values instead of a range of crisp values to deal with the vagueness, uncertainties, and ambiguities associated with the decision-making process for decision-makers [10]. FAHP is a decision-making method used to analyze and rank alternatives in a hierarchal structure. The overall objective was to prioritize experts considering three different alternatives as experts (D1, D2, and D3) based on three criteria: experience, education, and individual level. The majority of Fuzzy AHP applications use the simple extent analysis method proposed by [11] because fuzzy weights are difficult to compute in the form of crisp weights. Research in various literature shows that a very effective FAHP approach is the extent analysis method. This method was introduced by D.Y. Chang in 1992. Da-Yong Chang, in his paper, vividly illustrates the application of the FAHP extent analysis method for introducing the synthetic extent value S of a pairwise comparison by applying the principle of the comparison of fuzzy numbers and also demonstrates the decision process by an example. The extent analysis method was also used [12]. They generated fuzzy synthetic extent values and applied the method of comparison of fuzzy numbers in calculating the normalized weight vectors. These vectors were used to obtain the final score of each student for breaking a tie situation and deciding the rank among the students in the case of having the same marks in an examination.
The failure modes and effects analysis (FMEA) is used to identify hazards in requirements or possible failure of software. FMEA is used to estimate the risk priorities of failure modes associated with an event or product. In the case of this study, FMEA is used to evaluate the potential failure of MLPs through the risk priority number (RPN) value. The risk priority number (RPN) is the estimated numerical value of the assessment of risk assigned to a process or steps in a process as part of its FMEA. RPN is calculated by multiplying the risk factors occurrence (O), severity (S), and detection (D) of a potential failure. [13] in his paper, he proposed the possible use and application of FMEA in the analysis stage by identifying potential hazards in a bid to develop software defenses by developing fault-tolerant or self-checking techniques to reduce the probability of their occurrence. Also, researchers in [14] used failure mode and effect analysis (FMEA) as a tool for evaluating Enterprise architecture (EA) risks. However, due to some drawbacks of the traditional FMEA, as stated by their research, instead of calculating RPN—the study prioritized EA risk factors with fuzzy VIKOR. Researchers in [10] employed the use of FMEA to compute the RPN value of Maintenance Policy Selection (MPS) of equipment using three dimensions, including Severity, Occurrence, and Detection. However, they also used FAHP to assign weights to the identified sub-dimensions, which is a further indication that the traditional FMEA process is not sufficient for ranking in an MCDM process. Due to the drawbacks associated with the traditional FMEA process, the traditional RPN calculation in the FMEA process does not consider the significance of individual risk factors to the peculiarity of the processor type of software being used. FAHP is therefore employed to enable us to assign priority to individual risk factors (severity occurrence and detection).
Various studies have employed the use of GRA- FAHP-FMEA in different aspects of evaluation research. However, before this study, no research has employed this approach in evaluating mobile learning platforms. Researchers in [15] proposed a model that involves the combination of the fuzzy analytic hierarchy process (FAHP) with FMEA to obtain accurate and timely results when carrying out maintenance, particularly in the paper industry. The model was applied to a numerical sample to show its efficiency. Researchers in [16] also proposed a method for analyzing the risks of green components in the electronic manufacturing process in a bid to improve quality control. They used FMEA to analyze the failure modes and effects of green components and used FAHP to determine the relative weightings of four factors. This helped them to calculate a green component risk priority number (GC-RPN) for each of the components that a supplier provided to identify potential risks and manage them appropriately. The most frequently used MCDM methods were summarized as TOPSIS, ELECTRE, and FAHP in [17].
The core features in the mobile learning platform must be for the MLPs to be able to offer learning irrespective of location and time. Likewise, knowing the best method for the selection criteria is another problem, and manual selections among several MLPs are both tiresome and time-consuming, which leads to poor choices. Therefore, there is a need to devise a means of independently selecting and evaluating MLPs. For this purpose, utilizing FMEA-FAHP approaches in the selection and evaluation of MLPs provides time efficiency, optimum results, and the best technique for future study. Hence, locating the most suitable mobile learning platforms using a uniquely developed framework is proposed from the standardized model of ISO/IEC 9126.
The number of studies located in the context of mobile learning literature that applied either multi-criteria decision-making or machine learning approaches is limited. There are no located studies that apply grey relational analysis or FMEA techniques in the context of mobile learning.
A study used the analytical hierarchy process (AHP) to determine the critical success factors for the successful implementation of cloud-based mobile learning as Pedagogy, Mobile Compatibility, Data Security, Learning Management, Content Quality, Portability, and Support for Teachers [18].
Another study evaluated and ranked the several sub-factors of cloud-based mobile learning through the fuzzy analytic hierarchy process (FAHP) and analytical hierarchy process-group decision-making (AHP-GDM) [19]. Another study employed a machine learning-based algorithm to select the most suitable mobile learning application for children by using textual analysis of app reviews [20]. Another study uses fuzzy-TOPSIS group decision-making with 25 criteria and 125 students as decision-makers to rank educational applications [21].
The diversity and ease of availability of mobile learning applications encouraged researchers and educators to use mobile learning platforms to enhance teaching and learning [22]. The mobile learning platforms are another perspective of implementing e-learning, which uses both the technical and non-technical ability of mobile devices to deliver content to users. In some situations, MLPs are taken as simple extensions of e-learning, which makes it another channel for delivering content; in fact, quality mobile learning can only be delivered properly when special limitations and benefits of mobile devices are acknowledged [23]. Instead, it is relevant to consider features special to M-learning when putting into perspective how to offer a quality learning experience to users [23].
Even though the International Standard Organization modeled standards (ISO/IEC 9126) for both technical and non-technical quality features, researchers also recommend that studying mobile learning platforms should be considered based on two perspectives, which are Technical and Non-Technical Approaches to the quality of mobile learning platforms. In the study, both the technical and non-technical aspects were considered within the scope of the study [24].
Platforms that aid mobile learning are defined by technologies, and the learning approach is specified by pedagogy or non-technical approach [25]. Many studies applied ISO/IEC 9126 quality-in-use standards to evaluate e-learning and mobile learning environments. In one study, ISO/IEC 9126 was linked and used in the evaluation of e-learning systems [26,27]. Meanwhile other studies focused on proposing usability aspects similar to the usability characteristic of the ISO/IEC9126 quality model for evaluating mobile learning applications [28,29].
The International Standard Organization, ISO/IEC 9126, developed a model for assessing the quality of software products such as mobile learning platforms. This model for assessing mobile learning platforms is based on criteria such as functionality, accessibility, usability, efficiency, maintainability, dependability, quality, and maintainability criteria. Researchers have developed several models to evaluate MLPs in [30,31,32,33,34,35], still, the most widely used model in research and software platform developments is ISO/IEC 9126 [36].
The quality model can be characterized as the set of features and the relationship that gives the premise for indicating quality necessities and assessing item quality. Different works have been discovered in literary research centered on program item assessment. Among the most acceptable models include McCall, Boehm, FURPS, Dromey, Bayesian, and ISO/IEC 9216 [37].
The ISO/IEC 9126 is a standard developed by the ISO for evaluating software reliability. It has proven to be very efficient in evaluating software in a bid to verify its reliability and durability after production. Its fundamental objective is to address familiar human biases that can adversely affect the perception and delivery of a software development project. Biases can range from not having a clear definition of success to a change of priorities after the start of the project. It consists of six major criteria, each with sub-criteria. These are the yardsticks used for the software evaluation process. A lot of research has employed this standard in developing a working framework for their software analysis research. Researchers in [38] sifted 21 out of the 32 software quality criteria of ISO/IEC 9126 standard in optimizing Enterprise resource planning (ERP) software problems. Also, researchers in [39] proposed a framework for evaluating both the quality and user satisfaction of Mobile learning applications for mathematics (MLAM). They used the ISO/IEC 9126 standard to generate a model for evaluating the technical aspect, which was combined with user satisfaction proposed for the non-technical aspects. In another study, ELECTRE I and quality standards rank Mobile Learning Applications for Mathematics [24].
Quality properties are classified into various leveled tree structures of features and sub-features. The most noteworthy level comprises quality characteristics, and the least level comprises quality criteria. ISO/IEC 9126 indicates six characteristics that are separated into twenty-one sub-characteristics. These sub-characteristics are shown remotely when the computer program is utilized as a portion of a computer framework, and then internal quality is achieved. The core merit of this model is that the characteristics stated are pertinent to each kind of computer program, giving steady wording for software item quality [40].
Conclusively, ISO/IEC 9126, since it is based on past activities and models, is more finalized than other models and reasonable to be utilized within the assessment of MLPs. ISO/IEC 9126 engages all pivotal attributes such as progressive structure, criteria for assessment, comprehensive expression and terms, basic and exact definitions, and one to numerous connections between distinct layers of the model. In expansion, ISO/IEC 9126 aids key decision-making efforts and also exempts expensive errors.
Functionality is chosen based on the fact that it is dependent on the application domain (education), while reliability focuses on information presentation and content in academic products. Usability is also included because it is a vital figure, particularly for client-driven applications. Productivity is additionally added because it refers to the potential of the computer program to supply utilizable work to realize its purpose. Viability, be that as it may, is cleared out according to the framework as it may be assessed either by the designer or a third party when getting to the project’s technical documentation and the source code. Even though most products sense maintainability as a critical quality attribute, only in the early stages of development were assessed. This can be in line with computer program product assessment within the scholastic space, which makes maintainability and portability less vital features.
Upon reviewing the literature, certain studies were located to apply FMEA in combination with other techniques but in different contexts. For instance, one study applied FMEA combined with Pythagorean Fuzzy Sets (PFS) and Dimensional Analysis (DA) risk assessment in product design [41]. Meanwhile, another study employed the integrated Fuzzy-GRA-TOPSIS-FMEA method to investigate the personal information security risk assessment for e-waste recycling [42]. Another study examines a combined failure mode effect analysis (FMEA) with a hybrid analytic hierarchy process (AHP) and the preference ranking organization method for enrichment evaluation (PROMETHEE) in supply chain management [43]. In reviewing the literature for this study, it was observed that no study on MLPs has applied GRA-FAHP-FMEA techniques in evaluating the potential risk and quality of mobile learning platforms. This study is unique in comparison to other relevant studies that implement FMEA in collecting and evaluating data collected from various MLPs, which are then applied to FAHP for weighting risks. Studies that combine these methods in evaluating MLPs could not be located in the literature.

3. Methodology

The study is carried out in stages; this involves the selection of MLPs, developing the framework from the ISO standards, and determining the value of the three risk factors for each risk item with the aid of an Excel template for FMEA data collection. In addition, calculating the weight factors and assigning the corresponding weights leads to calculating the grey RPN. Therefore, the evaluation and analysis of the results from FMEA-FAHP.
The procedures involved are further described in steps as shown below;
(1)
The selected criteria were based on ISO/IEC 9126 and were arranged on the first column of the FMEA sheet. And the metrics (Occurrence, Severity, and Detection) were arranged on the subsequent columns of the same sheet.
(2)
The data on the columns were manually collected, as each MLP was observed based on individual criteria. Therefore, metrics were measured based on their corresponding rankings.
(3)
The classical Risk Priority number is then calculated, which is the multiplication of the three metrics: Occurrence, Severity, and Detection.
(4)
The grey priority number is calculated based on weights obtained through the FAHP technique.
(5)
The grey RPN is ranked through the application of the method of ranking [10].
Figure 1 shows the study’s procedural flowchart. In this flowchart, the MLPs are accessed based on the criteria from the adopted framework. The risk factors for each criterion are prioritized using the generated FAHP weights of the weighted criteria for occurrence, severity, and detection, respectively.

3.1. Selection and Evaluation Criteria

The FMEA method requires the use of multiple attributes to create a list of potential failures that can be used to fix design defects and expensive bugs that can be difficult to address at a later stage [44]. To evaluate the potential risk of the MLPs, only the technical quality criteria will be of relevance. The FAHP method uses the three components of FMEA, namely, occurrence, detection, and severity, as criteria to determine the weight of each component in order to further evaluate the criteria for FMEA outcomes. In this study, the proposed criteria are shown in Table 1.

3.2. Alternatives: Sample Mobile Learning Platforms

To guarantee a representative and diverse sample, the five MLPs assessed in our study—SoloLearn, Duolingo, EdX, Khan Academy, and Pluralsight—were chosen using the following standards.
We considered the platform ratings, active user numbers, content, technology used, and ease of access to the evaluation criteria by experts.
Our selection is based on the platforms with high user numbers and widespread recognition. We inspected reviews, user ratings, download statistics, and the number of active users on the Google Play Store. There are millions of users on EdX, Khan Academy, and Duolingo. The chosen MLPs have a variety of educational content, such as courses for higher education (EdX), general education (Khan Academy), professional development (Pluralsight), language acquisition (Duolingo, SoloLearn), and coding. Including varied content MLPs contributed well to the generalizability of the study. In addition, chosen platforms should meet the requirements of the evaluation framework. Our study made sure to capture a representative sample of well-known and established platforms accessible to learners today. The selected platforms range in focus from language learning (Duolingo) and K–12 education (Khan Academy) to specific professional skills (Pluralsight). Platforms selected have multilingual supports, enabling access to learners globally. We believe that the sample of MLPs reviewed in this study is both representative and valid because we were able to precisely define our selection criteria and select platforms that are widely used, diversified, and equipped with modern technology. Our findings can guide decisions in a variety of educational contexts because the platforms we have chosen offer a thorough understanding of the mobile learning environment. There are different forms of MLPs based on some relevant features such as functions, design structure, aim, targeted users, and drawbacks. On the Google Play Store, these MLPs are digitally delivered and are at the Play Store’s subscriber’s disposal, and they may either be free or at a price. There are millions of mobile platforms designed for different purposes, which are hosted on Google Play Store, and they have categories such as games, social, finance, education, and health. Google Play Store gave users the power to rate these software platforms based on their experience using the software. Every mobile software on the Google Play Store has been verified by Google Play Protect to avoid malicious software. The rating is from 0 to 5. However, for this study, five MLPs were selected to be designed for adults and able people. Therefore, MLPs chosen for this study are popular and mostly open source. Table 2 shows the list of MLPs and their Google Play Store rating.

3.2.1. Duolingo Platform

Duolingo, shown in Figure 2, is an MLP for learning different and popular languages such as Chinese, English, Arabic, German, Spanish, etc... It has millions of subscribed users. It can help users learn from basic to advanced levels through its aided features, such as reading and audio translation. The MLP can be used anytime and anywhere as long as there is internet connectivity. The MLP has a practice test readily available while learning on it. It has over 100 million downloads and 8 million reviews.

3.2.2. EdX Platform

EdX in Figure 3 is developed for learning information technology and information technology-related courses. It has millions of users due to the many free courses it offers. It has almost 50,000 reviews on the Google Play Store and over a million downloads. It provides learners with courses from Harvard, MIT, and Microsoft, such as data science, IT support, project management, business and management, integrated digital media, Microsoft Professional program in IoT. It has over 1 million downloads and 49,000 reviews.

3.2.3. Pluralsight Platform

Pluralsight, as depicted in Figure 4, is also an MLP for learning IT courses, mostly programming languages and technologies such as AngularJS19.0.0, HTML5, JavaScript ECMAScript 2023, Bootstrapv4.6, cloud solutions, etc. It has thousands of users across the world. The MLP is designed to be used anywhere, irrespective of location, as long as the user has access to an internet connection. It has over 500,000 downloads and 12,000 reviews.

3.2.4. Khan Academy Platform

Khan Academy, given in Figure 5, is an MLP developed for learning information technology and information technology-related courses as well as other fields of study. Khan Academy has millions of users due to the availability of several free courses. It provides learners with courses such as Math, Economics, Finance, etc. It has over 10 million downloads and 105,000 reviews.

3.2.5. SoloLearn Platform

Another popular MLP is SoloLearn, given in Figure 6, which is built to aid and support beginners in code learning. It courses and IT-related courses such as HTMLv5, CSSv3, JavaScript ECMAScript2023, Ruby on Railsv7.1, Pythonv3.13, Javav23, C++v20, Kotlinv2.0.0, etc. Though it does not support offline functionality, it tracks student performance. This MLP has over 5 million downloads and 387,000 reviews.

3.3. Fuzzy Analytic Hierarchy (FAHP) Procedure

In the 1980s, Thomas Saaty presented an analytical hierarchy process for solving complex problems by breaking the problem down into subgroups that were hierarchically arranged [9]. Even though AHP has merit in evaluating multiple criteria, it has shortcomings that the decision maker came up with, as it substitutes for pairwise comparison. Deng also observed the fallbacks of the AHP method when comparing alternatives [46]. Also, in its comparison procedure, the scale used for judgment in this technique did not extract the meaning of unpredictability and delinquency in the data [46].
Therefore, Chang presented an improved AHP referred to as the fuzzy analytical hierarchy process [11]. The FAHP technique was used in this research in order to avoid fallbacks. Fuzzy offers more versatility to expression. There are several types of fuzzy numbers: triangular, trapezoidal fuzzy numbers, etc. [17]. The triangular fuzzy number contains the group of three real integer numbers starting from the least, most ascertain, and the highest weights for the triangular fuzzy. However, trapezoidal and triangular fuzzy numbers are said to be the most popularly used. Triangular fuzzy numbers were used in this study.
The procedures involved in the improved AHP are described as follows, as presented by [47].
Let’s declare G as a set of objects, where G = {G1, G2, G3,……..…,Gn} and P is a goal set where P = {P1,P2,P3,………..Pn}. Based on this technique, the individual object is taken, and the extent analysis set for every goal is carried out respectively. Therefore, for an individual object, the respective m extent analysis values are acquired.
In this case, the individual value of M g p j where j = 1, 2, 3, 4,…..,m are triangular fuzzy number. The steps for Chang extent analysis are described as follows:
Step 1: Fuzzy synthetic based on ith objects is stated by an equation as follows;
S i   = j = 1 m M g p j i = 1 n j = 1 m M g p j 1
To acquire the value of j = 1 m M g p j , m extent analysis is used based on fuzzy summation.
j = 1 m M g p j = j = 1 m l j ,   j = 1 m m j , j = 1 m u j
i = 1 n j = 1 m M g p j = i = 1 n l i ,   i = 1 n m i , i = 1 n u i
Equation (2) is a vector that is then reversed to acquire Equation (4), shown below;
i = 1 n j = 1 m M g p j 1 = 1 i = 1 n u i , 1 i = 1 n m i , 1 i = 1 n l i  
Step 2: The degree of possibility depends on the values of M1 and M2, of which, while M1 remains (l1, m1, u1) and M2 remains (l2, m2, u2) and can be evaluated as stated in Equation (5);
l i   = m i n k a i k ;   m i   = 1 k k = 1 b i k ; u i j = m a x k   c i k
Therefore, M2  M1 for the degree of possibility, which is stated as follows:
V ( M 2 M 1 ) = h g t ( M 2 M 1 ) = μ M 2   ( d ) = 1 ,                                                             m 2   m 1 0 ,                                                               l 1       u 2 l 1 u 2 ( m 2 u 2 ) ( m 1 l 1 )                   O t h e r w i s e
V(M2  M1) is equal to d; this indicates that the highest convergence within M1 and M2 is the ordinate of D, as shown in Figure 7.
In Figure 7, d defines the highest convergence point among μ M 1 and μ M 2 . Comparing M1 and M2, the values of V(M2  M1) and V(M1  M2) becomes relevant.
Step 3: So, the degree of possibility of a single convex fuzzy number that could be greater than the value of k convex numbers Mi where i = 1, 2, 3,…….., k; is stated as follows:
V ( M   M 1 ,   M   M 2 , .   M   M k ) = V [ ( M   M 1 )   and   ( M   M 2 ) ,   and   ( M     M k ) ]
minV(M > Mi), where i = 1,2,3,…k
Suppose   d ( A ) = min V ( S i     S k )
Therefore, for k = 1, 2, 3,…n where k i; the weight vector can be given as;
w = d A 1 , d A 2 , d A 3 , d ( A n ) T
For Ai: i = 1, 2, 3,…,n-elements
Step 4: The weight vectors are normalized, as stated in Equation (9); the weight is a number, but note that it is a non-fuzzy number.
w = d A 1 , d A 2 , d A 3 , d ( A n ) T
Step 5: The Grey RPN [45] will then be calculated based on the equations below;
Sr = i = 1 n w s S i ; Or = i = 1 n w o O i ; Dr = i = 1 n w D D i ; where i = 1,2,3,….,n
Therefore, Grey RPN is
RPN = i = 1 n W s S i + W O O i + W D D i
Step 6: The computed information is then ranked based on the order of preference.

3.4. Proposed Evaluation Model

The study aims to evaluate the quality and risks associated with mobile learning platforms with the help of FMEA-FAHP techniques. Several other related models are related to this study in the quality evaluation of software previously discussed in the literature. This study is unique in that there has been no previous combination of FMEA and FAHP in the evaluation process for the inspection of mobile learning platforms. This study embraces both the technical and non-technical perspectives as the core reason for establishing the assessment of the mobile learning platform. ISO/IEC 9126 technical and non-technical aspects constitute the essentials of the model for this study.
This study adopts the following model for assessing mobile learning platforms, which was scrutinized from the extant literature. Figure 8 indicates the adopted model for assessing mobile learning platforms.
Seven criteria were identified and were used in this study. The reason for the selected criteria is that the FMEA can be utilized to evaluate technical quality and functionality. The adopted framework in this study emphasizes key characteristics of the ISO/IEC 9126 standards, namely functionality, efficiency, reliability, and usability. The steps followed were given in pseudo-code:
  • Decide about the set of criteria for evaluation from ISO/IEC 9126 standard
  • Decide about the experts involved in the evaluation
  • Decide about the alternatives of MLPs for evaluation
  • Decide about the linguistic scale and their triangular fuzzy reciprocal scale
  • Experts establish a decision matrix regarding the weights for occurrence, severity, and detection
  • Fuzzy pairwise comparison matrix was established from experts’ decision matrices
  • Synthetic extent values were established from a Fuzzy pairwise comparison matrix
  • Weights for Occurrence, severity, and detection were calculated with respect to Synthetic extent values
  • Experts evaluate five alternatives concerning severity, occurrence, and detection using FMEA rating from 1–10
  • Total RPN value calculated with respect to each alternative
  • Calculated weights for severity occurrence detection applied for each alternative for each criterion to calculate Grey RPN
  • Alternatives yielding to total Grey RPN for each alternative are ranked from least to most.
  • Alternatives with the least Grey RPN value yield better results in terms of evaluated aspects.

3.5. Selection and Evaluation Criteria for MLP Based on Adopted Framework

FMEA implementation goes through a series of processes that involve the development of an FMEA data-gathering template. Software analysis using FMEA is achieved by evaluating each criterion derived from the developed framework and estimating their potential failure modes. These failure modes are called risk factors. In this study, the criteria for evaluation are adopted from the developed framework ISO/IEC 9126 standard. These attributes represent a detailed model for evaluating any software system [48]. Additionally, these criteria are used to evaluate the quality of the mobile learning platforms.
The adopted model from the ISO/IEC 9126 standards has seven sub-criteria: time−behavior, fault tolerance, understandability, learnability, operability, attractiveness, and suitability. These sub-criteria cut across four main characteristics: efficiency, reliability, usability, and functionality. The choice of the criteria listed in the adapted framework was based on the criteria modified specifically for mobile learning platforms. The characteristics and sub-characteristics of the adopted framework are given earlier in Table 1.
In this study, the sub-characteristics selected for evaluation are suitability, fault tolerance, understandability, learnability, operability, attractiveness, and time behavior.
(1)
Suitability: The suitability of software refers to the essential functionality characteristic. It helps to answer the question, “Can the software perform the tasks required?’ It enumerates a set of attributes used for the explicit assessment of functions to prescribed tasks, and to determine their adequacy for performing the tasks. It answers the question; “How appropriate are the functions of software applied to the specification?”
A quality model is used to determine the quality of the software. Usability, which is one of the sub-characteristics of the quality model, is defined as the capability of the software product to be understood, learned, and used and the ability to provide visual appeal under certain specific conditions of usage (the effort needed for use).
(2)
Fault tolerance: Fault tolerance helps to determine if the software is capable of handling errors. It refers to the ability of a system to withstand component failure. It is defined as a set of attributes used to assess a software’s ability to maintain the desired performance level in the unlikely event of operational defects or infringement of specified interfaces.
(3)
Understandability: This refers to how well or easily a user comprehends the functionality of a system. System functions and things like layout consistency, functions of buttons, clear prompts for input, consistent use of terms throughout the system, and proper documentation will all contribute to the ease of the system being understood.
(4)
Learnability: The learnability of software depicts how easily a user can learn to use the system. Although learnability is very similar to understandability, a major sub-characteristic is that it has to do with how long it takes users to learn how to use specific functions and the effectiveness of help systems. It is the learning effort required by users of various levels of difficulty during the learning process.
(5)
Operability: This is also one of the sub-characteristics of usability. It characterizes the ability to use the system with minimal effort. The operability of software depends on whether a user can operate and control the software in a given environment.
(6)
Attractiveness: Factors such as screen design, color, and general software interface are considered in determining the attractiveness of software.
(7)
Time Behavior: This characterizes how quickly the system responds and processing time when carrying out functions like file upload and download. The rate at which transactions are being processed and its ability to give an appropriate response time for a given throughput.

3.6. Failure Modes and Effects Analysis (FMEA)

FMEA implementation goes through a series of processes that involve the development of an FMEA data-gathering template. The template includes tables that show information gathered based on individual alternatives considered. Where to enter data, guidelines on the type of data to be entered, and when those data are collated. Below are pictures and descriptions of each FMEA file.
The FMEA technique accesses applications based on the three risk criteria known as risk factors severity, occurrence, and detection. A heuristic scale is used to estimate the values of each risk factor. The scales are ranked from 1–10. This ranking is based on the effect of failure modes, which are further categorized into minor, moderate, and hazardous effects. From Table 3, the sections are further categorized with colors: red for hazardous effects (ranking: 7–10), yellow for moderate effects (3–6), and green for minor effects (1–2).
(1)
Occurrence: This is the probability of occurrence, which is the likely number of times the failure is bound to take place. Its ranking is from 1 to 10. From unlikely chance of occurrence through occasional failures to almost inevitable probability of failure. From Table 3, the scale categorizes risk and estimates a value for different ranges or categories of risk from increasing to decreasing order.
(2)
Detection: This relates to how well a system can recognize abnormality in its functionality. Assuming a failure occurs, a system able to assess its capabilities to prevent failure mode is called detection. It helps rank the level at which failures can be detected. This is sub-categorized into most certain to be detected, moderate level of detection, and failure is almost uncertain to be detected. It is similar to both occurrence and severity in ranking, as it ranks from 1 to 10.
(3)
Severity: This can be defined as the extent to which the consequences of failure on a software affect the internal and external functionality occurring during and after failure. The sub-category is sectioned into failures that are ‘hazardous in the absence of warning, hazardous with warning, moderate and minor severity in terms of damage. It ranks from 1 to 10.
The risk priority number is the multiplication of these three risk factors: severity, probability of occurrence, and detectability.
Risk Priority Number (RPN) = Severity × Occurrence × Detection
In the detection of potential failures of the software functionality, for example, the sub-categories of the functionality tab are considered, and the mobile learning platform is compared against the sub-characteristics to ascertain the potential failures associated with the functionality. The following questions were asked:
  • Suitability: Does the software perform its required tasks?
  • Accuracy: Does the result produced conform to the expectation?
  • Interoperability: Can the system relate or interact with a similar system?
  • Security: Is the software able to restrict unauthorized access?
With respect to the above questions, if the software is found to have any shortcomings, the expert is then able to identify if the software has potential failures associated with the functional characteristics in question via the sub-characteristics. Table 4 explains the types of questions experts use in the evaluation and estimation process for each criterion. This process is the source of the input for the system, and the estimations are approximated in Table 3.
The detected potential failure metrics are then accessed using three different metrics severity, occurrence, and detection. These metrics are called the risk factors, namely, severity, detection, and probability of occurrence. Values for severity are estimated from Table 3.
The processes involved in the application of the FMEA technique on MLPs based on this study guidelines are;
Step 1: Evaluate the process
  • Use the developed framework to signify individual components.
  • On the FMEA table, itemize the components.
Step 2: Deliberate on all the modes of potential failure.
  • What are the causes of failure and the possibility of occurrence?
  • The scale for determining the level of impact of a failure.
Step 3: Itemize the potential effects of each failure
  • The effect of failure on the other preceding processes or failures is measured and valued.
  • There will likely be more than one effect for each failure.
Step 4: Assign the risk factor rankings
  • Fundamentally, on each risk factor and implication of failure.
  • Use the ranking scale to determine the value of individual risk factors.
Step 5: Calculate the RPN and the Grey RPN

4. Results

4.1. FAHP Procedure for Evaluating MLPs

As stated earlier, the traditional RPN calculation in the FMEA process does not consider the significance of individual risk factors for the type of software. Therefore, FAHP is used to assign priority to individual risk factors (severity occurrence and detection). The results of the FAHP process are then merged with the FMEA results in Table 5 using Equation (11). This is the point of integration to FMEA from FAHP.
With the aid of triangular fuzzy numbers, the experiment was performed with five alternatives. To begin with the FAHP assessment procedure, a pairwise matrix for comparison is generated with the aid of a linguistics scale. Table 5 shows the proposed scale adopted from [49,50].
The three decision-makers’ opinions were considered in this study. The decision-makers are involved in determining the weights of the risk factors, such as severity, occurrence, and detection, depending on their experience. They also participated in the evaluation of the five mobile learning platforms using FAHP combined with GRA.
We carefully selected three experts based on their extensive experience and relevance to the evaluation of mobile learning platforms (MLPs). The criteria for selecting these decision-makers included:
Each expert has substantial experience in mobile learning technologies, distance educational tools, or instructional design, ensuring that they are well-versed in the technical and pedagogical aspects of MLPs. Specifically, one expert is a Computer and Instructional Technology professor specializing in educational technology with 13 years of experience in the development and maintenance of distance educational tools, and another expert is a senior instructional designer with a Computer Science background with experience in implementing mobile learning solutions with 4 years of experience, and the third expert is an information systems engineer with a focus on usability and system functionality in educational platforms with 7 years of experience. All selected experts have been involved in the development, implementation, or evaluation of mobile learning platforms in various educational institutions. Their hands-on experience with MLPs provided valuable insights into both the technical and user-experience aspects of the platforms. The experts use the linguistic variables given in Table 5 to generate the decision matrix separately.
The matrix for individual decision-makers, as obtained from the linguistic table and assigned to the three decision-makers, is stated as follows, where D1, D2, and D3 represent the decision-makers obtained from the linguistic scales table. Table 6, Table 7 and Table 8 show the process of obtaining synthetic extent values from pairwise comparison matrix.
D 1 = 1 1 7 1 9 7 1 1 9 9 9 1 ;   D 2 = 1 1 6 1 9 6 1 1 9 9 9 1 ;   D 3 = 1 1 6 1 9 6 1 1 8 9 8 1
C1, C2, and C3 are the criteria selected based on occurrence, detection, and severity, respectively, and the matrices are evaluated using Equation (5). Equation (5) is used to sample the opinions of all three decision-makers to produce a single pairwise comparison matrix shown in Table 6. The pairwise comparison matrix takes into consideration the opinion of all three decision-makers depicted in Table 7.
Equation (12) is used to obtain the synthetic extent value given in Table 8 that resulted from Table 9, Table 10 and Table 11; the calculation involved is shown below:
SC 1 = ( 1.2684 ,   7.2562 ,   15 )     1 40.1231 1 23.4172 1 20.6791   =   ( 0.0321 , 0.3051 , 1.4142 )
SC 2 = ( 1.2684 ,   3.6012 ,   8.1432 )     1 40.1231 1 23.4172 1 20.6791   =   ( 0.0321 , 0.1534 , 0.7622 )
SC 3 = ( 8.1432 ,   12.7141 ,   17 )     1 40.1231 1 23.4172 1 20.6791   =   ( 0.2032 , 0.5021 , 1.5913 )
Table 9, Table 10 and Table 11 show the pairwise comparison matrix obtained through the assessment of the decision-makers. The weights of each component were calculated using the FAHP method.
Equation (6) is then used to compare the results in Table 12 to obtain the results shown as follows;
V ( S O     S S ) = 0.83512 ;   V ( S D     S S ) = 0.58973 ;   V ( S O     S D ) = 1
V ( S S     S O ) = 1 ;   V ( S S     S D ) = 1 ;   V ( S D     S O ) = 0.53206
Therefore, Equation (9) is used to determine the priority of individual risk factors.
d′(Qo) = min (1: 0.83512) = 0.83512
d′(Qs) = min (1:1) = 1
d′(QD) = min (0.58973: 0.532062) = 0.532062
l = 1 3 d ( Q l ) = 0.8353783383091131 + 1 + 0.5903867040418067 = 2.4258
The corresponding risk factor weight can then be calculated using Equation (10) as follows;
σ o = 0.8354 2.4258 = 0.3444
σ D = 0.5904 2.4258 = 0.2433
σ S = 1 2.4258 = 0.4123
Researchers considered more attention to severity regarding the influence of severity in FMEA applications in compliance with the literature. Therefore, the severity tends to have the highest weight factor.
As the study continues to unfold, the merging of FMEA and fuzzy will be executed based on the assumption of utilizing different weights on risk factors. Note that the following results of risk factor weights were retrieved through FAHP techniques;
σ o = 0.3444
σ D = 0.2433
σ S = 0.4123
The values of σ o ,   ( σ D ) and ( σ S ) are the generated weights of occurrence, detection, and severity, respectively.

4.2. Results of FMEA Applied to MLPs

The developed framework was used to itemize components (criteria) on the FMEA sheet. Having evaluated possible potential failures, severity, occurrence, and detection ranking are assigned based on the impact of failure. Table 12 shows the results attained for the five MLPs, which are arranged based on MLPs because individual MLPs must be considered.
In the FMEA process, each alternative (mobile learning platform) is assessed based on all criteria presented in the adopted framework (time behavior, fault tolerance, understandability, learnability, operability, attractiveness, and suitability). The FMEA assessment factors (severity, occurrence, and probability of detection) are therefore tested against a particular software testing criteria based on the adopted framework, and a value is estimated in terms of severity, occurrence, and detection using the ranking tables for each risk factor for each of the above mention testing criteria.
The results generated from the FMEA assessment are shown in Table 12. The expert opinions are given in Table 9, Table 10 and Table 11. The average of the data collected from experts for severity, occurrence, and detection are given in Table 12, and these are the data used as input for developing the multi-criteria decision-making software.
The risk priority number is evaluated across each MLP, and the result is based on Equation (11) above. Table 7 shows the result generated from the FMEA evaluation of each alternative for risk factors. Having collected values for the risk factors, it is necessary to evaluate the risk priority number of the criteria and, subsequently, a value for the overall risk priority for each alternative or MLP. Table 12 shows the result of each mobile learning platform with the risk priority number evaluated for each criterion it is tested against as well as the overall risk priority of the platform.

4.3. Integration of FMEA and FAHP

At the traditional RPN calculation, the relative significance of these three risk factors is neglected as they are all assumed to be equal. These factors could be assumed to be equal, but this is not always the case in practice, as noted by previous research. Research has shown that the relative significance of the risk factors varies based on the software or the nature upon which it is being used.
In this study, the FAHP is employed as a decision-making method to rank and analyze these risk factors in a hierarchal structure. The objective of the FAHP is to prioritize risk factors. From Equation (11), the FAHP RPN number for each characteristic is the sum of the weighted risk factor;
i = 1 n W s S i + W O O i + W D D i
We can, therefore, calculate the grey RPN using the corresponding weight of each FMEA risk factor as follows:
The FAHP helps the realization of better and more efficient alternative ranking by prioritizing risk factors based on their relative importance, unlike the traditional FMEA RPN, which assumes all risk factors to be equal yet is not feasible in practice. The grey RPN is the risk priority number derived from applying prioritized risk factors of the FMEA process. The applied FAHP results for each mobile learning platform are given in Table 13, Table 14, Table 15, Table 16 and Table 17. The final ranking for individual mobile learning platforms is given in Table 18. From the calculated results, it is shown that EdX MLP with a grey RPN of 19.0023 is most reliable based on the quality evaluation process carried out in this research. This is followed by the Khan Academy MLP with a grey RPN of 19.4994. Plural sight is ranked the lowest in terms of reliability, considering the grey RPN value.

4.4. Drawbacks of Conventional FMEA Method

In the FMEA application, for each failure mode, the likelihood of occurrence, detection, and severity is assessed by the experts who have experienced FMEA at least once in their jobs. The experts give points of 1 to 10 for each parameter (1 is the best and 10 is the worst case). In this study, these parameters are called risk factors. By multiplying values for severity (S), occurrence (O), and detectability (D), the risk priority number (RPN) is obtained. RPN values change between 1 and 1000. These risk priority numbers help to identify the parts or processes that need certain actions for improvement. Failure modes are prioritized, and corrective measures are applied according to the RPN values. Since RPN values are obtained by multiplying severity, occurrence and detection values, different severity, occurrence and detection values could yield to same RPN values, one of the prevailing limitations of classical FMEA procedure. In addition, different weights are omitted in conventional FMEA. Table 9, Table 10, Table 11 and Table 12 are the FMEA results and average FMEA results of the three decision-makers.

4.5. Drawbacks of Grey Relational Analysis to FMEA

In Grey relational analysis method risk factors have equal weights which may lead to limitation for the findings.

4.6. Multi-Criteria Decision-Making (MCDM) Software for MLP Evaluation

The developed MCDM tool is a standalone Windows command shell application designed using Python programming language version 3. The application contains six separate modules. Four of these modules are the codes for FAHP, FMEA, grey calculation, and ranking algorithm. The remaining two are for the input data files of the FMEA and FAHP, respectively. The first page of the application displays the introduction page. This page explains the function of the application to the user. After the introduction page, the user is presented with several options, as shown in Figure 9. Since the data input for FAHP and FMEA are based on expert decisions, the data are entered into the system by writing them into a JSON file and placing them in a folder inside the software directory, as shown in Figure 10 and Figure 11. FMEA and FAHP have two separate files for input. These options enable the user to view FAHP data, FMEA data, grey RPN, and ranking, as shown in Figure 9. The information required is displayed depending on what the user wishes to view. The input of the FAHP is in the file FAHP_data.json. This file enables the input of the FAHP process to be altered according to the user preference, as shown in Figure 10. Figure 12 displays the calculated grey RPN for each MLP, and Figure 13 shows the ranking of these MLPs based on the calculated grey values.

5. Sensitivity Analysis

Sensitivity analysis was conducted to compare the new rankings with the initial rankings to assess how sensitive the final results are to changes in the weights. If significant shifts occur in the rankings, the model is sensitive to those changes, suggesting that the results are influenced by the specific weight settings. In the sensitivity analysis, we focused on three basic scenarios as follows:
High Weight on Severity: We’ll further increase the weight on the severity and equally decrease the occurrence and detection weights.
Balanced Weights: We’ll assume all three factors (severity, occurrence, detection) have equal weights.
Low Weight on Severity: We will lower the weight of severity and increase the weights for occurrence and detection.
The corresponding weights are given in Table 19.
As demonstrated in Figure 14, EdX and Khan Academy consistently demonstrate low sensitivity to changes in severity, occurrence, and detection, making them the most reliable platforms based on the analysis. Pluralsight and SoloLearn are more sensitive to increases in severity, indicating they are more prone to significant issues when critical failures occur. Changes in the detection weight seem to have the least impact across all platforms, suggesting that detection is not as influential in determining the overall risk as severity or occurrence. This sensitivity analysis helps to confirm the robustness of the rankings and shows that certain platforms (like EdX and Khan Academy) maintain stability, while others (like SoloLearn and Pluralsight) are more sensitive to specific risk factors.

6. Comparative Analysis

The study compares three baseline methods, namely, traditional FMEA, Grey RPN with equal weights, and the proposed methodology. In traditional FMEA, risk priority numbers (RPNs) are calculated by multiplying the risk factors (severity, occurrence, and detection), which can lead to biased results as it does not account for the relative significance of each risk factor. In FMEA-GRA, risk priority numbers (RPNs) are calculated by assigning equal importance to the risk factors (severity, occurrence, and detection). In the proposed methodology, different weights were assigned to risk factors determined by the experts, and then expert judgments of alternatives were ranked by using FAHP. This allows for a more flexible and accurate representation of risks, especially when subjective judgments are involved. Table 20 shows the average of the results for three experts. Table 21 shows the overall ranking based on the traditional FMEA method.
For calculating GRA-FMEA with equal weights, these steps are as follows: To reduce the potential risk, the values of all risk factors should be as small as possible. Thus, the standard series is defined as X i k = 1 , 1 , 1 . For each decision maker, to reveal the degree of fuzzy relation, the difference between values of risk factors and standard series is determined and expressed as the matrix shown below:
The relational coefficient is calculated with an equation which is X 0 k   d e n o t e s   s e r i e s ,   X i k   s h o w s   c o m p a r a t i v e   s e r i e s   w h e r e
i = 1,2 , 3 , n u m b e r   o f   f a i l u r e   m o d e s ,   k = 1,2 , 3 , n u m b e r   o f   r i s k   f a c t o r s , Δ m i n   m i n i m u m   o f   a l l   Δ i k ,   Δ m a x   m a x i m u m   o f   a l l   Δ i ( k ) , ζ 0 , 1   i s   r e l a t i v e   v a l u e   o f   r i s k   c o e f f i e n t ,   w a s   c o n s i d e r e d   a s   0.5 .
The corresponding values are calculated as follows:
γ X 0 k , X i k   = Δ m i n + ζ Δ m a x Δ 0 i k + ζ Δ m a x
If all three risk factors have equal weights, the following equation is applied to determine the RPN value.
τ 01 k = 1 3 1 3 γ 0 i ( k )
An example of calculated difference FMEA results for expert1 is shown in Table 22. Table 23 shows the GRA-FMEA results for expert 1. The same calculations were carried out with expert2 and expert2 evaluations. Table 24 shows the GRA-FMEA average results for three experts. Table 25 indicates the final ranking according to GRA-FMEA.
The visualization for the comparative analysis of the three methods is given in Figure 15, Figure 16, Figure 17 and Figure 18.
FMEA provides a baseline evaluation of the platforms based on fault tolerance, time behavior, and other criteria, but it treats all risk factors equally without accounting for subjective differences in the importance of each factor. This leads to outliers like Pluralsight’s high fault tolerance score. GRA-FMEA refines the evaluation by considering the relational closeness to an ideal platform, moderating extreme scores like Pluralsight’s, and offering a more balanced comparison across platforms. Platforms like Khan Academy and Duolingo improve in ranking based on user-friendly criteria such as learnability and understandability. GRA-FMEA-FAHP enhances the evaluation further by introducing fuzzy logic to handle uncertainty in expert judgments. This results in a more realistic and balanced ranking of the platforms, with EdX and SoloLearn emerging as more well-rounded platforms across all criteria and Khan Academy standing out in learnability and user accessibility. The combination of FMEA, GRA, and FAHP shows the evolution of the evaluation process. FMEA provides the base risk analysis. GRA helps in making better relational comparisons. FAHP adds flexibility and refinement to adjust for expert input, making it the most robust and balanced evaluation method among the three. GRA-FMEA-FAHP is the most reliable method because it balances the quantitative risk analysis of FMEA with relational comparisons from GRA and adjusts for subjective importance and uncertainty through FAHP. This produces more robust, well-rounded results that reflect both objective and subjective factors relevant to the decision-making process. Traditional FMEA tends to produce extreme values—either high (SoloLearn) or low (Khan Academy, EdX), which may miss some nuances. The GRA-RPN method provides a relational comparison to an ideal solution, which results in more moderate scores. It highlights the relative performance differences more than absolute risk values. The FMEA-GRA-FAHP method shows the most balanced results, reflecting how risk perception changes when expert opinions and uncertainty are considered. It is a more nuanced and realistic method for comparing the platforms.

7. Discussion

The study aims to propose an efficient way to rank and choose the most suitable mobile learning platform by integrating risk analysis and multi-criteria decision-making methods to remedy the difficulty of selecting an adequate mobile learning platform among many that are available.
The proposed method uses unique, integrated risk analysis and multi-criteria decision-making approaches to overcome the shortcomings of the traditional FMEA approach. An integrated FMEA-GRA-FAHP method which is a combination of risk analysis and multi-criteria decision-making method for the suitability of mobile learning platforms. The initial implementation of grey relational analysis involved assuming that risk factors carried equal weights. In the subsequent application, distinct weights were assigned to risk factors through the utilization of Fuzzy AHP, as determined by evaluations from experts. This represents a significant advancement in the realm of failure mode and effect analysis (FMEA), as the introduction of varied weights through Fuzzy AHP introduces a novel approach to the FMEA methodology. This study aims to employ a multi-criteria method such as FMEA-GRA-FAHP to evaluate and rank the five selected MLPs. Furthermore, the study adopts the technical usability requirements concerning the ISO/IEC 9126 model to generate a suitable framework for the effective evaluation of the selected MLPs using such techniques. It was found that EdX, with a grey RPN of 19.0023, is the most reliable based on the quality evaluation process carried out in this research. Khan Academy is the 2nd in place with a grey RPN of 19.4994. For example, studies employing VIKOR, Analytical Network Process (ANP) have also found that platforms like EdX and Khan Academy consistently rank highly in terms of user satisfaction, hence usability and reliability, which is in alignment with the study findings [51,52,53]. Duolingo was third in place with an RPN value of 22.91, followed by SoloLearn with an RPN value of 25.10. Plural sight is ranked the lowest, considering the grey RPN value. According to the results, mobile learning platforms with higher grey RPN values are found to be less reliable and demonstrate highly significant risks or lower performance in certain criteria. In this case, Pluralsight, with the highest RPN, ranks the lowest, suggesting more potential risks exist as compared to the other platforms. Our study introduces the concept of risk analysis through FMEA, which has led to a few notable differences in ranking compared to other studies that emphasize user satisfaction alone. Meanwhile, Pluralsight, highly rated in terms of content quality employing the FAHP-TOPSIS approach [54], ranked lower in our analysis due to higher risks in fault tolerance and time behavior. This finding underlines the significance of integrating a risk assessment with multi-criteria decision-making as a framework for evaluating learning platforms, particularly for long-term usage in educational environments.
These findings show that the proposed methodology is robust and could be used effectively to locate a suitable mobile learning platform. The integrated FMEA-FAHP-GRA methodology provides a more nuanced analysis by prioritizing risk factors such as fault tolerance and time behavior, which previous studies using standalone methods may not have considered with the same rigor. The proposed methodology incorporates the risk assessment well with multi-criteria decision-making while compensating for the shortcomings of the conventional FMEA, where risk factors have equal weights. Another prevalent feature of this approach is the consideration of each factor using expert decision-making. GRA also has shortcomings in that risk factors have equal weights; however, in the proposed approach, different weights could be assigned by using the FAHP method, which improves the accuracy of results. Additionally, adopting a framework using ISO/IEC 9126 quality in-use international standards for software quality strengthens the evaluation framework.
The FMEA-GRA-FAHP approach has shown to be a reliable and effective strategy for assessing and classifying mobile learning platforms. The study’s findings are consistent with the idea that adding expert knowledge via Fuzzy AHP to risk analysis offers a more complex and trustworthy framework for making decisions, particularly when several platforms need to be assessed according to various criteria.
Table 13, Table 14, Table 15, Table 16 and Table 17 showed FMEA-FAHP grey RPN results obtained for each MLP. The conventional RPN is evaluated based on Equation (12). The reason was that for every FMEA assessment process carried out, the risk priority number was evaluated. The grey RPN is achieved using Equation (11).
The impact of the assigned weights is obvious after the calculation of grey RPN. It is seen that, even when the value of each risk factor is the same for two different criteria under an MLP, the grey RPN seems to be different, while the conventional RPN remains the same because the weight has a significant influence on the results. Instances are observed under EdX and Sololearn MLPs. In Table 4, for EdX MLP, under the criteria Understandability and Learnability, the RPN remain the same because of the same value of risk factors. However, the grey RPN is different. This is repeated under Sololearn MLP for learnability and operability.
The conclusive ranking of the MLPs depends on the summation of the grey RPN, which is obtained from the values of each criterion [10]. From Table 18, the MLP ranking using the FMEA-FAHP technique indicates Pluralsight MLP is ranked as the most vulnerable and has a high potential of failure, followed by SoloLearn. However, EdX was found to be the most reliable based on quality evaluation using the FMEA-FAHP technique. EdX and Khan Academy are ranked to have the least potential for failure. The final ranking of the mobile learning platforms is shown in Table 18.
The adopted framework in this study emerged from the ISO/IEC 9126 standards, and decisions made to access them using the FMEA were based on the judgment of the three experts. The FMEA was used to assess the potential risks or failures based on the adopted framework, and the FAHP was used to prioritize the weights of the risk factors of the FMEA; hence, a ranking was created for the mobile learning platforms with the integration of both FMEA and FAHP.
In this study, alternatives are ranked based on seven criteria with the use of the FMEA-FAHP technique. In previous studies where AHP was used, it was observed that AHP was unable to account for ranking inconsistencies, and hence, FAHP is preferred and proven to be more efficient than the regular AHP approach [55].
Researchers in [39] evaluated mathematical learning applications with a combination of the FAHP and TOPSIS methods. In their work, a model was proposed that considered technical and non-technical aspects to provide a framework for evaluation. The technical aspect was coined from the ISO/IEC 9126 and was based on quality evaluation, while the non-technical aspects of the framework focused on user satisfaction. In their study, the FAHP was used to assign weights to each criterion in the framework, and TOPSIS was used in ranking the mathematical applications, but in the case of this study, the FAHP is used to assign weight to the risk factors associated with FMEA and these risk factors are used to access potential risk or failures associated with each criterion in the working framework.
Other studies have employed the use of MCDM in comparing approaches to evaluation and combined them for a software evaluation, but this study combines FMEA and recognizes FMEA as an MCDM situation in prioritizing the risk factors, and hence, the combination of FMEA-FAHP is unique in terms of MLPs evaluation. This is the first study to combine both approaches and effectively rank mobile learning platforms with the use of Fuzzy AHP from the failure mode effect analysis perspective.
While this study focused on five mobile learning platforms (MLPs) in a specific educational context, the methodology—an integrated approach combining Failure Mode and Effects Analysis (FMEA), Fuzzy AHP (FAHP), and Grey relational analysis (GRA)—is designed to be scalable and adaptable to various settings. The strength of this method lies in its ability to handle multi-criteria decision-making (MCDM) problems by weighing qualitative and quantitative data. This makes it applicable to a wide range of platforms and educational environments.
The proposed method can easily handle a larger number of platforms by scaling the FMEA and FAHP processes. The ranking procedure, based on GRA –FAHP and the calculated risk priority numbers (RPNs), is not limited to a specific number of platforms. Increasing the number of platforms would primarily require more expert input for the FAHP process, but the methodology itself remains robust and can handle the added complexity.
The methodology is not only limited to the evaluation of mobile learning platforms. The evaluation framework can be easily applied to assess a variety of educational contexts using software. Risk factors, such as time behavior, fault tolerance, and operability, can be tailored to match the critical features of different applications as well. The FAHP process provides the adjustment of criteria and their relative importance depending on the specific needs of the educational context.
FMEA assessment ensures that platform-specific risks are identified and assessed based on the context. The proposed methodology provides flexible and adaptable features beyond the scope of mobile learning platforms such as e-learning systems, training platforms, or hybrid learning solutions.

8. Practical Implications

The relevant literature heavily focused on user satisfaction and content-system quality; however, educators should consider potential technical risks, and usability features are advised to since such issues may cause risks and could provide limited functionality, which can easily disrupt the learning process. Therefore, educators should consider the proposed aspects of the study in the careful selection of mobile learning platforms for their students. The findings of the study could be used as a guideline for educators who are aiming for reliable, mobile learning environments.
Developers of mobile learning platforms could benefit from the findings of the study by improving their products over problematic areas. Developers can significantly minimize the risks associated with fault tolerance and optimize time behavior to make platforms more attractive and usable. Maintaining a technically robust platform also reduces costs and platform failures.
The results of the study amplify a comprehensive and holistic risk-based approach for the integration of mobile learning tools into the curriculum, which aligns with policies supporting equitable and sustainable education for policymakers. Mobile learning platforms with high technical performance should also be available and accessible to all learners, encouraging lifelong learning.

9. Conclusions

The combination of FMEA-GRA-FAHP methods proved to be an effective technique in evaluating the quality of a mobile learning platform as well as ranking mobile learning platforms in order of their reliability.
Multi-criteria decision-making techniques play a vital role in the quality evaluation and testing of mobile learning platforms. The last COVID-19 pandemic situation has triggered rapid growth in the use of mobile learning platforms, especially in the educational sector. Evaluating such platforms by employing robust and reliable multi-criteria decision-making techniques plays a vital role in selecting proper mobile learning platforms and improving the vulnerabilities, hence providing better management of online learning.
This study utilizes the FMEA-GRA-FAHP technique in prioritizing risk factors. It is one of the most efficient weight-assigning MCDM techniques. FMEA is used to evaluate the potential risk associated with each platform based on the adopted framework which is coined from the ISO/IEC 9126 standard. The results of the FMEA are prioritized using the FAHP weights to assign risk priority numbers and, in turn, rank mobile learning platforms in their order of reliability from a quality perspective.
Other multi-criteria decision-making techniques, such as AHP, ANP, PROMETHEE, etc., can be integrated to further the selection and assessment of MLPs. Also, additional MLP samples can be added, and increasing the number of experts could be recommended for future work.
FMEA-GRA-FAHP is an efficient tool for evaluating the potential risk of a system considering the fact that the FAHP is able to compensate for the drawbacks of the traditional FMEA technique. This makes it recommendable in various industries, including the aerospace industries, electronic industries, and safety management. These are a few of the areas where the risk of failure is highly prioritized to maintain safety. Educators can leverage the use of the FMEA-GRA-FAHP approach for MLP ranking to ease the learning process and use the most efficient MLPs from a pool of available options. Institutions can also leverage the use of this technique in the teaching process to enable easier and more efficient means of learning for students and a more efficient means of teaching for lecturers.
Some potential limitations of the study could be listed as follows:
This study is delimited to enhancing FMEA methodology through the use of GRA-FAHP methods by assigning different weights to risk factors and ranking mobile learning platforms. The traditional FMEA, GRA-FMEA and the proposed methodology are also compared and contrasted in terms of robustness. However, in the future, different risk assessments and multi-criteria methods could be incorporated to perform fine-tuning of the results. In this study, the same importance was assigned to each expert judgment, as for future work, different weights could also be assigned to expert judgments. The evaluation framework emphasizes technical aspects of mobile learning platforms; however, researchers acknowledge the importance of non-technical aspects such as user engagement, accessibility, and equity for the pedagogical effectiveness of mobile learning platforms to enhance user experiences and outcomes in the educational milieu. Therefore, the crucial attributes should be considered for further evaluation. Another limitation of the study is that the number of experts is limited to three. For future work, the inclusion of more experts will yield a more representative sample of judgments. The same issue is also valid with the number of alternatives, including more mobile learning platforms for evaluation, which will improve the representability of the sample.

Author Contributions

Corresponding author, S.B., contributed to the manuscript’s theoretical development, analysis, interpretation, and writing. Conceptualization, S.B.; Data curation, S.B.; Formal analysis, O.A.I.; Investigation, S.B. and O.A.I.; Methodology, S.B.; Project administration, O.A.I.; Resources, S.B. and O.A.I.; Software use, O.A.I.; Supervision, S.B.; Validation, S.B.; Visualization, O.A.I.; Writing—original draft, O.A.I.; Writing—review and editing, S.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Every ethical standard for research was followed in this article. The University’s Scientific Research Ethics Committee granted the study’s ethical approval 01/06/2020 (NEU/FB/2020/01/06). The participation was completely voluntary, and that consent might be revoked at any moment was also made clear to the participants.

Informed Consent Statement

Informed Consent Statement: Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Behera, S.K. E-and M-Learning: A comparative study. Int. J. New Trends Educ. Their Implic. 2013, 4, 65–78. [Google Scholar]
  2. Chandran, S. Application and Impacts of Mobile Learning: A Case Study in Technology for Sustainable Distance Learning. In Proceedings of the Eighth International Network Conference (INC 2010), Heidelberg, Germany, 6–8 July 2010; pp. 237–240. [Google Scholar]
  3. Anggrainingsih, R.; Umam, M.Z.; Setiadi, H. Determining e-learning success factor in higher education based on user perspective using Fuzzy AHP. MATEC Web Conf. 2018, 154, 03011. [Google Scholar] [CrossRef]
  4. Işık, A.H.; İnce, M.; Yiğit, T. A Fuzzy AHP Approach to Select Learning Management System. Int. J. Comput. Theory Eng. 2015, 7, 499–502. [Google Scholar] [CrossRef]
  5. Miguel, J.P.; Mauricio, D.; Rodríguez, G. A Review of Software Quality Models for the Evaluation of Software Products. Int. J. Softw. Eng. Appl. 2014, 5, 31–53. [Google Scholar] [CrossRef]
  6. Pocatilu, P.; Boja, C. Quality characteristics and metrics related to m-learning process. Amfiteatru Econ. J. Acad. Econ. Stud. 2009, 11, 346–354. [Google Scholar]
  7. Abou El Hassan, D.S.; Elsherpieny, E.A.; Kholif, A.M.; Khorshid, M.A. The role of failure mode and effects analysis in improving the quality performance of dairy laboratories. J. Food Saf. 2017, 37, e12364. [Google Scholar] [CrossRef]
  8. Pomerol, J.-C.; Barba-Romero, S. Multicriterion Decision in Management; International Series in Operations Research & Management Science; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  9. Saaty, T.L. The analytic hierarchy process (AHP). J. Oper. Res. Soc. 1980, 41, 1073–1076. [Google Scholar]
  10. Farajiparvar, N.; Mayorga, R.V. Maintenance policy selection using fuzzy failure modes and effective analysis and key performance indicators. Int. J. Product. Qual. Manag. 2018, 25, 170–197. [Google Scholar] [CrossRef]
  11. Chang, D.-Y. Applications of the extent analysis method on fuzzy AHP. Eur. J. Oper. Res. 1996, 95, 649–655. [Google Scholar] [CrossRef]
  12. Iftikhar, M.A.; Siddiqui, A.S. A Study on Fuzzy AHP Method and Its Applications in a Tie Breaking Procedure. Glob. J. Pure Appl. Math. 2017, 13, 1619–1630. [Google Scholar]
  13. Reifer, D.J. Software Failure Modes and Effects Analysis. IEEE Trans. Reliab. 1979, R–28, 247–249. [Google Scholar] [CrossRef]
  14. Safari, H.; Faraji, Z.; Majidian, S. Identifying and evaluating enterprise architecture risks using FMEA and fuzzy VIKOR. J. Intell. Manuf. 2016, 27, 475–486. [Google Scholar] [CrossRef]
  15. Ilangkumaran, M.; Shanmugam, P.; Sakthivel, G.; Visagavel, K. Failure mode and effect analysis using fuzzy analytic hierarchy process. Int. J. Product. Qual. Manag. 2014, 14, 296–313. [Google Scholar] [CrossRef]
  16. Hu, A.H.; Hsu, C.W.; Kuo, T.C.; Wu, W.C. Risk evaluation of green components to hazardous substance using FMEA and FAHP. Expert Syst. Appl. 2009, 36, 7142–7147. [Google Scholar] [CrossRef]
  17. Başaran, S. Multi-Criteria Decision Analysis Approaches for Selecting and Evaluating Digital Learning Objects. Procedia Comput. Sci. 2016, 102, 251–258. [Google Scholar] [CrossRef]
  18. Yousif, E.H.; Abd Elwahab Mekki, W. Prioritizing the CSFS of Cloud-Based M-Learning Analytic Hiearchy Process and Blackboard Case Study. Int. J. Learn. Manag. Syst. 2023, 10, 65–78. [Google Scholar] [CrossRef]
  19. Naveed, Q.N.; Qahmash, A.I.; Qureshi, M.R.N.; Ahmad, N.; Abdul Rasheed, M.A.; Akhtaruzzaman, M. Analyzing Critical Success Factors for Sustainable Cloud-Based Mobile Learning (CBML) in Crisp and Fuzzy Environment. Sustainability 2023, 15, 1017. [Google Scholar] [CrossRef]
  20. Meng, X.; Li, S.; Malik, M.M.; Umer, Q. Machine-Learning-Based Suitability Prediction for Mobile Applications for Kids. Sustainability 2022, 14, 12400. [Google Scholar] [CrossRef]
  21. Singh, K.; Naicker, N.; Rajkoomar, M. Selection of Learning Apps to Promote Critical Thinking in Programming Students using Fuzzy TOPSIS. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 383–392. [Google Scholar] [CrossRef]
  22. Alonso-Martínez, D.; Jiménez-Parra, B.; González-Álvarez, N.; Godos-Díez, J.-L.; Cabeza-García, L. Taking Advantage of Students’ Passion for Apps in Sustainability and CSR Teaching. Sustainability 2019, 11, 779. [Google Scholar] [CrossRef]
  23. Parsons, D.; Hokyoung, R. A framework for assessing the quality of mobile learning. In Proceedings of the International Conference for Process Improvement, Research and Education, Amsterdam, The Netherlands, 12–14 June 2006; pp. 17–27. [Google Scholar]
  24. Basaran, S.; Oluwatobi, J.A. A Multi-Criteria Decision Making to Rank Android based Mobile Applications for Mathematics. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 99–107. [Google Scholar] [CrossRef]
  25. Ozan, O.; Yamamoto, G.T.; Demiray, U. Mobile learning technologies and educational applications. Mob. Learn. Technol. Educ. Appl. 2015, 9, 97–109. [Google Scholar]
  26. Djouab, R.; Bari, M. An ISO 9126 Based Quality Model for the e-Learning Systems. Int. J. Inf. Educ. Technol. 2016, 6, 370–375. [Google Scholar] [CrossRef]
  27. Idri, A.; Moumane, K.; Abran, A. On the Use of Software Quality Standard ISO/IEC9126 in Mobile Environments. In Proceedings of the 2013 20th Asia-Pacific Software Engineering Conference (APSEC), Bangkok, Thailand, 2–5 December 2013; Volume 26, pp. 1–8. [Google Scholar] [CrossRef]
  28. Navarro-Cota, C.X.; Molina, A.I.; Redondo, M.A.; Lacave, C. A Comprehensive Usability Measurement Tool for m-Learning Applications. IEEE Trans. Educ. 2024, 67, 209–223. [Google Scholar] [CrossRef]
  29. Paul, A.; Aaron, A.; Victor, T.; Muheise, H.; Brian, M.; Joe, M. A Framework for Evaluating the Usability of Mobile Learning Applications in Universities. J. Sci. Technol. 2022, 7, 42–59. [Google Scholar] [CrossRef]
  30. Parsons, D.; Ryu, H.; Cranshaw, M. A design requirements framework for mobile learning environments. J. Comput. 2007, 2, 1–8. [Google Scholar] [CrossRef]
  31. Zurita, G.; Nussbaum, M. Computer supported collaborative learning using wirelessly interconnected handheld computers. Comput. Educ. 2004, 42, 289–314. [Google Scholar] [CrossRef]
  32. Abdullah, M.R.T.L.; Hussin, Z.; Asra, B.; Zakaria, A.R. MLearning scaffolding model for undergraduate English language learning: Bridging formal and informal learning. Turk. Online J. Educ. Technol. 2013, 12, 217–233. [Google Scholar]
  33. Peng, H.; Su, Y.J.; Chou, C.; Tsai, C.C. Ubiquitous knowledge construction: Mobile learning re-defined and a conceptual framework. Innov. Educ. Teach. Int. 2009, 46, 171–183. [Google Scholar] [CrossRef]
  34. Huang, J.H.; Lin, Y.R.; Chuang, S.T. Elucidating user behavior of mobile learning: A perspective of the extended technology acceptance model. Electron. Libr. 2007, 25, 585–598. [Google Scholar] [CrossRef]
  35. Yau, J.; Joy, M. Proposal of a Mobile Learning Preferences Model. Int. J. Interact. Mob. Technol. 2010, 4, 49–51. [Google Scholar] [CrossRef]
  36. Hsu, Y.C.; Ching, Y.H. A review of models and frameworks for designing mobile learning experiences and environments. Can. J. Learn. Technol. 2015, 41, 1–22. [Google Scholar] [CrossRef]
  37. Fahmy, S.; Haslinda, N.; Roslina, W.; Fariha, Z. Evaluating the quality of software in e-book using the ISO 9126 model. Int. J. Control Autom. 2012, 5, 115–122. [Google Scholar]
  38. Liang, S.K.; Lien, C.T. Selecting the optimal ERP software by combining the ISO 9126 standard and fuzzy AHP approach. Contemp. Manag. Res. 2007, 3, 23. [Google Scholar] [CrossRef]
  39. Başaran, S.; Haruna, Y. Integrating FAHP and TOPSIS to evaluate mobile learning applications for mathematics. Procedia Comput. Sci. 2017, 120, 91–98. [Google Scholar] [CrossRef]
  40. Al Nawaiseh, A.J.; Helmy, Y.; Khalil, E. A new software quality model for academic information systems ‘case study e-learning system’. Int. J. Sci. Technol. Res. 2020, 9, 271–282. [Google Scholar]
  41. García Aguirre, P.A.; Pérez-Domínguez, L.; Luviano-Cruz, D.; Solano Noriega, J.J.; Martínez Gómez, E.; Callejas-Cuervo, M. PFDA-FMEA, an Integrated Method Improving FMEA Assessment in Product Design. Appl. Sci. 2021, 11, 1406. [Google Scholar] [CrossRef]
  42. Li, X. Personal Information Security Risk Assessment for E-Waste Recycling Based on Fuzzy–GRA–TOPSIS and FMEA. IEEE Trans. Eng. Manag. 2024, 71, 8240–8250. [Google Scholar] [CrossRef]
  43. Altubaishe, B.; Desai, S. Multicriteria Decision Making in Supply Chain Management Using FMEA and Hybrid AHP-PROMETHEE Algorithms. Sensors 2023, 23, 4041. [Google Scholar] [CrossRef] [PubMed]
  44. Vijayaraghavan, G. A Taxonomy of the E-Commerce Risks and Failures. Master’s Thesis, Florida Institute of Technology, Melbourne, FL, USA, 2003. [Google Scholar]
  45. Botella, P.; Burgués, X.; Carvallo, J.P.; Franch, X.; Grau, G.; Marco, J.; Quer, C. ISO/IEC 9126 in practice: What do we need to know. In Software Measurement European Forum; Data Processing Organization S.r.l., Via Valentino Mazzola: Rome, Italy, 2004. [Google Scholar]
  46. Deng, H. Multicriteria analysis with fuzzy pairwise comparison. Int. J. Approx. Reason. 1999, 21, 215–231. [Google Scholar] [CrossRef]
  47. Sofyalioğlu, Ç.; Öztürk, Ş. Application of grey relational analysis with fuzzy AHP to FMEA method. Doğuş Üniversitesi Derg. 2012, 13, 114–130. [Google Scholar] [CrossRef]
  48. Chua, B.B.; Dyson, L.E. Applying the ISO 9126 model to the evaluation of an e-learning system. In Proceedings of the ASCILITE, Perth, Australia, 5–8 December 2004; Volume 5. [Google Scholar]
  49. Akman, G.; Alkan, A. Tedarik Zinciri Yönetiminde Bulanık AHP yöntemi kullanılarak tedarikçilerin performansının ölçülmesi: Otomotiv Yan Sanayiinde bir uygulama. İstanbul Ticaret Üniversitesi Fen Bilim. Derg. 2006, 5, 23–46. [Google Scholar]
  50. Anagnostopoulos, K.P.; Gratziou, M.; Vavatsikos, A.P. Using the fuzzy analytic hierarchy process for selecting wastewater facilities at prefecture level. Eur. Water 2007, 19, 15–24. [Google Scholar]
  51. Dina, N.Z.; Juniarta, N. Analysis of the Educational Application Users Satisfaction Levels by Using the Integrated Entropy Weight VIKOR Method. DESIDOC J. Libr. Inf. Technol. 2023, 43, 69–77. [Google Scholar] [CrossRef]
  52. Dina, N.Z.; Yunardi, R.T.; Firdaus, A.A.; Juniarta, N. Measuring User Satisfaction of Educational Service Applications using Text Mining and Multicriteria Decision-Making Approach. Int. J. Emerg. Technol. Learn. (iJET) 2021, 16, 76. [Google Scholar] [CrossRef]
  53. Raji, M.; Zualkernan, I. A Decision Tool for Selecting a Sustainable Learning Technology Intervention. Educ. Technol. Soc. 2016, 19, 306–320. [Google Scholar]
  54. Alojaiman, B. Toward Selection of Trustworthy and Efficient E-Learning Platform. IEEE Access 2021, 9, 133889–133901. [Google Scholar] [CrossRef]
  55. Hsu, C.W.; Hu, A.H.; Wu, W.C. Using FMEA and FAHP to risk evaluation of green components. In Proceedings of the 2008 IEEE International Symposium on Electronics and the Environment, San Francisco, CA, USA, 19–22 May 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–6. [Google Scholar]
Figure 1. Flowchart of the study.
Figure 1. Flowchart of the study.
Applsci 14 08844 g001
Figure 2. Duolingo mobile learning platform.
Figure 2. Duolingo mobile learning platform.
Applsci 14 08844 g002
Figure 3. EdX mobile learning platform.
Figure 3. EdX mobile learning platform.
Applsci 14 08844 g003
Figure 4. Pluralsight Platform.
Figure 4. Pluralsight Platform.
Applsci 14 08844 g004
Figure 5. Khan Academy Platform.
Figure 5. Khan Academy Platform.
Applsci 14 08844 g005
Figure 6. Sololearn Platform.
Figure 6. Sololearn Platform.
Applsci 14 08844 g006
Figure 7. Point of intersection between M1 and M2 [44].
Figure 7. Point of intersection between M1 and M2 [44].
Applsci 14 08844 g007
Figure 8. Quality evaluation model based on ISO/IEC 9126.
Figure 8. Quality evaluation model based on ISO/IEC 9126.
Applsci 14 08844 g008
Figure 9. MCDM Tool Description Page.
Figure 9. MCDM Tool Description Page.
Applsci 14 08844 g009
Figure 10. FAHP input file.
Figure 10. FAHP input file.
Applsci 14 08844 g010
Figure 11. FMEA input file.
Figure 11. FMEA input file.
Applsci 14 08844 g011
Figure 12. Grey scale calculation.
Figure 12. Grey scale calculation.
Applsci 14 08844 g012
Figure 13. Mobile Learning Platform Final Ranking.
Figure 13. Mobile Learning Platform Final Ranking.
Applsci 14 08844 g013
Figure 14. Sensitivity Analysis Results.
Figure 14. Sensitivity Analysis Results.
Applsci 14 08844 g014
Figure 15. FMEA.
Figure 15. FMEA.
Applsci 14 08844 g015
Figure 16. GRA-FMEA.
Figure 16. GRA-FMEA.
Applsci 14 08844 g016
Figure 17. GRA-FMEA-FAHP.
Figure 17. GRA-FMEA-FAHP.
Applsci 14 08844 g017
Figure 18. Comparison of Different Methods.
Figure 18. Comparison of Different Methods.
Applsci 14 08844 g018
Table 1. The Criteria for FMEA data collection [45].
Table 1. The Criteria for FMEA data collection [45].
LabelCharacteristicsSub-Characteristics
Technical Essentials based on ISO/IEC 9126EfficiencyTime Behavior
ReliabilityFault Tolerance
UsabilityUnderstandability
Learnability
Operability
Attractiveness
FunctionalitySuitability
Table 2. MLP Google Play ratings.
Table 2. MLP Google Play ratings.
MLPsRatings (0–5)
SoloLearn4.8
Duolingo4.7
edX4.7
Khan Academy4.5
Pluralsight4.5
Table 3. Ranking for probability of occurrence/detection/severity.
Table 3. Ranking for probability of occurrence/detection/severity.
Probability of FailureFailure ProbabilityRanking
Very High: Failure is almost Certain>1/210
1/39
High: Repeats failures1/88
1/207
Moderate: Failures are Occasional 1/806
1/4005
1/20004
Low: Rare failures1/15,0003
1/150,0002
Remote: low chance of failure <1/1,500,0001
Table 4. Description of the criteria for evaluation.
Table 4. Description of the criteria for evaluation.
Sub-CharacteristicsExplanation
Time BehaviorHow quickly does the system respond?
Fault ToleranceIs the MLP capable of handling error?
UnderstandabilityDoes the user understand how to use the system without much efforts?
LearnabilityCan the user learn to use the system easily?
OperabilityCan a user use the system without much effort?
AttractivenessHow pleasant does the interface look?
SuitabilityDoes the software perform its required tasks?
Table 5. Linguistic variables and scales.
Table 5. Linguistic variables and scales.
Linguistic VariablesScaleReciprocal ScaleTriangular Fuzzy ScaleTriangular Fuzzy Reciprocal Scale
Equally Preferred11(1,1,1)(1/1,1/1,1/1)
Equally to Moderately Preferred2½(1,2,3)(1/3,1/2,1)
Moderately Preferred31/3(2,3,4)(1/4,1/3,1/2)
Moderately to Strongly Preferred4¼(3,4,5)(1/5,1/4,1/3)
Strongly Preferred51/5(4,5,6)(1/6,1/5,1/4)
Strongly to Very Strongly Preferred61/6(5,6,7)(1/7,1/6,1/5)
Very strongly preferred71/7(6,7,8)(1/8,1/7,1/6)
Very strong to Extremely preferred81/8(7,8,9)(1/9,1/8,1/7)
Extremely preferred91/9(8,9,9)(1/9,1/9,1/8)
Strongly to Very Strongly Preferred61/6(5,6,7)(1/7,1/6,1/5)
Table 6. Fuzzy-based pairwise comparison matrix.
Table 6. Fuzzy-based pairwise comparison matrix.
C1C2C3
C1(1, 1, 1)(0.1432, 3.7143, 7)(0.1253, 2.4422, 7)
C2(0.1432, 2.4641, 7)(1, 1, 1)(0.1253, 0.1372, 0.1432)
C3(0.1432, 4.3813, 8)(7, 7.3333, 8)(1, 1, 1)
Table 7. Sum of rows and columns.
Table 7. Sum of rows and columns.
CriteriaSum of RowsSum of Columns
C1(1.2685, 7.1565, 15)(1.2864, 7.8454, 16)
C2(1.2685, 3.6013, 8.1432)(8.1432, 12.0476, 16)
C3(8.1432, 12.7146, 17)(1.2506, 3.5794, 8.1432)
10.6802, 23.4724, 40.1432
Table 13. FAHP results for SoloLearn mobile learning platform.
Table 13. FAHP results for SoloLearn mobile learning platform.
CriteriaWSWOWDGrey RPN
Fault Tolerance2.06151.03320.97324.0679
Time Behavior2.47380.68881.21654.3791
Suitability2.47381.03321.21654.7235
Understandability1.23691.37761.45984.0743
Learnability0.82461.03320.24332.1011
Operability1.23690.68880.24332.169
Attractiveness2.06151.03320.48663.5813
Overall Grey RPN 25.0962
Table 14. FAHP results for Khan Academy mobile learning platform.
Table 14. FAHP results for Khan Academy mobile learning platform.
CriteriaWSWOWDGrey RPN
Fault Tolerance0.82461.7220.97323.5198
Time Behavior1.64921.7220.72994.1011
Suitability0.82460.68880.48662
Understandability0.41230.68880.48661.5877
Learnability0.82461.03320.97322.831
Operability0.41230.68880.48661.5877
Attractiveness1.23690.68881.94643.8721
Overall Grey RPN 19.4994
Table 15. FAHP result for EdX mobile learning platform.
Table 15. FAHP result for EdX mobile learning platform.
CriteriaWSWOWDGrey RPN
Fault Tolerance1.64921.03320.97323.65
Time Behavior2.06151.03320.97324.0679
Suitability1.23690.68881.21653.1422
Understandability0.82460.34440.48661.6556
Learnability0.82460.68880.24331.7567
Operability0.82461.03321.21653.0743
Attractiveness0.82460.34440.48661.6556
Overall Grey RPN 19.0023
Table 16. FAHP result for Duolingo mobile learning platform.
Table 16. FAHP result for Duolingo mobile learning platform.
CriteriaWSWOWDGrey RPN
Fault Tolerance1.23691.37760.97323.5877
Time Behavior1.23691.03320.72993
Suitability0.82460.68880.48662
Understandability0.82460.68880.48662
Learnability1.23691.03320.48662.7567
Operability0.41231.7222.4334.5673
Attractiveness0.82463.4440.72994.9985
Overall Grey RPN 22.9102
Table 17. FAHP result for Pluralsight mobile learning platform.
Table 17. FAHP result for Pluralsight mobile learning platform.
CriteriaWSWOWDGrey RPN
Fault Tolerance4.1232.75520.97327.8514
Time Behavior1.23690.34440.97322.5545
Suitability0.82461.7220.24332.7899
Understandability1.64921.03321.45984.1422
Learnability1.23690.68880.24332.169
Operability0.41230.68880.24331.3444
Attractiveness2.06151.37760.97324.4123
Overall Grey RPN 25.2637
Table 18. MLP ranking based on FMEA-FAHP.
Table 18. MLP ranking based on FMEA-FAHP.
MLPsTotal Grey RPNRanking
EdX19.001
Khan Academy19.502
Duolingo22.913
Sololearn25.104
Plural sight25.265
Table 8. Synthetic extent value results for individual criteria.
Table 8. Synthetic extent value results for individual criteria.
CriteriaSCi
C1(0.031599, 0.3048, 1.40445)
C2(0.031599, 0.1534, 0.76245)
C3(0.20285, 0.54167, 1.59173)
Table 9. FMEA results of expert 1 (D1).
Table 9. FMEA results of expert 1 (D1).
EdXKhan AcDoulingoPluralsightSololearn
SODRSODRSODRSODRSODR
Fault tolerance342242321225220106742046496
Time behavior4254042432342242136539135
Suitability32742312612243712172570
Understandability312613264128658240455100
Learnability3216235303411242182418
Operability34672122413103012124114
Attractiveness21122374221036086419443224
Overall RPN192132158891437
Note: S = severity, O = occurrence, D = Detection, R = Individual RPN per criteria, and overall RPN is the RPN value of the MLP.
Table 10. FMEA results of expert 2 (D2).
Table 10. FMEA results of expert 2 (D2).
EdXKhan AcDoulingoPluralsightSololearn
SODRSODRSODRSODRSODR
Fault tolerance725701954545360109327062224
Time behavior8541605512532212213673484
Suitability32424242161236221452550
Understandability2112112213263253042972
Learnability1212234243141232162316
Operability22312132617107011111212
Attractiveness2148518402103604243283248
Overall RPN278158226349286
Note: S = severity, O = occurrence, D = Detection, R = Individual RPN per criteria and overall RPN is the RPN value of the MLP.
Table 11. FMEA results of expert 3(D3).
Table 11. FMEA results of expert 3(D3).
EdXKhan AcDoulingoPluralsightSololearn
SODRSODRSODRSODRSODR
Fault tolerance235303354532742109218051630
Time behavior3231838496335455142043224
Suitability32424112242181616655150
Understandability1133122412243253015210
Learnability2214233183411222142214
Operability136181122151050131343112
Attractiveness2112229362103603647233218
Overall RPN99203221315248
Note: S = severity, O = occurrence, D = Detection, R = Individual RPN per criteria, and overall RPN is the RPN value of the MLP.
Table 12. FMEA average of the results.
Table 12. FMEA average of the results.
EdXKhan AcDoulingoPluralsightSololearn
SODRSODRSODRSODRSODR
Fault tolerance434482544034448108432053460
Time behavior5346045360333273141262560
Suitability32530222822282511063590
Understandability2124122422284367234672
Learnability2214234243321832162316
Operability23530122415105012123216
Attractiveness2124328482103605448053230
Overall RPN172188219262324
Note: S = severity, O = occurrence, D = Detection, R = Individual RPN per criteria, and overall RPN is the RPN value of the MLP.
Table 19. RPN risk factor weights for sensitivity analysis.
Table 19. RPN risk factor weights for sensitivity analysis.
MLPOriginal_RPNOriginal_RankingRPN_Severity_IncreasedRPN_Occurrence_IncreasedRPN_Detection_IncreasedRPN_Severity_Low
EdX19119.518.819.218.5
Khan Academy19.522019.319.819.1
Duolingo22.91323.522.52322.3
Sololearn25.1425.824.725.524.5
Pluralsight25.26525.924.8525.724.7
Table 20. FMEA results for MLPs.
Table 20. FMEA results for MLPs.
EdXKhan AcDoulingoPluralsightSololearn
RPNRPNRPNRPNRPN
Fault tolerance48404832060
Time behavior6060271260
Suitability30881090
Understandability4487272
Learnability4241866
Operability3045026
Attractiveness448608030
Overall RPN172188219262324
Table 21. MLP ranking based on conventional FMEA.
Table 21. MLP ranking based on conventional FMEA.
MLPsOverall RPNRanking
EdX1721
Khan Academy1882
Duolingo2193
Plural sight2624
Sololearn3245
Table 22. Difference FMEA results of expert 1(D1).
Table 22. Difference FMEA results of expert 1(D1).
EdXKhan AcDoulingoPluralsightSololearn
SODSODSODSODSOD
Fault tolerance231121141956353
Time behavior314313231102428
Suitability216201011160614
Understandability201021301547344
Learnability210124230310130
Operability235011029010300
Attractiveness100126192753321
Table 23. GRA-FMEA results of expert 1(D1).
Table 23. GRA-FMEA results of expert 1(D1).
EdXKhan AcDoulingoPluralsightSololearn
SODRSODRSODRSODRSODR
Fault tolerance0.710.490.660.621.000.670.780.811.000.500.670.720.700.460.380.521.000.470.440.63
Time behavior0.600.870.430.630.560.820.460.610.710.490.660.620.501.000.500.670.750.790.400.65
Suitability0.800.950.420.720.331.000.560.631.001.001.001.000.750.331.000.690.440.810.410.56
Understandability0.331.000.560.631.000.670.780.810.331.000.560.630.880.580.390.621.000.500.420.64
Learnability0.330.331.000.561.000.750.460.740.430.331.000.590.330.331.000.560.600.331.000.64
Operability1.000.640.420.681.001.001.001.001.000.850.400.751.000.331.000.780.331.001.000.78
Attractiveness0.331.001.000.781.000.800.420.741.000.410.470.630.620.420.430.490.560.520.680.58
Table 24. GRA-FMEA Average of the results of experts.
Table 24. GRA-FMEA Average of the results of experts.
EdXKhan AcDoulingoPluralsightSololearn
SODRSODRSODRSODRSODR
Fault tolerance0.720.690.510.251.000.610.550.330.870.640.530.290.530.400.470.100.620.700.490.21
Time behavior0.630.710.470.210.630.530.630.210.790.670.630.340.441.000.470.210.620.630.510.20
Suitability0.740.910.470.320.780.850.750.500.780.780.890.540.690.331.000.230.600.710.410.18
Understandability0.561.000.740.411.000.890.930.820.780.890.780.540.790.810.440.280.900.650.490.29
Learnability0.560.331.000.191.000.710.500.360.430.560.810.190.330.331.000.110.480.331.000.16
Operability1.000.800.510.411.000.890.930.821.000.670.380.261.000.561.000.560.560.561.000.31
Attractiveness0.421.000.820.350.820.930.410.311.000.410.470.190.720.570.440.180.550.520.670.19
Overall GRA- RPN2.133.362.351.671.53
Table 25. MLP ranking based on conventional GRA-FMEA.
Table 25. MLP ranking based on conventional GRA-FMEA.
MLPsOverall RPNRanking
SoloLearn1.531
Pluralsight1.672
EdX2.133
Duolingo2.354
Khan Academy2.365
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Başaran, S.; Ighagbon, O.A. Enhanced FMEA Methodology for Evaluating Mobile Learning Platforms Using Grey Relational Analysis and Fuzzy AHP. Appl. Sci. 2024, 14, 8844. https://doi.org/10.3390/app14198844

AMA Style

Başaran S, Ighagbon OA. Enhanced FMEA Methodology for Evaluating Mobile Learning Platforms Using Grey Relational Analysis and Fuzzy AHP. Applied Sciences. 2024; 14(19):8844. https://doi.org/10.3390/app14198844

Chicago/Turabian Style

Başaran, Seren, and Odianosen Anthony Ighagbon. 2024. "Enhanced FMEA Methodology for Evaluating Mobile Learning Platforms Using Grey Relational Analysis and Fuzzy AHP" Applied Sciences 14, no. 19: 8844. https://doi.org/10.3390/app14198844

APA Style

Başaran, S., & Ighagbon, O. A. (2024). Enhanced FMEA Methodology for Evaluating Mobile Learning Platforms Using Grey Relational Analysis and Fuzzy AHP. Applied Sciences, 14(19), 8844. https://doi.org/10.3390/app14198844

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop