Next Article in Journal
Feasibility of the Fintech Industry as an Innovation Platform for Sustainable Economic Growth in Korea
Next Article in Special Issue
Options to Continue the EU ETS for Aviation in a CORSIA-World
Previous Article in Journal
Travelers’ Acceptance of Electric Carsharing Systems in Developing Countries: The Case of China
Previous Article in Special Issue
Fuel Consumption Model of the Climbing Phase of Departure Aircraft Based on Flight Data Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrating Qualitative Comparative Analysis and Support Vector Machine Methods to Reduce Passengers’ Resistance to Biometric E-Gates for Sustainable Airport Operations

by
Cheong Kim
1,2,
Francis Joseph Costello
1 and
Kun Chang Lee
1,3,4,*
1
SKK Business School, Sungkyunkwan University, Seoul 03063, Korea
2
Korea Airports Corporation (KAC), Seoul 07505, Korea
3
Samsung Advanced Institute for Health Sciences & Technology (SAIHST), Sungkyunkwan University, Seoul 03063, Korea
4
Creativity Science Research Institute (CSRI), Sungkyunkwan University, Seoul 03063, Korea
*
Author to whom correspondence should be addressed.
Sustainability 2019, 11(19), 5349; https://doi.org/10.3390/su11195349
Submission received: 16 August 2019 / Revised: 18 September 2019 / Accepted: 21 September 2019 / Published: 27 September 2019

Abstract

:
For the sake of maintaining sustainable airport operations, biometric e-gates security systems started receiving significant attention from managers of airports around the world. Therefore, how to reduce flight passengers’ perceived resistance to the biometric e-gates security system became much more important than ever. In this sense, the purpose of this study is to analyze the factors which contribute to passenger’s resistance to adopt biometric e-gate technology within the airport security setting. Our focus lies on exploring the effects that perceived risks and benefits as well as user characteristics and propagation mechanisms had on causing such resistance. With survey data from 339 airport users, a support vector machine (SVM) model was implemented to provide a tool for classifying resistance causes correctly, and csQCA (crisp set Qualitative Comparative Analysis) was implemented in order to understand the complex underlying causes. The results showed that the presence of perceived risks and the absence of perceived benefits were the main contributing factors, with propagation mechanisms also showing a significant effect on weak and strong resistance. This study is distinct in that it has attempted to explore innovation adoption through the lens of resistance and in doing so has uncovered important complex causation conditions that need to be considered before service quality can be enhanced within airports. This study’s implications should therefore help steer airport managers in the right direction towards maintaining service quality while implementing sustainable new technologies within their current airport security ecosystem.

1. Introduction

Global leading airports, such as Incheon International Airport (IIA), are looking to provide improved passenger services in order to maintain the overall airport experience. This has been cited as one of the key factors for developing sustainable airport operations and their own competitiveness [1]. Therefore, passenger service quality within airports are evaluated every year. For example, the Airport Service Quality (ASQ) of Airport Council International (ACI) help to ensure that airports provide appropriate passenger services that are sustainable in the long term. One important component of the passenger service quality measurement is the waiting time. However, airports have started to suffer from various issues recently, including prolonged waiting times to board, and this has subsequently decreased the passenger service quality. These are attributable to extreme congestion in passenger terminals. In fact, according to the International Civil Aviation Organization (ICAO), the number of global air-passengers rose by 78% to 36.96 billion in 2016, up from 20.72 billion in 2006. This is equivalent to an annual growth rate of 7.8%, leading to concerns over the sustainability of the state of current airport operations in the perspective of passenger service quality. Thus, recent attempts to mitigate these problems are an ongoing research problem in sustainable airport operation [2,3]. In order to sustain this continuous growth in air passengers, many international airports globally are expanding their facilities, or even building new airports to increase their operation capacity. Other airports have looked to introduce self-service systems, as well as automated devices, such as Self Check-in, Self-Bag-Drop, or Biometric Identification [4]. According to the International Air Transport Association (IATA), the implementation of self-service during the boarding process will increase to 80% by 2020 [5].
Although automated devices are increasing in power and number, it is not enough. Rising congestion and queuing at airports is a much more complex problem to deal with. Up till now much research has gone into optimization methods in helping to reduce the burden of numbers at airports [3]. One such area that needs optimization, but raises great concern, is the ways in which to sustain the levels of security. This has become something of a top priority for most airports as well as governments. While optimizing security is necessary, there are a few fundamentals of a security network for sustainable airport operations that need to be considered: (1) it needs to be less costly, (2) include better operating procedures, (3) be able to reduce passenger’s waiting time, and (4) improve the overall experience for the passengers [3].
In the case of Korea (where this study was based) biometric technology has become a considered sustainable alternative to the common human agents for identification of passengers at pre-security areas in airports. The aim is to reduce congestion by installing biometric e-gates in parallel at these pre-security areas. The need for this technology has been identified due to a 2017 Air Traffic Service Report (Ministry of Land, Infrastructure and Transport, 2017), where they identified that the number of air passengers in Korea had grown rapidly, reaching 109.36 million in 2017, surpassing the 100 million mark for two years in a row. Alongside this, the Korean Airports Corporation (KAC) found in an online study that a safe airport, including security measures, was ranked number one amongst internal (19.6%) and external (17.4%) stakeholders, based on sentiment towards social value prioritization [6]. Due to these findings, in January 2018 KAC identified the domestic lines of Gimpo Airport and Jeju Airport as routes for potentially implementing biometric technology. According to a survey of 224 airport users conducted by KAC in October 2017, 78% of the respondents expected that the biometric boarding system would reduce waiting time and only 8% of them said that the biometric boarding system would provide no value to the passengers.
With this indication from the public, the biometric e-gate boarding service was introduced by KAC for the Gimpo and Jeju route to enhance passenger service quality. However, contrary to the expectations of high numbers of passengers choosing to use such a service, data released by KAC showed only a total of 12,475 passengers (daily average of 870 passengers) had used the biometric e-gates on this domestic route during January 29~February 13. If we consider that an average of 67,633 passengers used this domestic line on a daily basis in 2017, it means that only 1.28% of the passengers used the biometric e-gates. With this realization it became apparent that biometric e-gate security might not be the sustainable system for the future that was much anticipated.
Our research aim, therefore, lies in exploring the reasons why this technology was not yet adopted enthusiastically by air passengers despite the airports’ best efforts in providing this beneficial service that will help to sustain low queue times at security areas within airports.
To fulfill this research aim, we adopt two methods: csQCA (crisp set Qualitative Comparative Analysis) and Support Vector Machine (SVM) classification. By implementing csQCA we wanted to explore the underlying complex conditions which support the resistance of biometric e-gate technology. csQCA analysis greatly enables the identification of cases that can be further analyzed for causal process observations. With csQCA not all conditions have to imply a significant meaning, but rather that a researcher can select a few cases based on relevant configurations and prior theoretical knowledge [7]. Through csQCA we were able to find relevant conditions that help to support our prior hypothesis and, therefore, can help to provide unique implications, which can help to guide airport managers in implementing similar biometric e-gate technologies.
The SVM methodology, which belongs to intellectual machine-learning algorithms, has been actively used within the field of sustainability research [8,9,10,11,12]. Among the machine-learning algorithms, such as linear discriminate analysis, decision trees, logistic regression, naïve Bayes, artificial neural networks and k-nearest neighbor, SVM is a tried and tested algorithm that has gained much trust amongst academics [13,14]. Compared to other machine-learning algorithms, SVM stands out to show greater performance [15,16,17,18,19,20], by specifically changing kernel function techniques, such as the polynomial kernel, radial basis function (RBF) kernel and Pearson VII universal function (PUF) kernel. Despite this, only a few researches have shed light on these kernel functions used alongside SVM [13]. On top of kernel functions, ensemble classifiers have shown generally better results compared with single classifiers [21]. Hence, it would be expected that SVM classifier ensembles, such as SVM with bagging (BA), random subspace (RSS) and adaptive boosting (ADA), might reveal greater performance than a single SVM classifier in prediction. Therefore, this paper looks to identify the reasons why passengers resist to adopt the biometric e-gate security by exploring the performance of SVM models in combinations of various types of kernel functions and ensemble classifiers as well.
The rest of this research is arranged as follows. In the second section we explore the reasons why people may resist the adoption of technology, specifically biometric e-gate technology as the technology of choice within this paper. Next we will present the first part of our study, the QCA methodology. In particular, an analysis of the csQCA will be made. Additionally, an analysis of SVMs, kernel functions and ensemble techniques will be made before we present the dataset used and assessment metrics applied within both methodologies. Results will then be presented from our empirical testing with both research methods, before a full discussion, conclusions and implications of this research will be made.

2. Theoretical Background

2.1. Biometrics for Passenger Service Quality

After the 9.11 crisis in 2001, safety and security at airports has become extremely serious and is one of the key tools identified for airports operations [6]. In response to it, most airports have enhanced the level of security, for instance, the US government legislated an extra decree, the Aviation and Transportation Security Act (ATSA), that commanded several momentous changes in aviation security policy [22]. Although this new regulation of screening processes at airports has greatly increased aviation security on the one hand, on the other hand, it also adds to the waiting time of passengers. This endeavor has produced a negative effect on passenger service quality [22]. According to an air passenger survey of 17 countries by SITA in 2015, 76% of passengers claimed that aviation security screening procedures at airports brought forward the most negative reactions in passenger services at the airports [23].
Key stakeholders of the aviation industry, such as airports, airlines and governments, are all aware of this negative response from passengers to the new security policy at airports. Security should not be compromised, however, as it is a very significant matter for the sustainability of the industry. Hence, it is not easy to make compromises for enhancing passenger service quality. Biometric e-gates could be implemented to automate airport security, providing a solution to the paradoxical desire for both improved security and a more productive procedure for passenger service quality to derive sustainable airport operations [23]. Implementing a single-token method can replace the current system and passengers could finish the entire boarding process through the use of their biometric features.
With biometric e-gates, identities of passengers would be automatically identified and examined using their distinct physical biometric features [24]. Because this biometric information is immediately received from the passengers, airports could confirm that the passenger is really the same person who is attempting to get approval to enter airside. This would help to reduce identity deception and improve security [23,24]. In addition, biometrics at airports could support upgraded passenger services with quicker processing speeds and greater accuracy compared to human agents [25]. Thus, biometric e-gates has become an appreciable substitution of human agents for passenger identification at security entrances, with global airports now aggressively adopting or is in the process of adopting biometric services in parallel at security areas to improve passenger service quality [4].

2.2. Underlying Theory: Resistance to Innovation

In an attempt to answer the question of why resistance to biometric e-gate technology was present, we next explore the theory which guided our exploration into this phenomena (refer to Table 3 for a full breakdown of each construct).
As seen in KAC’s implementation of biometric e-gate technology, high expectations were not matched. A lack of active use by passengers in the airport showed a resistance to the use of this innovative technology [1]. This indicated that it is not easy to adopt to an innovation despite numerous benefits, such as the elimination of bringing an ID card and reduced waiting time [26]. Moreover, it becomes very difficult to adopt an innovation if the actual user does not accept, or sometimes denies the “benefits” suggested by the providers of a new system. If the users also think that there are serious risk factors in adopting it, there can be a possibility that they even show “resistance” against innovation. In some cases, the majority of the users who do not adopt an innovation might be making a more rational judgment than the few users whose objective is just to adopt innovation [27].
Currently there exists various types of perceived risk which can influence a consumer. This can then lead to resistance from a consumer in adopting a new innovation. Currently, research on resistance has identified the following factors as contributing factors: temporal risk, social risk, physical risk and functional risk [28,29,30,31,32,33,34]. Firstly, temporal risk is the concern about an unnecessary waste of one’s time that might occur from adopting the new innovation [33]. Social risk is the potential concern that adoption of a new innovative product or service might create a negative impression to others [34]. Physical risks are ones that reflect upon the threat of the new innovative product or service potentially causing bodily harm, inflicting injuries [33]. Functional risks are concerns about whether the functions of the new innovative product or service will actually work properly [33].
Notably, research up till now has focused on the “Diffusion of Innovation” theory originally proposed by Rogers [26]. This theory has been used in various fields when conducting research about innovation. Examples have been seen whereby its use was utilized as a theoretical background when examining the factors that influence the adoption of innovation in education [35] or policies [36]. In the “Diffusion of Innovation” model, it proposes that perceived benefits, such as relative advantage, compatibility and trialability have an important influence on the adoption of innovation [26,35]. In other words, relative advantage is about how much benefit the innovative product or service can provide to an end user compared to the existing products or services on offer in the current marketplace [26]. Compatibility refers to whether the innovation fits the values of the society in which the consumer belongs to. Furthermore, it asks whether it fits the personal experience of the consumer well [26]. Trialability is the concept that says consumers are getting more used to the innovation, or easing anxiety, through pilot use of the product or service [26].
Next, Rogers [37] defined innovation tendency as “the degree of how fast a certain consumer adopts an innovation.” Also, Ram [38] presented the consumers’ tendency to seek “new things” as one of the main driving factors that can have an effect upon innovation resistance. Additionally, openness to innovation, which refers to the consumers’ innovation tendency [39], and an independent decision-making tendency [40] can be seen as a user characteristic that has a potential effect on resistance to innovation.
Moreover, previous studies have highlighted that publicity gained through mass media is a very important way of spreading the acceptance of an innovation throughout a society [26]. The various media platforms include online and offline channels which are effective in easing any resistance by consumers through creating a type of intimacy with the innovative product or service [41]. With mass media, negative images of a product or service can be mitigated through the provision of information using mass media outlets [42,43]. In coherence with this notion, this study predicts that the provision of information through online and offline promotions can help to reduce the level of resistance towards a new innovation [32].
However, up till now, previous studies based on the adoption of innovation have not addressed the “resistance” in a variety of forms. In other words, it is highly necessary to conduct research on this factor in order to identify which features influence the adoption of innovation by classifying the types of resistance in the identified forms [32]. As this innovative biometric system is in its early stage of adoption [44], some airports have done more research on the regulations, functions and implementations of biometric e-gates. This paper posits that rather only focusing on these areas, researchers and managers should also concentrate on how to effectively introduce innovations such as biometric e-gate technology to passengers. For example, passengers could simply postpone accepting the biometric e-gate service until it is standardized. On the other hand, passengers could abortively disagree with adopting the service, or even have an effect on other passengers not accepting it. Since there are various types of passengers at airports, it is required to consider different levels of resistance to innovation, such as (1) Weak Resistance and (2) Strong Resistance, in order to attempt to improve the passenger service quality of airports [45].
This research therefore proposes the use of a prediction system that would detect factors which would decrease each different level of resistance of innovation as seen in the biometric e-gate technology to discover the gateway to the solution of passenger service quality enhancement for sustainable airport operations. We have used an SVM in this paper to provide a straightforward and practical solution to identify and help to mitigate resistance of passengers to biometric e-gate services at airports. If this can be successfully achieved, identification of factors that will help to sustain airport operations could be acquired promoting new innovation for passenger service quality improvement [6].

2.3. Crisp Set Qualitative Comparative Analysis (csQCA)

Before any manager can implement a successful biometric e-gate security system, they must understand the complex thought processes of air passengers. In our search for a suitable method of analysis, especially one that can capture the underlying complex notions at play, the literature led us to the use of csQCA [46]. csQCA is one part of the overarching QCA social research approach and is best summarized as the original modelling system based on two truth values: True or False [47]. This research approach is based upon set theory and Boolean algebra [32] and allows for “conjunctural causation” across observed cases [48]. Put another way, when implementing QCA research, different constellations of the same factors (X), e.g., theoretical constructs, can lead to the same result on the dependent variable (Y). Specifically, QCA allows for the development of causality while leaving room for complexity in a model, also known as “multiple conjunctural causation” [48]. Thus, different causal “paths” may lead to the same outcome as each path can be just as relevant in its own distinct way compared with other found “paths” [47,48]. Therefore, with the implementation of QCA, a researcher can look for new ideas, propositions or conjectures formulated within the researchers’ theory, rather than testing theory already entrenched within prior literature [7].
In this paper, we implement the use of csQCA for three main reasons: (1) The original data was based upon a Likert scale response survey which had a scale from 1–5. Duşa [47] argues binary crisp sets (csQCA) are the most general type of calibrated data in which any kind of raw data can be transformed. Furthermore, due to the bipolar nature of a Likert response scale, the use of fuzzy sets has been questioned due to logical flaws. In an example, this paper’s dependent variable resistance is constructed on a Likert response scale from a positive end (1 = low resistance) to a negative end (5 = strong resistance). Transforming this bipolar scale into a unidimensional set introduces conceptual flaws. This is because there is no logical reason for which the end “low resistance” can be calibrated as a numerical zero point in the set of “resistance”. An air passenger who has very low resistance is certainly outside the set of a passenger with high resistance, but the same can be argued about the mid-point (3 = neutral to resistance). In csQCA set theory, low resistance and high resistance can be treated as two separate sets, rather than two ends of the same set, like seen in a fuzzy set fsQCA. Thus, implementing csQCA reflects the true nature of responses given in our survey. (2) The number of samples in our dataset was considered within the small- to intermediate-N size. Previous research has shown that QCA performs most effectively when data samples are within this range and explanatory conditions stay between two and ten [7,46]. Lastly, (3), within this paper we attempted to study the adoption of innovation through “resistance” in two forms—postponement and rejection [32]. As this approach to studying adoption to innovation flips the conceptual lens, we argue that QCA is best placed to capture new ideas or conjectures [7] that previous research has failed to capture while studying adoption through more traditional research methods [26].
The analytical process of QCA is split into four main steps. Once all of these steps are complete, complex solutions for the problem are presented in the form of the model(s) sufficiency of fit [46,47]. First, a decision must be made on what cases to include within a model [49]. Second, a data table must be converted into a “truth table” that gives all the possible combinations of the conditions found. A truth table with K causal conditions leads to 2K rows of possible causal combinations [46]. Rows without cases that are empirically significant are logical reminders and those cases where the outcome is both present and absent are considered contradictory configurations. Boolean algebra is then applied in order to present the most parsimonious solution. The parsimonious solution is the smallest common denominator of a causal condition that leads to the same outcome. Lastly, the results can be explored for all the factors that are necessary and sufficient for the outcome: resistance [46,47,49].
The QCA package within the R software environment [47] was used to perform all the tests. First, we had to organize the constructs for testing into two separate files. Our reasoning for this was due to fact that our dependent variable, resistance, was split into two separate constructs and thus gave us “Weak Resistance” (WR) and “Strong Resistance” (SR). These dependent variables were then tested through four identified independent variables, “Perceived Risk” (PR), “Perceived Benefits” (PB), “User Characteristics” (UC), and “Propagation Mechanism” (PM).
The first challenge is how to calibrate data. The data collected within this study is based on a 5-point Likert scale survey. This means the calibration of categorical data is needed, for this reason we choose to use a crisp set calibration (0, 1) based on recommendations by [47]. Next, another serious consideration is with the skewed nature of Likert scale responses. As no prior scale based on our constructs came to our knowledge, we failed to find a theoretical solution to calibration. In the absence of ground theoretical solutions for calibration, statistical clustering is a solution. This mathematically looks to find naturally occurring clusters within a given dataset and can be used as a threshold for labelling the data into crisp sets [47].
Next, necessity and sufficiency tests were made. However, first a theoretical definition is needed. Necessity looks to assess whether cases in each outcome have the same causal conditions. In other words, the condition X is necessary when the outcome Y is present alongside the presence of condition X. Given this, Y cannot be achieved without the condition X [49,50]. With sufficiency, the opposite logic applies. A condition is considered sufficient if, whenever it is present across each case, the outcome is also present. With sufficiency there should not be any single case whereby it shows the condition but not the outcome. Put differently, if X is sufficient for Y, then X cannot exist without the presence of Y [46].
Within QCA there are two main parameters assessing the model’s fit: consistency and coverage. Necessity coverage measures the extent to which the elements of an outcome set are also elements of the condition set. Sufficiency consistency measures how much a case shares a given condition or combination of conditions based on a subset of the outcome in question. Sufficiency coverage measures the degree to which a subset relation exists between the outcome and conditions. Likewise, necessity consistency measures the amount that a subset relation exists between the outcome of a found condition (see Ragin [46] and Schneider and Wagemann [49] for a more detailed explanation).

2.4. Support Vector Machine (SVM) and Kernels

SVM was introduced by Vapnik (1998) and has performed greatly in pattern recognition problems [51]. In addition, SVM provides enhanced classification compared to other machine-learning methods [13]. SVM builds a hyperplane for classification that distinguishes between positive (+1) and negative (−1) sets with the greatest margin. [51]. For example, (x1, y1), (x2, y2) … (xm, ym), where xi Rd signifies vectors in a d-dimensional hyperplane and yi [ 1 ,   + 1 ] is a label of the class. Next, SVM indications are generated by transforming the input vectors into a new higher dimensional analogue for the plane in dimensions marked as Φ :Rd → Hf where d < f [13]. The optimal hyperplane is then formed using a kernel K(xi,xj), which is the product of input vectors xi and xj and where K(xi,xj) = Φ (xi)∙ Φ (xj) [13,52].
In the modified space, the training instance closest to the hyperplane is called the support vector. The number of such support vectors is typically small relative to the extent of the training set and decides the blanks and crystalline surface of the hyperplane. Among SVM kernels, the Gaussian radial basis function (RBF), polynomial and Pearson VII universal function (PUF) kernels show immense performance, which are Kpoly(xi, xj) = (xi∙xj + 1)p (p is the degree of polynomial), KGaussian(xi, xj) = e x i x j 2 2 σ ( σ is Gaussian sigma), and KPearson(xi, xj) = 1 / [ 1 + ( 2 | x i x j | 2 2 ( 1 / ω ) 1 / σ ) 2 ω ] (σ and ω are the half-width and the tailing factor of the peak). Selection of the adequate kernel for SVM highly depends on the dataset and domain of researches. Nonetheless, polynomial and radial basis function (RBF) kernels are frequently employed for miscellaneous purposes [53,54,55,56,57], while the Pearson VII universal function (PUF) kernel is vigorous and has stronger mapping power than the other kernels [58].

2.5. Synthesis of Recent SVM Literature

Recent studies on SVM have several limitations. First, most studies have used the RBF kernel function exclusively, despite the possibility of other kernel functions. Second, most related works have focused on accuracy as the main performance metric. However, it is considerable that datasets of prediction are in multiple classes of an imbalanced nature. Therefore, it is inappropriate to rely on just accuracy for deciding the performance of a model. Table 1 below shows recent studies on predictions using SVM in terms of the kernels and evaluation methods appointed.
Performance measurements on prediction models using SVM should also include the F-Score and AUC. When used alongside accuracy, especially for complex prediction models including multiclass and imbalanced classification tasks. Accuracy is used to show a model’s true performance in a classification task. When this is considered, Accuracy can be easily understood and can quickly represent the performance of each individual outcome. Accuracy can be measured by calculating the confusion matrix in the following way:
T P   +   T N T P   +   F P   +   T N   +   F N .
Solely relying on accuracy is not enough as overfitting can be an issue. For this reason, the F-score metric can be used with accuracy. The F-score gives the results of the harmonic mean of precision and recall. Precision is the result of all the accurate positives and recall being the number of accurate positive results divided by the relevant samples. Its best value is seen at one and can be measured in the following way:
2 × P r e c i s i o n   ×   R e c a l l P r e c i s i o n   +   R e c a l l .
In addition, Area Under the Curve (AUC) can be used with accuracy and the F-score in order to improve confidence of the evaluation model. The AUC uses two distinct metrics and plots them in order to find out how well a model can classify true results. The curve plots two parameters; recall or true positive rate and the false positive rate. These could be represented in the following way:
T P R =   T P T P + F N ,       F P R = F P F P + T N .
The AUC curve displays TPR against FPR at different classification thresholds and aggregates a performance measurement for all possible classification thresholds. Another way to view this metric is through the model’s interpretation of the probability that the model ranks a random positive instance more highly than a random negative instance.
In this study, accuracy, the F-score and AUC will be used as evaluation metrics to extract the best performance on predicting “resistance” against innovation based on biometric e-gates.

2.6. Ensemble Classifiers

Ensemble classifiers are coalescence of various classifiers and can provide enhanced performance when implementing them against single classifiers [21]. Ensemble classifiers are based on the modular processing of the brain [13]. In addition, each function can be segregated into subdivisions without mutual intervention [61]. Ensemble classifier use the idea of the divide-and-conquer algorithm, which is based on the multi boughed recursive function in computer science. This algorithm recursively breaks down intractable problems into more recognizable problems to use more adequate techniques for each sub-problem. There are three common techniques of ensemble classifiers; Bagging (BA), Random Subspace (RSS) and Adaptive Boosting (ADA) [62].
Bagging, also known as bootstrap aggregating, produces n new datasets from the original dataset for training to build n individual models by random sampling with replacement. Then, n models are aggregated to generate an output average by using a blending approach, for example, majority voting [63] for improving the stability and accuracy of the model.
Random Subspace (RSS), also called attribute bagging [64], reinforces performance of classifications by reducing the correlation between prediction values of datasets with randomly chosen samples [65]. RS constructs a classifier using a decision tree that preserves veracity on training sets. This classifier has several trees that are built by quasi-random selection for training with replacement to derive posterior probabilities.
The curse of dimensionality occasionally occurs to hyperplane machine-learning cases [66]. AdaBoost (ADA), also known as adaptive boosting, only chooses features with high predictive probability, reducing the dimensionality as well as improving the running time. At first, each classifier has the same weight, then, the weight of each trained subset will be changed according to the significance [13]. After replaying this process, the model will make the final result according to the weight of the classifiers [67].

3. Methodology

3.1. csQCA Procedure

First, the calibration process transformed all the data into either a 0 = the outcome is not present, or 1 = the outcome is present. From this newly calibrated data a truth table was created finding all the relevant conditions for Y (0;1) (as the dataset and truth tables were too large to fit within this paper, they are available upon request). Conforming with previous QCA research, we used the threshold of 0.90 for consistency in order to reduce the likelihood of true logical contradictions within the cases found. In line with recent QCA research, we too also implement the use of the parsimonious solution as this has been shown to be the best minimization technique for causal inference [50].

3.2. SVM Procedure

First of all, the 10-fold cross validation technique was used for the dataset, therefore, 10% of the given dataset was for testing and 90% was for training sets [68]. Second, SVM classifiers were constructed using polynomial, RBF and Pearson VII universal function (PUF) kernels. In addition, Bagging (BA), Random Subspace (RSS) and AdaBoost (ADA) were also used as ensemble classifiers of the SVM model. Lastly, accuracy, the F-score and AUC were used to observe the performance of the SVM models from a diverse perspective.

3.3. Dataset

The dataset for this study was conducted through a professional online survey company from March 13 to March 16, 2018. The sample used a proportional allocation extraction method based on age and gender by using the online surveyor’s panel. The final samples of the survey consisted of 339 people who are in their 20~50s, have experience in using the airport in the past but did not experience the biometric boarding system. The demographic composition of the samples is shown in Table 2. The questionnaire used in the dataset of this study was measured on a Likert five-point scale (“totally disagree” to “totally agree”). The questionnaires are shown in Table 3.

3.4. Constructing Classifiers

In order to test our dataset, the open source tool WEKA (https://www.cs.waikato.ac.nz/ml/index.html) on a Macintosh with a 3.2GHz Intel Core i7 CPU and 32GB 2400MHz DDR4 RAM was utilized. Three separate tests were implemented so that all three dependent variables (Weak Resistance and Strong Resistance) could be individually tested based on the identified independent variables. Within these three separate tests we explored SVM classifiers with polynomial, RBF and Pearson VII universal function (PUF) kernels. Consequently, comparisons between single SVM classifiers and ensemble SVM classifiers such as BA, RSS and ADA were performed. As we were dealing with a class imbalance and a multiclass problem, traditional binary classification metrics were not enough to serve a true representation of our results. Thus, we implemented the use of three metrics that are better suited to cope with these issues, these being accuracy, F-measure and ROC.

4. Results

4.1. csQCA

To begin with, we performed necessity tests on both the dependent variables in order to find out the necessary conditions. Before approaching necessity tests, one must be careful to not overestimate the ability of the computer to find necessity relations. Duşa advices “It is the researcher who should make sense out of all these necessity relations and select only those which have a theoretical meaning” [47] (p.122). By looking at Table 4, the first test based on the presence of WR, Models 1 and 4 show no relation to prior literature and thus we reject these necessity conditions. Models 2 and 3 however show potential conditions. First, Model 2 claims that the presence of PR OR UC is necessary to explain WR. Furthermore, Model 3 claims that the presence of PR OR PM is necessary to explain WR. It is clear that the presence of PR is the common factor present, however as it is within two models where the factors UC and PM could be the factors present, we have to negate the outcome to see if any further conditions appear that add rigor to this claim. As can be seen in the results, when we make the outcome no resistance, the condition whereby ~PR + PB is found. This model backs up the previous literature as it is clear that in order to have no Weak Resistance you need to have no perceived risks or perceived benefits. Thus, we conclude that PR is a necessary condition of WR.
The next test which was made on the dependent variable SR shows five possible models. Models 1, 3 and 5 do not have any prior theoretical explanation and thus we reject these necessity conditions. Models 2 and 3 shows the exact same conditions as seen within the WR dependent variable test and therefore we conclude the same explanation for this outcome as was made in test one. Once again in order to clarify the potential necessity of PR on SR we once again negated the outcome to see what factors were necessary. Two models were presented and first we rejected Model 2 as this model does not have any theoretical rigor based on prior literature. Next, and more importantly, we can see that ~PR is a necessary condition to explain ~SR. In layman’s terms, this shows that when there is no presence of SR, there is also no presence of PR. This condition conforms to prior literature and as this variable is alone in its condition, we can strongly affirm that PR is a necessary condition of the outcome SR.
Next we undertook sufficiency tests. As can be seen in Table 5, Weak Resistance is represented in two constellations combined by the logic AND function (+) creating an overall model consistency of 0.957. If we look at the model it explains that the presence of PR is a sufficient condition to explain Weak Resistance. In the first configuration, the presence of PR alongside the absence of PB and UC shows Weak Resistance with a consistency of 1. The next configuration shows that the presence of PR and the absence of PB and PM is a sufficient condition for Weak Resistance. Notably, in both configurations we can see the presence of PR and absence of PB, suggesting these are the main constructs that affect Weak Resistance.
For Strong Resistance, the configurations are slightly more complex. Based on the raw coverage, we can see that Configurations 1, 2 and 5 are the most prominent for explaining Strong Resistance, and therefore our focus will be put on explaining these. The first configuration conforms with the results from Weak Resistance whereby PR is present and PB is absent, suggesting that once again these two constructs when in this state have sufficiency in explaining Strong Resistance. Next, Configuration 2 once again shows that PR is present. However, interestingly, it shows that this is when PM is present too. This suggests that online information can potentially have a positive impact on the resistance of biometric e-gate technology adoption. Lastly, the most interesting outcome was seen in Configuration 5. First, it is clear that the absence of PB has an impact on Strong Resistance conforming to prior literature and our own hypothesis. However, the csQCA analysis also states that no PB OR UC OR PM is a sufficient condition. This could be explained through the notion that people who have an openness to new technology or to trying new innovations will negatively react to an innovation in the presence of perceiving no benefits. With this we argue that this can be because the feeling of no PB outweighs the feelings seen within UC and PM, and thus the outcome would be Strong Resistance as expectations would be outweighed by disappointment. Both models presented to explain both types of resistance can be visualized in Figure 1, whereby the red color represents instances of resistance as the output condition based on the sufficiency of the areas of each construct.

4.2. Single SVM Classifiers

Figure 2and Figure 3 indicate the evaluation results of performances of the single SVM classifiers with three different kernel functions (polynomial, RBF and PUF) on the given dataset. The evaluation of each performance was measured in accuracy, F-score and AUC.
As Figure 2 shows the performance evaluation for the dependent variable Weak Resistance, the polynomial kernel with an SVM classifier has a better performance than the others in terms of accuracy (58.42%). In addition, in terms of F-score, polynomial and RBF kernels had the best score (0.58), while the polynomial kernel acquired the best (0.74) for performance evaluation of AUC.
For the dependent variable Strong Resistance in Figure 3, contrary to the performance evaluation seen in the dependent variable Weak Resistance, the result shows that RBF calculates the best result (64.56%) in terms of accuracy. As is shown in Table 6, a single SVM classifier with a PUF kernel makes the best score (0.45) for the F-score. Nevertheless, the polynomial kernel has the best results in AUC performance evaluation (0.79)

4.3. Ensemble SVM Classifiers

Figure 4 and Figure 5 represent the performances of the ensemble SVM classifiers with Bagging. In addition, the models are processed with polynomial, RBF and PUF kernels in the given dataset. As the first experiment with single SVM classifier models, accuracy, F-score and AUC were measured.
Figure 4 shows BA with an SVM classifier, where the combination with the PUF function acquires the best performance results in accuracy (58.72%) and AUC (0.77) for the target variable Weak Resistance. On the other hand, polynomial and RBF obtain better performance measurement in the F-score (0.58).
For the dependent variable Strong Resistance in Figure 5, the performance with a single kernel function can be found for the SVM classifier with Bagging: RBF in accuracy (65.75%), PUF in F-score (0.50) and polynomial in AUC (0.86) show the best performance, respectively. Table 7 addresses best performance for ensemble SVM (BA).
The performance result in Figure 6, and the result of PUF with RSS SVM derives the best score in all performance metrics for Weak Resistance: 57.25% in accuracy, 0.51 in F-score and 0.75 in AUC, respectively.
For the dependent variable Strong Resistance in Figure 7, using an RBF kernel with RSS ensemble SVM models shows the superior performance in accuracy (65.16%). However, polynomial and RBF cannot be used in terms of F-score, since their datasets were overfitted, while the PUF kernel has the highest score (0.40). In terms of AUC, the polynomial kernel have a relatively better performance than other kernels (0.86). Table 8 summarizes the best performance for ensemble SVM classifiers (RSS).
Figure 8 and Figure 9 reveal the performances of the ADA ensemble SVM classifiers with three different kernel functions (polynomial, RBF and PUF) in terms of accuracy, F-score and AUC of the given dataset. As Figure 7 shows, the polynomial kernel has the highest performance result (58.42%) compared with RBF and PUF kernels with ADA ensemble SVM classifiers in accuracy for the dependent variable Weak Resistance. In terms of the F-score, polynomial and RBF both have the best performance (0.58), while, interestingly, PUF acquires superior performance (0.74) compared with other kernels in AUC evaluation.
For the dependent variable Strong Resistance in Figure 8, RBF presents relatively better performance in accuracy (64.56%) and F-score (0.30) performance measurement while PUF shows the best AUC performance (0.76) with ADA ensemble SVM models. Table 9 also states the best performance for ensemble SVM classifiers (ADA).
Below Table 10 shows overall results of SVM classifier performance evaluations for each dependent variable (Weak Resistance and Strong Resistance) that were used in this research.

5. Discussion

5.1. Key Findings

This study looked to explore the reasons why flight passengers are reluctant to adopt the biometric e-gate boarding system. The sustainability of airport operations relies on the airports’ ability to adapt to using more advanced, non-invasive and reliable technologies that can provide a satisfactory passenger service for generations to come. However, these technologies do not always provide the best solution. This paper has attempted to explore one such example of where a failed attempt to increase passenger service quality has led to managers looking for alternative solutions that can help to aid the sustainability of service quality despite the ever-increasing demands made through passenger numbers. Through an extensive self-report survey effort, 336 valid questionnaire samples were garnered and two types of resistance to the adoption of innovative technology, like biometric e-gate security, were analyzed in the form of Weak Resistance and Strong Resistance.
Our first key findings within this paper was using csQCA [29]. With csQCA we were able to find two sufficient configurations that explain Weak Resistance and three configurations that explain Strong Resistance to the biometric e-gate security system. Furthermore, all these configurations fit within the commonly used threshold of 0.90 [34]. Weak Resistance was firstly explained through (1) perceived risks * no perceived benefits * low user characteristics and (2) perceived risks * no perceived benefits * low propagation mechanisms. For Strong Resistance, the following three configurations were found: (1) perceived risks * no perceived benefits, (2) perceived risks * high propagation mechanism and (3) no perceived benefits * high user characteristics * high propagation mechanism. The overall findings from these configurations show that presence of perceived risks and absence of perceived benefits were the overarching key issues promoting resistance to air passengers from using the biometric e-gate technology. Therefore, future implementation of these services will have to take into consideration these factors in order to increase the chances of a successful implementation. Interestingly, the inclusion of propagation mechanisms was also seen in both dependent variables and therefore online information was seen to have an impact on the resistance to the technology. This means that managers should take into consideration online presence as well as the risks and benefits associated with the use of biometric e-gate services.
With csQCA one can find patterns within data and make decisions based on these findings, however it lacks the power to make quick and fast predictions based on the data. Therefore, we next wanted to explore the power of a classification technique that would help managers quickly assess the implementation of new technologies within the airport setting, including our case of biometric e-gate technology. Starting with the first dependent variable which attempted to predict people’s feelings on Weak Resistance. We found that a Bagging algorithm with a PUF kernel SVM model yields the best overall model with an accuracy of 58.72%, F-score of 0.57 and AUC of 0.77. Overall, in our test of this dependent variable, it is safe to presume that the best performance of the models lies with the use of BA in most cases. Another point to note about this test is that the overall level of accuracy and F-score is fairly low, proving this task to be a challenging one in the machine-learning domain. Another key outcome of this test was seen in the F-score by RSS SVM polynomial and RBF with the score of 0.0 meaning that the model was overfit.
Next, if we look at the second dependent variable Strong Resistance, the best accuracy was seen within the BA SVM RBF kernel model. F-score evaluation shows the best score with a BA SVM PUF of 0.50. However, as previously stated, due to the type of problem we are trying to learn with these models, the F-score is not reliable enough on its own. However, the AUC score presented by the BA SVM polynomial with 0.86 is reliable enough to compare to the result of the F-score. Once again, an ensemble model has been the best performer in this test compared to single SVM models. The overall performance of the BA SVM model shows more confidence in these models’ ability to predict the target variable (Strong Resistance) and yield greater reliability compared to the model’s usage seen in the dependent variable Weak Resistance.

5.2. Theoretical Implications

Theoretically, this paper has the following implications. First, this paper enriches the innovation adoption field of research by exploring resistance to innovation. This allows for a reverse analysis of the reasons people adopt innovation and thus can provide unique implications based on the results [27]. The context of biometric e-gate technology provides a case in which complex decisions have to be made by passengers who are entrenched within an old security system that has been in place for decades at airports. However, these systems are slowly coming to capacity and thus their sustainability is in question. With our analysis we have used an SVM classification methodology which can help managers quickly assess the resistance to biometric e-gate technology. Secondly we have shown how they can successfully determine what the problems are based on the use of csQCA, which allows for the identification of “multiple conjunctural causation” [48].

5.3. Limitations and Future Research

This study has limitations. First, this research was not able to measure the detailed experience of the consumers’ past encounters with biometric-related services. Considering the fact that biometric identification is already being used in various areas such as building entries, mobile payment and ATM authentication, it will also be meaningful to test the performance of this prediction model for resistance against similar services from other fields. In particular, analyzing the case where the usage rate of the “Automated Immigration Service of Korea” using fingerprints have risen from 24.8% in 2014 to 43.8% in 2017 [69] may let us find an important issue that this research was not able to consider for another sustainability issue at airports. Therefore, it is necessary to carry out a study comparing performance of SVM prediction models on resistance in the adoption process of various innovative products and services that utilizes biometric recognition.

Author Contributions

Conceptualization, K.C.L. and C.K.; methodology, K.C.L. and C.K.; software, F.J.C. and C.K.; validation, C.K. and F.J.C.; formal analysis, C.K. and F.J.C.; investigation, C.K., F.J.C. and K.C.L.; data curation, C.K.; writing—original draft preparation, C.K. and F.J.C; writing—review and editing, K.C.L.; supervision, K.C.L.

Funding

The APC was funded by BK21 project by the government of Korea.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Park, J.-W.; Ryu, Y.K. Investigating the Effects of Airport Servicescape on Airport Users’ Behavioral Intentions: A Case Study of Incheon International Airport Terminal 2 (T2). Sustainability 2019, 11, 4171. [Google Scholar] [CrossRef]
  2. Wensveen, J. Air Transportation: A Management Perspective; Routledge: London, UK, 2018. [Google Scholar]
  3. Xu, X.; Zhou, Z.; Dou, Y.; Tan, Y.; Liao, T. Sustainable Queuing-Network Design for Airport Security Based on the Monte Carlo Method. Sustainability 2018, 10, 1. [Google Scholar] [CrossRef]
  4. Morosan, C. An empirical examination of US travelers’ intentions to use biometric e-gates in airports. J. Air Trans. Manag. 2016, 55, 120–128. [Google Scholar] [CrossRef]
  5. Ali, A.B. Travel and Trourism Management; Delhi: PHI Learning Private Limited: Delhi, India, 2015. [Google Scholar]
  6. Kim, M.W. Social Responsibility Report 2018; Korea Airports Corporation: Seoul, Korea, 2018. [Google Scholar]
  7. Marx, A.; Rihoux, B.; Ragin, C. The origins, development, and application of Qualitative Comparative Analysis: The first 25 years. Eur. Political Sci. Rev. 2014, 6, 115–142. [Google Scholar] [CrossRef]
  8. Zu, X.; Wu, Y.; Zhang, Z.; Yu, L. Prediction of Consumption Choices of Low-Income Groups in a Mixed-Income Community Using a Support Vector Machine Method. Sustainability 2019, 11, 3981. [Google Scholar] [CrossRef]
  9. Alimi, O.A.; Ouahada, K.; Abu-Mahfouz, A.M. Real Time Security Assessment of the Power System Using a Hybrid Support Vector Machine and Multilayer Perceptron Neural Network Algorithms. Sustainability 2019, 11, 3586. [Google Scholar] [CrossRef]
  10. Xue, Y.; Ren, J.; Bi, X. Impact of Influencing Factors on CO2 Emissions in the Yangtze River Delta during Urbanization. Sustainability 2019, 11, 4183. [Google Scholar] [CrossRef]
  11. Mrówczyńska, M.; Sztubecka, M.; Skiba, M.; Bazan-Krzywoszańska, A.; Bejga, P. The Use of Artificial Intelligence as a Tool Supporting Sustainable Development Local Policy. Sustainability 2019, 11, 4199. [Google Scholar] [CrossRef]
  12. Olivares-Mercado, J.; Toscano-Medina, K.; Sánchez-Perez, G.; Hernandez-Suarez, A.; Perez-Meana, H.; Sandoval Orozco, A.L.; García Villalba, L.J. Early Fire Detection on Video Using LBP and Spread Ascending of Smoke. Sustainability 2019, 11, 3261. [Google Scholar] [CrossRef]
  13. Huang, M.; Chen, C.; Lin, W.; Ke, S.; Tsai, C. SVM and SVM Ensembles in Breast Cancer Prediction. PLoS ONE 2017, 12, 1–14. [Google Scholar] [CrossRef]
  14. Sasikala, B.S.; Biju, V.G.; Prashanth, C.M. Kappa and accuracy evaluations of machine learning classifiers. In Proceedings of the 2nd IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology, Bangalore, India, 19–20 May 2017. [Google Scholar]
  15. Abdelaal, M.M.A.; Sena, H.A.; Farouq, M.W.; Salem, A.-B.M. Using data mining for assessing diagnosis of breast cancer. In Proceedings of the IEEE International Multiconference on Computer Science and Information Technology, Wisła, Poland, 18–20 October 2010; pp. 11–17. [Google Scholar]
  16. Ahmad, L.G.; Eshlaghy, A.T.; Poorebrahimi, A.; Ebrahimi, M.; Razavi, A.R. Using three machine learning techniques for predicting breast cancer recurrence. J. Health Med. Inform. 2013, 4, 124. [Google Scholar]
  17. Huang, C.L.; Liao, H.C.; Chen, M.C. Prediction model building and feature selection with support vector machines in breast cancer diagnosis. Expert Syst. Appl. 2008, 34, 578–587. [Google Scholar] [CrossRef]
  18. Salama, G.I.; Abdelhalim, M.B.; Zeid, M.A.E. Experimental comparison of classifiers for breast cancer diagnosis. In Proceedings of the 2012 Seventh International Conference on Computer Engineering & Systems (ICCES), Cairo, Egypt, 27–29 November 2012; pp. 180–185. [Google Scholar]
  19. Senturk, Z.K.; Kara, R. Breast cancer diagnosis via data mining: Performance analysis of seven different algorithms. Comput. Sci. Eng. 2014, 4, 35–46. [Google Scholar] [CrossRef]
  20. You, H.; Rumbe, G. Comparative study of classification techniques on breast cancer FNA biopsy data. Int. J. Artif. Intell. Interact. Multimed. 2010, 1, 6–13. [Google Scholar] [CrossRef]
  21. Kittler, J.; Hatef, M.; Duin, R.P.W.; Matas, J. On combining classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 226–239. [Google Scholar] [CrossRef] [Green Version]
  22. Blalock, G.; Kadiyali, V.; Simon, D.H. The Impact of Post-9/11 Airport Security Measures on the Demand for Air Travel. J. Law Econ. 2007, 50, 731–755. [Google Scholar] [CrossRef]
  23. Farrell, S. How airports can fly to self-service biometrics. Biom. Technol. Today 2016, 2016, 5–7. [Google Scholar] [CrossRef]
  24. Wayman, J.; Jain, A.; Maltoni, D.; Maio, D. An Introduction to Biometric Authentication Systems; Springer: London, UK, 2005. [Google Scholar]
  25. Byun, S. Perceptions towards High-tech Service among its Early Adopters: In the Case of Fingerprint-enabled ATMs. Mark. Manag. Res. 2010, 15, 43–69. [Google Scholar]
  26. Rogers, E.M. Diffusion of Innovations; Free Press: New York, NY, USA, 2003. [Google Scholar]
  27. Sheth, J.N. Psychology of innovation resistance: The less developed concept. Res. Mark. 1981, 4, 273–283. [Google Scholar]
  28. Garner, S.J. Perceived risk and information sources in services purchasing. Mid-Atl. J. Bus. 1986, 24, 49–58. [Google Scholar]
  29. Kaplan, L.B.; Szybillo, G.J.; Jacoby, J. Components of perceived risk in product purchase: A cross-validation. J. Appl. Psychol. 1974, 59, 287–291. [Google Scholar] [CrossRef]
  30. Ko, H.; Jung, J.; Kim, J.; Shim, S.W. Cross-cultural differences in perceived risk of online shopping. J. Interact. Advert. 2004, 4, 20–29. [Google Scholar] [CrossRef]
  31. Shimp, T.A.; Bearden, W.O. Warranty and other extrinsic cue effects on consumers’ risk perceptions. J. Consum. Res. 1982, 9, 38–46. [Google Scholar] [CrossRef]
  32. Song, H.Y.; Jung, J.M.; Kim, Y. Factors affecting postponement, rejection, and opposition to 3DTV adoption. Media Econ. Cult. 2013, 11, 7–50. [Google Scholar]
  33. Stone, R.N.; Grønhaug, K. Perceived risk: Further considerations for the marketing discipline. Eur. J. Mark. 1993, 27, 39–50. [Google Scholar] [CrossRef]
  34. Featherman, M.S.; Pavlou, P.A. Predicting e-services adoption: A perceived risk facets perspective. Int. J. Hum.-Comput. Stud. 2003, 59, 451–474. [Google Scholar] [CrossRef]
  35. Moon, D.Y. Diffusion Factors of STEAM Education-Based on Rogers’ Diffusion of Innovations. Theory-J. Korean Pract. Arts Educ. 2016, 29, 133–150. [Google Scholar] [CrossRef]
  36. Mintrom, M. Policy Entrepreneurs and the Diffusion of Innovation. Am. J. Political Sci. 1997, 41, 738–770. [Google Scholar] [CrossRef]
  37. Rogers, E.M.; Shoemaker, F.F. Communication of Innovations: A Cross-Cultural Approach; Free Press: New York, NY, USA, 1971. [Google Scholar]
  38. Ram, S. A model of innovation resistance. Adv. Consum. Res. 1987, 14, 208–212. [Google Scholar]
  39. Bearden, W.O.; Netemeyer, R.G.; Teel, J.E. Measurement of consumer susceptibility to interpersonal influence. J. Consum. Res. 1989, 15, 473–481. [Google Scholar] [CrossRef]
  40. van Rijnsoever, F.J.; Oppewal, H. Predicting early adoption of successive video player generations. Technol. Forecast. Soc. Chang. 2011, 79, 558–569. [Google Scholar] [CrossRef]
  41. Fain, D.; Roberts, M.L. Technology vs. consumer behavior: The battle for the financial services customer. J. Direct Mark. 1997, 11, 44–54. [Google Scholar] [CrossRef]
  42. Laukkanen, T.; Kiviniemi, V. The role of information in mobile banking resistance. Int. J. Bank Mark. 2010, 28, 372–388. [Google Scholar] [CrossRef]
  43. Laukkanen, T.; Sinkkonen, S. Consumer resistance to internet banking: Postponers, opponents and rejectors. Int. J. Bank Mark. 2008, 26, 440–455. [Google Scholar] [CrossRef]
  44. Haas, E.P. Back to the Future-The Use of Biometrics, Its Impact of Airport Security, and How This Technology Should Be Governed. J. Air Law Commer. 2004, 69, 459–489. [Google Scholar]
  45. Szmigin, I.; Foxall, G. Three forms of innovation resistance: The case of retail payment methods. Technovation 1998, 18, 459–468. [Google Scholar] [CrossRef]
  46. Ragin, C.C. The Comparative Method: Moving beyond Qualitative and Quantitative Strategies; Univercity of California Press: Berkeley, CA, USA, 2014. [Google Scholar]
  47. Duşa, A. QCA with R: A Comprehensive Resource; Springer: Berlin, Germany, 2018. [Google Scholar]
  48. Rihoux, B.; Ragin, C.C. Configurational Comparative Methods: Qualitative Comparative Analysis (QCA) and Related Techniques; Sage Publications: Thousand Oaks, CA, USA, 2008. [Google Scholar]
  49. Schneider, C.Q.; Wagemann, C. Set-Theoretic Methods for the Social Sciences: A Guide to Qualitative Comparative Analysis; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  50. Liu, H.; Subramanian, A.M.; Hang, C. In Search of the Perfect Match: A Configurational Approach to Technology Transfer in Singapore. IEEE Trans. Eng. Manag. 2019, 1–12. [Google Scholar] [CrossRef]
  51. Byun, H.; Lee, S.W. A survey on pattern recognition applications of support vector machines. Int. J. Pattern Recognit. Artif. Intell. 2003, 17, 459–486. [Google Scholar] [CrossRef]
  52. Tsai, C.F.; Cheng, K.C. Simple instance selection for bankruptcy prediction. Knowl.-Based Syst. 2012, 27, 333–342. [Google Scholar] [CrossRef]
  53. Yekkehkhany, B.; Safari, A.; Homayouni, S.; Hasanlou, M. A comparison study of different kernel functions for SVM-based classification of multi-temporal polarimetry SAR data. Int. Arch. Photogram Rem. Sens. Spatial Inform. Sci. 2014, 2, W3. [Google Scholar] [CrossRef]
  54. Das, S.R.; Das, K.; Mishra, D.; Shaw, K.; Mishra, S. An empirical comparison study on kernel based support vector machine for classification of gene expression data set. Procedia Eng. 2012, 38, 1340–1345. [Google Scholar] [CrossRef]
  55. Buck, T.A.E.; Zhang, B. SVM kernels optimization: An example in yeast protein subcellular localization prediction. Project Report, School of Computer Science, Carnegie Mellon University, Pittsburgh, USA. Available online: https://pdfs.semanticscholar.org/0671/da2cf0d288d526a17aa2d7eddebe489429ff.pdf (accessed on 8 June 2018).
  56. Mezghani, D.B.A.; Boujelbene, S.Z.; Ellouze, N. Evaluation of SVM kernels and conventional machine learning algorithms for speaker identification. Int. J. Hybrid Inform. Tech. 2010, 3, 23–34. [Google Scholar]
  57. Baten, A.K.M.A.; Chang, B.C.H.; Halgamuge, S.K.; Li, J. Splice site identification using probabilistic parameters and SVM classification. BMC Bioinformatics 2006, 7, S15. [Google Scholar] [CrossRef]
  58. Zhang, G.; Ge, H. Support vector machine with a Pearson VII function kernel for discriminating halophilic and non-halophilic proteins. Comput. Biol. Chem. 2013, 46, 16–22. [Google Scholar] [CrossRef] [Green Version]
  59. Joshi, J.; Doshi, R.; Patel, J. Diagnosis and prognosis breast cancer using classification rules. International J. Eng. Res. Gen. Sci. 2014, 2, 315–323. [Google Scholar]
  60. Aruna, S.; Rajagopalan, S.P.; Nandakishore, L.V. Knowledge based analysis of various statistical tools in detecting breast cancer. In Proceedings of the International Conference on Computer Science, Engineering and Applications, Chennai, India, 15–17 July 2011; pp. 37–45. [Google Scholar]
  61. Happel, B.L.M.; Murre, J.M.J. The design and evolution of modular neural network architectures. Neural Netw. 1994, 7, 985–1004. [Google Scholar] [CrossRef]
  62. Schapire, R.E. The strength of weak learnabilty. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef]
  63. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  64. Bryll, R. Attribute bagging: Improving accuracy of classifier ensembles by using random feature subsets. Pattern Recognit. 2003, 36, 1291–1302. [Google Scholar] [CrossRef]
  65. Ho, T.K. The Random Subspace Method for Constructing Decision Forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar]
  66. Bellman, R.E. Dynamic Programming; Courier Dover Publications: New York, NY, USA, 2003. [Google Scholar]
  67. Freund, Y.; Schapire, R.E. Experiments with a New Boosting Algorithm; International Conference on Machine Learning: Bari, Italy, 1996; pp. 148–156. [Google Scholar]
  68. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the 14th International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 20–25 August 1995; pp. 1137–1143. [Google Scholar]
  69. Kim, Y. Incheon Airport Welcomed 100 Million Passengers with 72 Automatic Immigration Control; Edaily: Seoul, Korea, 2017. [Google Scholar]
Figure 1. Venn diagrams of sufficient conditions for Weak (a) and Strong (b) Resistance.
Figure 1. Venn diagrams of sufficient conditions for Weak (a) and Strong (b) Resistance.
Sustainability 11 05349 g001
Figure 2. Performance of single SVM classifiers for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 2. Performance of single SVM classifiers for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g002
Figure 3. Performance of single SVM classifiers for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 3. Performance of single SVM classifiers for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g003
Figure 4. Performance of an ensemble SVM classifier (BA) for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 4. Performance of an ensemble SVM classifier (BA) for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g004
Figure 5. Performance of an ensemble SVM classifier (BA) for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 5. Performance of an ensemble SVM classifier (BA) for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g005
Figure 6. Performance of an ensemble SVM classifier (RSS) for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 6. Performance of an ensemble SVM classifier (RSS) for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g006
Figure 7. Performance of an ensemble SVM classifier (RSS) for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 7. Performance of an ensemble SVM classifier (RSS) for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g007
Figure 8. Performance of an ensemble SVM classifier (ADA) for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 8. Performance of an ensemble SVM classifier (ADA) for Weak Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g008
Figure 9. Performance of an ensemble SVM classifier (ADA) for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Figure 9. Performance of an ensemble SVM classifier (ADA) for Strong Resistance based on accuracy (a), F-Score (b) and AUC (c).
Sustainability 11 05349 g009
Table 1. Recent studies on prediction using a support vector machine (SVM).
Table 1. Recent studies on prediction using a support vector machine (SVM).
StudiesKernelsEvaluation
Zu et al. [8]OptimalPrecision
Alimi et al. [9]Linear, Poly, RBFPrecision, Recall, F-Score
Xue et al. [10]RBFAccuracy
Olivares-Mercado et al. [12]RBFPrecision, Recall, Accuracy, F-Score
Joshi et al. [59]RBFAccuracy
Ahmad et al. [16] RBFAccuracy
Aruna et al. [60]RBFAccuracy
Abdelaal et al. [15]RBFAUC
You and Rumbe [20]Poly, RBF, SigmoidAccuracy
Huang et al. [17]RBFAccuracy
Table 2. Demographic information of the dataset.
Table 2. Demographic information of the dataset.
VariablesTypeN%
GenderMale16849.6
Female17150.4
Age20s8424.8
30s8424.8
40s8424.8
50s8725.7
Frequency of Airport Use 1~3 Times29787.6
4~6 Times308.8
6~9 Times61.8
Over 10 Times61.8
Table 3. Survey questions of the dataset.
Table 3. Survey questions of the dataset.
Type
(the Number of Questions)
Avg.
(SD)
QuestionSource
Perceived
Risk
Temporal Risk
(3)
3.27
(0.879)
Biometric e-gates will take too much time to learn how to use.Stone and Grønhaug [33]
3.40
(0.899)
If I use biometric e-gates, I think it will require more unnecessary procedures. It will take more time.
3.33
(0.885)
I think I will waste even more time since I have to learn how to use biometric e-gates.
Socia Risk
(3)
3.45
(0.874)
Using biometric e-gates will have a negative impact on my perception/reputation.Featherman and Pavlou [34]
3.82
(0.884)
If I use biometric e-gates, I’m sure that other people will find me strange.
2.65
(0.848)
I’ll use biometric e-gates after seeing other people using it without going through any trouble.
Physical Risk
(3)
3.43
(0.925)
I am concerned about the physical fatigue caused by the use of biometric e-gates.Stone and Grønhaug [33]
3.27
(0.976)
I’m worried that biometric e-gates might cause physical discomfort or other physical side effects.
2.96
(0.966)
Since biometric e-gates has yet to be verified as safe, I am concerned about the physical risks using the system.
Functional Risk
(3)
2.53
(0.904)
I’m not sure whether the biometric e-gates system will work as well as it is said.
2.86
(0.870)
I’m not sure if biometric e-gates will provide me with as much convenience and benefits as I’d expect.
2.86
(0.877)
I’m not sure if the speed and convenience offered by biometric e-gates will be practical enough.
Perceived
Benefits
Relative Advantage
(3)
2.43
(0.838)
Biometric e-gates will be more interesting compared to the conventional boarding system.Rogers [26]
3.09
(0.864)
If I use biometric e-gates, other people will think of me as being smart.
2.36
(0.758)
The use of biometric e-gates will require less work from my side compared to the conventional boarding procedure.
Compatibility
(3)
2.92
(0.788)
Using biometric e-gates is in line with my values.
3.03
(0.837)
Biometric e-gates is not much different from the conventional boarding system.
2.65
(0.787)
Introducing the biometric e-gates procedure to the current boarding system is not a big problem.
Trialability
(3)
2.40
(0.776)
If I could try biometric e-gates right now, I think it would make me more open to the idea of using it in the future.
2.41
(0.842)
If airlines or airports hold events about using biometric e-gates, it would definitely motivate me to use the system in the future.
2.27
(0.793)
If there are more types of biometric e-gates in the future, it would be helpful for us to use the system with ease.
User
Characteristics
Openness to Innovation
(3)
2.71
(0.776)
I tend to pursue new ideas or experiences.Bearden, Netemeyer [39]
2.67
(0.841)
I like to experience new things and make changes in my daily life.
2.73
(0.856)
I am more interested in using new products or services than other people.
Independent Decision Making
(3)
2.42
(0.751)
Before using a new product or service, I tend to look for information provided by others.van Rijnsoever and Oppewal [40]
2.50
(0.719)
I decide whether to use a new product or service after hearing the opinions from others who have already adopted them.
2.48
(0.715)
I tend to be influenced by the evaluation from existing users when using new products or services.
Propagation
Mechanism
Provision of Online Information
(3)
2.65
(0.775)
Online information such as an Internet article or a SNS event gave me good information about biometric e-gates.Song et al. [32]
2.82
(0.787)
I was able to understand the biometric e-gates process through the Internet articles or SNS.
3.32
(0.880)
I have often heard and read about biometric e-gates through the Internet articles or SNS.
Provision of Offline Information
(3)
3.04
(0.821)
Newspaper articles, TV reports gave me good information about biometric e-gates.
2.95
(0.812)
I was able to understand the newspaper articles, TV reports and installation ads about biometric e-gates.
3.45
(0.856)
I have often been approached with newspaper articles, TV reports and installation ads about biometric e-gates.
ResistanceWeak Resistance
(3)
2.71
(0.835)
I’ll postpone my decision on using biometric e-gates until it turns out to be beneficial to me.Szmigin and Foxall [45]
2.84
(0.889)
I won’t use biometric boarding until I’m sure I need biometric e-gates.
2.88
(0.879)
I won’t use biometric e-gates until I reach the conclusion that biometric boarding is a great system for me.
Strong Resistance
(3)
3.40
(0.869)
I don’t want to use biometric e-gates.
3.24
(0.813)
Biometric e-gates isn’t for me.
3.35
(0.810)
I don’t need biometric e-gates.
Table 4. Analysis of necessary conditions on resistance.
Table 4. Analysis of necessary conditions on resistance.
ModelincIN *RoN **covN ***
Weak Resistance is Present (WR)~PR + ~PB0.9210.2080.670
PR + UC0.9300.2110.676
PR + PM0.9210.2460.681
UC + ~PM0.9210.2080.670
Weak Resistance is not present (~WR)~PR + PB0.9550.2240.373
~PR + ~PM0.9020.2310.356
Strong Resistance is present (SR)~PR + ~PB0.9220.4910.910
PR + UC0.9220.4910.910
PR + PM0.9160.5610.919
UC + ~PM0.9190.4820.907
~PR + PB + ~UC + ~PM0.9060.4830.900
Strong Resistance is not present (~SR)~PR0.9030.2540.108
PB + ~UC + ~PM0.9030.2730.110
* Coverage; ** Relevance of Necessity; *** Raw Coverage
Table 5. Analysis of sufficient conditions on resistance.
Table 5. Analysis of sufficient conditions on resistance.
Weak ResistanceStrong Resistance
ConstructConfiguration
1212345
PR ×
PB××× ×
UC× ×
PM × ×
Consistency1.0000.9521.0000.9640.9521.0000.965
PRI1.0000.9521.0000.9640.9521.0000.965
Raw Coverage0.0260.0880.1690.1720.0650.0320.266
Unique coverage0.0090.0700.0680.0620.0490.0260.172
Overall solution consistency0.9570.962
Overall solution coverage0.0970.494
Table 6. Best performance for single SVM classifiers.
Table 6. Best performance for single SVM classifiers.
Dependent VariableAccuracyF-ScoreAUC
Weak ResistancePolynomialPolynomial, RBFPolynomial
Strong ResistanceRBFPUFPolynomial
Table 7. Best Performance for ensemble SVM classifiers (BA).
Table 7. Best Performance for ensemble SVM classifiers (BA).
Dependent VariableAccuracyF-ScoreAUC
Weak ResistancePUFPolynomial, RBFPUF
Strong ResistanceRBFPUFPolynomial
Table 8. Best performance for ensemble SVM classifiers (RSS)
Table 8. Best performance for ensemble SVM classifiers (RSS)
Dependent VariableAccuracyF-ScoreAUC
Weak ResistancePUFPUFPolynomial, PUF
Strong ResistanceRBFPUFPolynomial
Table 9. Best performance for ensemble SVM classifiers (ADA).
Table 9. Best performance for ensemble SVM classifiers (ADA).
Dependent VariableAccuracyF-ScoreAUC
Weak ResistancePolynomialPolynomial, RBFPUF
Strong ResistanceRBFRBF, PUFPUF
Table 10. The best performance results of SVM classifiers for each dependent variable.
Table 10. The best performance results of SVM classifiers for each dependent variable.
Weak ResistanceSingle SVMBA SVMRSS SVMADA SVM
AccuracyPolynomialPUFPUFPolynomial
58.42%58.72% *57.25%58.42%
F-ScorePolynomial, RBFPolynomial, RBFPUFPolynomial, RBF
0.58 *0.58 *0.510.58 *
AUCPolynomialPUFPolynomial, PUFPUF
0.740.77 *0.750.74
Strong ResistanceSingle SVMBA SVMRSS SVMADA SVM
AccuracyRBFRBFRBFRBF
64.56%65.75% *65.16%64.56%
F-ScorePUFPUFPUFRBF, PUF
0.450.50 *0.400.30
AUCPolynomialPolynomialPolynomialPUF
0.790.86 *0.850.76
* The highest score in row

Share and Cite

MDPI and ACS Style

Kim, C.; Costello, F.J.; Lee, K.C. Integrating Qualitative Comparative Analysis and Support Vector Machine Methods to Reduce Passengers’ Resistance to Biometric E-Gates for Sustainable Airport Operations. Sustainability 2019, 11, 5349. https://doi.org/10.3390/su11195349

AMA Style

Kim C, Costello FJ, Lee KC. Integrating Qualitative Comparative Analysis and Support Vector Machine Methods to Reduce Passengers’ Resistance to Biometric E-Gates for Sustainable Airport Operations. Sustainability. 2019; 11(19):5349. https://doi.org/10.3390/su11195349

Chicago/Turabian Style

Kim, Cheong, Francis Joseph Costello, and Kun Chang Lee. 2019. "Integrating Qualitative Comparative Analysis and Support Vector Machine Methods to Reduce Passengers’ Resistance to Biometric E-Gates for Sustainable Airport Operations" Sustainability 11, no. 19: 5349. https://doi.org/10.3390/su11195349

APA Style

Kim, C., Costello, F. J., & Lee, K. C. (2019). Integrating Qualitative Comparative Analysis and Support Vector Machine Methods to Reduce Passengers’ Resistance to Biometric E-Gates for Sustainable Airport Operations. Sustainability, 11(19), 5349. https://doi.org/10.3390/su11195349

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop