Next Article in Journal
Numerical Investigation on the Thermal Performance of Nanofluid-Based Cooling System for Synchronous Generators
Previous Article in Journal
Role of Quantum Entropy and Establishment of H-Theorems in the Presence of Graviton Sinks for Manifestly-Covariant Quantum Gravity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Entropic Statistical Description of Big Data Quality in Hotel Customer Relationship Management

by
Lydia González-Serrano
1,
Pilar Talón-Ballestero
1,*,
Sergio Muñoz-Romero
1,2,
Cristina Soguero-Ruiz
1 and
José Luis Rojo-Álvarez
1,2
1
Department of Business and Management, Rey Juan Carlos University, 28943 Madrid, Spain
2
Department of Theory and Comunications, Telematics and Computing Systems, Rey Juan Carlos University, 28943 Madrid, Spain
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(4), 419; https://doi.org/10.3390/e21040419
Submission received: 23 February 2019 / Revised: 7 April 2019 / Accepted: 17 April 2019 / Published: 19 April 2019
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
Customer Relationship Management (CRM) is a fundamental tool in the hospitality industry nowadays, which can be seen as a big-data scenario due to the large amount of recordings which are annually handled by managers. Data quality is crucial for the success of these systems, and one of the main issues to be solved by businesses in general and by hospitality businesses in particular in this setting is the identification of duplicated customers, which has not received much attention in recent literature, probably and partly because it is not an easy-to-state problem in statistical terms. In the present work, we address the problem statement of duplicated customer identification as a large-scale data analysis, and we propose and benchmark a general-purpose solution for it. Our system consists of four basic elements: (a) A generic feature representation for the customer fields in a simple table-shape database; (b) An efficient distance for comparison among feature values, in terms of the Wagner-Fischer algorithm to calculate the Levenshtein distance; (c) A big-data implementation using basic map-reduce techniques to readily support the comparison of strategies; (d) An X-from-M criterion to identify those possible neighbors to a duplicated-customer candidate. We analyze the mass density function of the distances in the CRM text-based fields and characterized their behavior and consistency in terms of the entropy and of the mutual information for these fields. Our experiments in a large CRM from a multinational hospitality chain show that the distance distributions are statistically consistent for each feature, and that neighbourhood thresholds are automatically adjusted by the system at a first step and they can be subsequently more-finely tuned according to the manager experience. The entropy distributions for the different variables, as well as the mutual information between pairs, are characterized by multimodal profiles, where a wide gap between close and far fields is often present. This motivates the proposal of the so-called X-from-M strategy, which is shown to be computationally affordable, and can provide the expert with a reduced number of duplicated candidates to supervise, with low X values being enough to warrant the sensitivity required at the automatic detection stage. The proposed system again encourages and supports the benefits of big-data technologies in CRM scenarios for hotel chains, and rather than the use of ad-hoc heuristic rules, it promotes the research and development of theoretically principled approaches.

1. Introduction

Customer Relationship Management (CRM) is a customer-centric business strategy which a company employs to improve customer experience and satisfaction by customizing products and services to their customers’ needs [1]. CRM systems offer companies the possibility of obtaining competitive advantages through better knowledge while maintaining a close relationship with their customers [2,3]. For the hospitality sector, CRM is considered as one of the best strategies to improve a company’s results and to ensure its long-term survival [4,5,6,7,8]. Accordingly, CRM systems are nowadays a fundamental tool in this sector, especially when properly implemented, as there is a large amount of customer data that are integrated by hotels today. These data can be turned into useful knowledge [9,10,11,12,13,14] to improve customer satisfaction and retention [15]. Furthermore, the growing importance of CRM systems has led to paying greater attention to the value of customer data as a strategic asset [16]. However, some studies [17] estimate that 15% of the data collected in a company’s customer databases are erroneous, missing, or incorrect, which all generates erratic data, i.e., databases held by such businesses that are inaccurate or partially lacking sense or cohesion. Overall, some authors have estimated that this could mean added costs as high as close to 10% of some company profits [18], and other authors aimed to measure the general average financial impact of poor data on businesses up to $9.7 million per year, though it is complex to give an accurate estimate nowadays as it is a problem raising during the last years after the increase of digital stored data [19]. In this scenario, data quality refers to the degree of accuracy, accessibility, relevance, timeliness, and data integrity [20,21]. The impact of data quality is crucial for companies and a key factor in the successful implementation of a CRM system [22], as well as one of the biggest challenges [23]. Despite this, it has been pointed out as one of the most underestimated issues in the implementation of CRM [18,24,25,26,27]. The data quality does not only depend on accurate data introduction in the system, but it is also affected by the way the data are stored, managed, and shared [28]. Among the main challenges in this area, we can consider the existence of duplicated entries. As the amount of data grows exponentially, the possibility of having redundant data notably increases, which in some cases cannot be acceptable or affordable, meaning that a preprocessing stage is required. The resolution of entities or the link of records is the task of detecting records that refer to the same entity in different databases. This task has many applications in different fields of research, and for this reason, different terms have been adopted for the same concept, such as duplicate detection, deduplication, reference matching, object identification, combination and purge, object consolidation, or reference conciliation. Entities mostly commonly correspond to people (for example, patients, clients, taxpayers, and travellers), companies, consumer products, or publications and appointments, among others [29]. In this work we use the term duplicate to broadly refer to this data-quality problem.
One origin of data duplicity that can be pointed out is the lexical heterogeneity that occurs when the databases use different representations to refer to the same real-world object [30]. For example, for a hotel chain with a large number of hotels that use management programs or Property Management Systems (PMS), the fusion of multiple data sources from these programs may be very complicated. In this scenario, a chain hotel can enter a customer’s data with a customer identification number (ID) that is slightly different from a previous record that was already entered in the PMS, so that several records can be created for the same client, and this could increase rapidly through the different hotels within the chain. On the other hand, common names [31], misspellings, nicknames or abbreviated names, phone numbers, and addresses can be particularly troublesome to be standardized, especially when it comes to guest data from many countries. For example, and according to Revinate (an international CRM technological firm) [32], considering a chain with 50,000 profiles, the combinations in the introduction of customer data can originate up to 2.5 billion records of profiles to be evaluated. Different problems can result from the presence of errors in the data collection, namely, hotels may not recognize repeating guests, duplicate communications could be sent to the same person, or even the expense or the nights could be not correctly calculated. In short, duplicities can lead to erroneous decisions by the chain that affect its profitability and its brand. Although the problem of duplicities is a partly avoidable organizational problem, for instance by having data collection rules for all employees in the chain (and most chains already have) [28], a question arises: What happens to the data that were collected before the implementation of these rules? Moreover, what happens if even after implementing these rules, mistakes still occur in the data collection? It cannot be assumed that the receptionist or booker will have access to the duplicate records, but the fundamental problem is that, in many cases, there are duplicate records that the CRM does not detect. It has been estimated that near 2% of contact data can go bad each month, which is almost a quarter of the database annually [33]. For all these reasons, computer programs are necessary to detect and eliminate duplicities.
On the other hand, there are different tools for data cleaning based on different criteria, such as individual fields or entire documents. In this paper, we focus on finding duplicated customers from the analysis of the individual fields of the CRM entry coming from lexical heterogeneity. Some authors analyze duplication manually by detecting variations in the values of certain variables (such as city or company) in a customer database [34]. Other authors have used different techniques based on character-based similarity metrics for matching fields with string data [30]. Some examples are the Edit Distance [35], the Affine Gap Distance [36], the Smith-Waterman Distance [37], or the Jaro Distance [38], among others. In this work, we focus on the possibilities to find duplicated fields in a hotel-chain CRM by using the Edit Distance, which is built according to the general idea that two values in a string field are duplicated when, by only changing a reduced number of characters, we can transform one value into another [39]. Specifically, we propose using one of the most popular Edit Distance algorithms, called the Levenshtein distance [35], to address duplicated-customer identification as a large-scale data analysis in the CRM from a multinational hospitality chain in Europe, with more than 300 hotels. To support the implementation of comparison strategies, a big-data implementation using map-reduce is used. Aiming to base it on a principled approach to generate efficient yet computationally affordable rules, the information conveyed by these distances was scrutinized in terms of their statistical properties. In particular, the entropy and the mutual information [40,41] of the fields in the CRM forms were analyzed with intensive calculations to show that convergence in their estimation can be readily obtained. Overall, these previous findings suggested creating an X-from-M criterion to reduce the computational complexity of the search algorithm and to estimate possible neighbours to a duplicated-customer candidate, hence providing the managers with a moderate number of candidates.
The rest of the paper is organized as follows. In Section 2, we summarize some existing literature related to the duplicate identification whose fundamentals have been useful for our proposal. In Section 3, the detection system for duplicated recordings is presented, together with the required notation and the statistical related concepts for its analysis. Then, Section 4 presents the obtained results throughout all the system elements. Section 5 discusses the significance of the results and the implications of the use of big data in the current CRM and hospitality industry scenarios, together with the statement of final conclusions.

2. Background

Managing all the available information in a company and ensuring that this information is of the highest-possible quality means that the value of this company grows significantly [42]. In this setting, the problem of Entity Reconciliation (ER) is one of the fundamentals in the integration of data. As noted above, this problem is also known as deduplication, conciliation of references, purge, and others. ER is a significant and common data cleaning problem, and it consists of detecting data duplicate representations for the same external entities, and merging them into single representations [43]. This problem can be applied to many different domains, such as deduplication in databases [44], duplicate detection in x m l data or hierarchical data [45], cross-document co-reference resolution methods and tools [46], blocking techniques [43,47], bug reports [48], customer recognition [31], and E-health [49]. Most of the existing studies have been validated using real-world datasets, but very few of them have applied their proposal in a real case in the industry [42]. In the literature, the duplication detection problem has been studied from three distinct analysis paradigms, namely, ranking, binary classification, and decision-making problems. Among these three problem types, the ranking problem attracts the most attention because of its feasibility, and text mining plays a crucial role in detecting duplicates [48]. Authors in [31] used Levenshtein Edit Distance for feature selection in combination with weights based on the Inverse Document Frequency (IDF) of terms. Matching dependencies, a new class of semantic constraints for data quality and cleaning, has been shown to be profitably integrated with traditional machine learning methods for developing classification models for ER [43].
As previously discussed, CRM is an important framework for managing an organisation’s interactions with its current and future customers [50,51]. Identification of duplicated customers in the CRM is also receiving increasing attention in the literature [52,53,54,55]. When there are many customers in the CRM database with similar names, or names with similar spellings to those of the customer to be identified, recognition becomes difficult. The current literature on identity recognition focuses on searching for and matching a given identity from all the available information in an organization’s database. A clean database improves performance and subsequently leads to better customer satisfaction [31]. It has also been clearly established that good data quality allows the staff to be more productive, as instead of spending time validating and fixing data errors, they can focus on their core mission [56]. For datasets which are noisy or use different data formats, they have to be normalized by employing data transformations prior to comparison. As comparing entities by a single property is usually not enough to decide whether both entities describe the same real-world object, the linkage rule has to aggregate the similarity of multiple property comparisons using appropriate aggregation functions [57]. An algorithm is presented in that work which combines genetic programming and active learning in order to learn expressive linkage rules which can include data transformations and combine different non-linear similarity measures. The user is only required to perform the much simpler task of confirming or declining a set of link candidates which are actively selected by the learning algorithm to include link candidates that yield a high information gain. However, few of these previous studies have devoted space to analyzing the statistical properties of the data fields usually found in a CRM in order to generate statistically principled and computationally operative methods in this arena.

3. Detection System for Duplicated Recordings

A database of customer recordings in a CRM can be seen as a multidimensional data structure, usually corresponding to an SQL structured type. The CRM data processing can often become unmanageable either by the dimensions or by the nature of the data, or by both of them. In order to achieve efficient and scalable data processing, it can be advantageous for instance to adapt the data to a file system (unstructured) or to a database (structured) and using queries on them, and sometimes it even compensates to distribute those data in different nodes. In this way, a map-reduce-based solution can be fruitful in order to extract useful information from the data in an embarrassingly parallel way [58] from all the above data arrangements. In this work, we focus on the result of a query obtained from a universe database and turned into a datastore, in such a way that CRM registers here are stored as a set of features for each recorded customer. We also stress here that there will be relevant and irrelevant interdependence among customers, especially when dealing with international delegations of a hospitality company. Moreover, a good number of customers (mostly the multinational ones) will have themselves a complex structure, in terms of headquarters, delegations, and their inter-dependencies. Overall, if we wanted to consider the problem with a strict-realistic view, we would probably need a graph-based structure, a tree-based structure, or even a combination of both. This is probably one of the main reasons for the moderate attention paid to the data duplication in CRM in recent years, despite the practical relevance of this problem.

3.1. Problem Notation

If we denote the set of customers in a CRM as { C n , n = 1 , , N } , the data structure for each customer can be seen as given by a concatenation of M features, denoted as { F m , m = 1 , M } . This represents a strong simplification on the data structure, as we are only working with a single-table view of the features. Nevertheless, several of the identification fields likely include implicit information about those relationships, so that it can be subsequently exploited. On the one hand, what we gain with this problem formulation is a simplification that allows us a first approach to the problem without being obscured by the implicit and explicit complex data structures, in other words, we are turning a multidimensional-database analysis problem into a simpler feature-comparison approach. On the other hand, our database now has form fields with different types. Let us assume that each feature belongs to one type in a set of different possible data types. Three of the most usual feature types are categorical, metric, and free text (denoted here as C , M , and T , respectively), and it is also very usual that the categorical features can be expressed in terms of text chains associated to each category value. Therefore, we can define a property type for each of the previous feature types, this is,
F m . t y p e { C I , C T , M , T }
where categorical features are either indexed values ( C I ), or each category value is associated to a descriptive text string ( C T ). In general, a categorical feature F m has several possible categorical values, which is generally denoted as { V m } (set of values for the mth feature), and the possible values are denoted by { v m 1 , , v m r m } , with r m the number of possible categorical values for F m . It can be seen then that even a simple table can have a complex data structure. Finally, we denote the instance matrix as C n , m = C n m , which conveys the recorded values of all the features for all the customers in the form table.
The multidimensional statistical density distribution can be denoted as f C ( C ) , and it can be seen as the joint probability distribution for each feature, taking into account that they can have their different types, this is,
f C ( C ) = f F 1 , , F M ( F 1 , , F M )
and hence, matrix C conveys sampled instances from this distribution. Let us assume in addition that we can distinguish two regions, which are statistically modeling two groups of instances in this multidimensional distribution. We denote as the target group to that set of instances corresponding to a given same customer, and the remaining are called the non-target group. Note that the non-target group can include other customer duplicated, but they are not considered as target in this model, so the targets are considered one each time. Then, the statistical region G 0 ( G 1 ) correspond to the support of the sampled instances of non-target (target) groups, given by G 0 ( G 1 ) sets. From a practical point of view, G 1 represents that region yielding the sampled set of CRM duplicated instances which the manager wants to identify. From a theoretical point of view, the data structure presents several limitations in different aspects which make their identification hard to address. First, we do not know the labels for the instances in both groups, hence this is a non-supervised problem in Statistical Learning Theory. Second, a method for the comparison among different features is not easy to define, given their categorical and in general heterogeneous nature. Third, even the comparison among instances of the values in the same fields is twofold complex, because we do not have a clear pattern to identify duplicated entries, and we do not either have theoretical knowledge about their statistical distributions.

3.2. Distance Distributions in Heterogeneous Recordings

To overcome these challenges, we state the practical problem of CRM duplicated-customer identification as follows. We select a specific test instance C t , which is a candidate to be scrutinized for the presence of its duplicates. The managerial criterion to select this test instance can be because it represents a relevant customer, or it is just a new customer to be introduced into the CRM and to be checked in advance that it is not in it. According to the previously described statistical formulation, C t G 1 . Our operative problem is now stated as the identification of other possible instances of G 1 in the database. A condition that can be established to design an appropriate algorithm is that it needs to be more sensitive (emphasis is on not excluding true duplicated instances) even if it represents decreased specificity (a larger number of non-duplicated customers are included as duplicate candidates). This is practically acceptable because the system aims to select a moderate set of candidates which can be subsequently scrutinized by the managerial team, hence connecting the automated tasks and the domain knowledge by the expert users.
In order to tackle the inter-feature issues, we initially assume independence among features, so that their statistical joint distribution can be expressed as the product of the individual feature distributions, i.e.,
f C ( C ) f F 1 ( F 1 ) · f F 2 ( F 2 ) · · f F M ( F M )
For an operative design of the intra-feature analysis, we distinguish both groups, i.e.,
f C ( C | G k ) f F 1 ( F 1 | G k ) · f F 2 ( F 2 | G k ) · · f F M ( F M | G k )
for k = 0 , 1 , and where each conditional distribution has its own shape and parameters. Let us assume now that we can use a distance between two values of a given feature in two customers n a and n b , denoted as d ( C n a m , C n b m ) . The statistical distribution of this distance for the mth feature is a random variable, with distribution
f d m ( d m ) = 1 s E G d ( C n a m , C n b m )
where E denotes statistical averaging and S is a normalization constant to unit area. In terms of this distribution, we can say that a customer is a possible neighbor of the test customer if their distance is small.In statistical terms, this can be given by its distance belonging to the threshold distance corresponding to a low percentile α m of the distribution, this is, d ( C n 0 m , C n 1 m ) < d α m . We call d α m the threshold distance for mth feature to consider two instances as neighbors in that feature.
A remarkably useful distance suitable for strings is the Wagner-Fischer algorithm to compute the Edit Distance between two string characters. This algorithm is extremely simple, and it has a history of multiple inventions [59,60,61]. As seen in the introduction, the Edit Distance determines how dissimilar two strings are to each other by counting the minimum number of operations that are needed to transform one of them into the other. Different definitions of an Edit Distance use different string operations.One of the most popular ones is the Levenshtein distance, in which the admitted operations are the deletion, the insertion, and the substitution of a single character in the string at each step. In the Wagner-Fischer algorithm, the Edit Distance between two strings a = a 1 a 2 a A and b = b 1 b 2 b B , is given by the following d i , j recurrence:
d i 0 = k = 1 i w d e l ( b k ) , for 1 i A
d 0 j = k = 1 j w i n s ( b k ) , for 1 j B d i j = d i 1 , j 1 , for a j = b i min d i 1 , j + w d e l ( b i ) d i , j 1 + w i n s ( a j ) d i 1 , j 1 + w s u b ( a j , b i ) for a j = b i , for 1 i A , 1 j B
which is the most basic form of this algorithm and can be readily programmed [60], and where w d e l ( w i n s , w s u b ) denotes the number of character deletions (insertions, substitutions). This distance provides us with a method to characterize the neighbourhood of test customer C t to detect its possible duplicated customers among its neighbors. Note that we need to identify distributions f d m ( d m ) and to define the set of thresholds { d m } , which can be either established according to the manager experience when convenient, or with a statistically supported percentile α m . In the following we describe the basic principles subsequently used to establish the multidimensional criterion according to the statistical properties of these distributions.

3.3. Entropy and Mutual Information Analysis of Distances

The most fundamental concept of Information Theory is the entropy, which characterizes the uncertainty within a random discrete variable. Whereas it was initially introduced in the telecommunication field, it has been subsequently expanded to a wide range of discrete statistics and its applications. Based on it, the mutual information informs us about the amount of information that a random discrete variable contains about another, which is sometimes also described as the reduction in the uncertainty of one variable thanks to the knowledge of the other one [40,41].
In this work, we study the entropy within a given field of a CRM form and the mutual information among the different fields with text type content. We restrict ourselves to a unified and simplified study, in which the fields can be represented one way or another by a not-too-long text. Let f d m ( d m ) and f d n ( d n ) be the described mass density functions of the n-th and m-th variables in the CRM form. Then, we define the entropy of the m-th variable as follows
H ( m ) = d m f ( d m ) l o g 2 ( f ( d m ) )
where the use of the base-2 logarithm implies that its units are bits. Accordingly, the mutual information between the n-th and m-th variables is given by
I ( m , n ) = d m , d n f d m , d n ( d m , d n ) log 2 f d m , d n ( d m , d n ) f ( d m ) f ( d n ) f d m ( d m ) f d n ( d n ) = E d m , d n f d m , d n ) f d m f d n
Therefore, in our problem entropy represents the amount of information of each of the variables in terms of the mass density of the Edit Distances among its instances. On the other hand, the mutual information represents the amount of information shared between two variables in terms of the same Edit Distance. Both represent a basic but fundamental statistical description to begin to analyze the distributions of distances in these abstract spaces of text strings, and they will allow us to scrutinize a principled way to adjust thresholds in the heuristic approaches to our problem.

3.4. Dealing Joint Distributions with X from M

The remaining issue to deal with is the inter-feature dependencies, which is really complex. For this purpose, and supported by the experiments related with the previous section, we propose here implementing the criterion so-called X from M. This means that, from the M available features, we consider that an instance is a global neighbour of C t if at least X features are at a distance below their threshold, independently of which features they are. This represents a robust yet simple criterion, which aims to give a tool to drive the search with flexibility of the origin of the duplication in the database.

3.5. Map-Reduce Implementation

Even with the above described simplifications on the data structure and on the statistical distributions, the amount of data to deal with in a CRM (several hundred thousands or often millions of entries) can be a strong limitation for any algorithm. In this scenario, as in many others, big-data techniques can be useful to solve serious and practical problems such as disk memory problems, random-access-memory problems or computational cost problems, making the problem unmanageable. These problems are mainly caused by the computational cost of the used algorithms and by the size of the data. Depending on the limitations of the problem, the most appropriate option among a variety of existing big-data techniques can be chosen [58]. One of the most popular methods is map-reduce, which allows division of the data and process them among different nodes and calculate the solution in an embarrassingly parallel way. Map-reduce divides the problem into different chunks, which could already be divided if the database or file system was previously distributed, and dispensed in different nodes (workers). These workers apply a function called map (mappers), whose operation must be commutative and associative and the obtained values are grouped according to the different existing required keys. This is known as a key-value paradigm. Once the values have been grouped by key in each node, all of these lists of values of a certain key can be reduced to a single value by means of a function called reduce. This function is executed by new assigned workers (reducers) who receive as input all of these lists of values assigned to a particular key.
In our work, map-reduce implementations were used basically to parallelize those tasks involved in running neighbor comparisons throughout the large database, which mostly involved the neighborhood loops, and the experimental estimations of mass density functions and probability density functions. Nevertheless, given their repetitive nature in the neighborhood comparison processes, it is evident that big-data technologies can be especially useful in order to support the proposed application, even with the use of basic tools available in them.

4. Experiments and Results

4.1. Database Description

Our method was built while using it on a real data problem, and a database was assembled from the CRM universe of a large hospitality company. It consisted of more than 800,000 recordings in a recovered table, in such a way that each customer was included in a row, and M = 18 features were included in text form, either categorical, integers, or text strings. In other words, the feature formats were of such a different nature as numerical integer identifiers, nominal identifiers (names), segmentation-categorical fields, localization fields, and binary values. These recordings corresponded to the hospitality-company data recorded during years 2016 and 2017.
We scrutinized and quantified the different theoretical elements and simplifications that have been described in Section 3. Note that those features containing names (nominal identifiers) sometimes can be optional and free-text input, such as names of people who are responsible for a given environment. Other kinds of variables with different a nature are those features related to segmentation or geographical localization, which can represent very different aspects, but they are often categorical features with a predetermined number of categories, and this number is often reduced compared to the number of recordings. This represents a different situation compared to nominal variables with free text, whose number of classes could be comparable to the number of recordings. Nevertheless, the same distance can be used for them with operative advantage, as shown in the following experiments.

4.2. Intra-feature Distance Distributions

We started by analyzing the distribution of the intra-feature distances for each of them individually. For this purpose, a customer test C t was randomly selected and compared with a subset of other 100,000 customers, { C n , n = 1 , , 100,000}, which was repeated for 100 independent realizations. Our comparison consisted of calculating distance d ( C t m , C n m ) for each m = 1 , , M feature, with n = 1 , , 100,000, i.e., we obtained the distances from the test customer to all the other ones for every feature. This allowed us to build an estimation of the empirical statistical distribution in terms of their histograms.
Figure 1 shows the averaged estimated distributions for these intra-feature distances for each mth feature. For visualization purposes, distributions f d m are represented as normalized to their maximum value. The shaded area represents their 95 % confidence interval. It is noticeable that this confidence interval is very narrow in practically all the variables, which is due to the facts that distributions are strongly discretized, and that a large number of examples is used for their construction. Hence, even the peaks in the fragmented distributions are not attributed to noise, but rather consistent among realizations. Note also that the distributions have been sorted in terms of descending position of their absolute maximum, and this in turn makes them group together in reference to their nature. Hence, the most widespread distributions correspond to mandatory names, which are free-text string characters. These distributions have a region of short distances (about less than 5), a single-modal distance and a tail (which is not shown as the trimmed histogram has been represented for visualization purposes). Then, features related with optional names exhibit narrower distribution mass, some tails, and short distances being around less than 10.
A similar pattern is followed then by numerical identifiers and location/segmentation variables, whose profile is in general narrow, with tails and with lower short distance. Obviously, binary features have the shortest distributions. These results point out that it seems feasible to use the same distance with all of these features, even despite their different nature, because neighborhood can be defined with a similar criterion. On the other hand, the neighborhood distance is to be defined for each feature, but it seems to be consistent with features of the same kind, which can guide their assignment, either from an automated or from a supervised approach.

4.3. Tails of the Generalized Edit Distance

We subsequently analyzed the dependence of the Edit Distance with the percentile. Note that in the problem of duplicated customer identification, as explained before, distances larger than 5 or 10 characters, depending on the feature, could yield completely different strings, independently of their length.
Figure 2 shows the representation and detail for lower percentiles of this relationship in the performed experiments. In order to determine the distance of interest, percentiles lower than 1 % should be used in this problem and in this database. The lower panel shows the distance for α m < 1 for each of the m features, including averaged value and the 95% confidence interval. These representation show a strong step-like aspect, with wide confidence intervals in the transitions, which suggests that the use of α m percentile as a threshold could turn unstable and even inefficient, because a range of percentiles will result in the same neighborhood distance. Therefore, we propose to use the distance value d t m as threshold, in number of characters or simple operations, as the threshold value to be fixed by managers using the system.
Nevertheless, and after these observations, further analysis of the text vector spaces induced by the Edit Distance can be done in terms of the asymptotic properties of the entropy and the mutual information of the variables, which in turn can be useful to design principled data quality algorithms in CRM environments, as introduced next.

4.4. Entropy Analysis of Text Features in Edit-Distance Induced Spaces

To characterize the fundamental properties of the text spaces induced by the Edit Distance in data quality applications in CRM, the following set of experiments was carried out. First, for a given length of available data (N), we randomly selected data from the total CRM, and the mass density function of each variable used was estimated within this subset. This experiment was repeated 100 times by taking random samples from the database with also randomized initialization, and the 95% CI of said estimation was calculated. In addition, the asymptotic behavior of the CI was studied by increasing the value of N. Secondly, considering the same data with which the estimates of the mass densities were calculated, the entropy of each characteristic and of the mutual information of all pairs was estimated. In the same way, its asymptotic evolution with N was also analyzed.
Figure 3 shows representative examples of these results. In particular, the figure represents in each panel (a,b,c) the estimated density mass function and their corresponding asymptotic CI for examples of features with different nature as a function of the distance for each of them, and calculated of a different set of compared samples N. Note that the densities are amplitude normalized for better visualization. The (non-amplitude-normalized) density distribution of the estimated entropies are represented in (d,e,f) for three different examples of representative types of features, and their asymptotic evolution can be observed when increasing N, which in general can be appreciated to be mostly stabilized for N > 1000 . Also the figure represents the (non-amplitude-normalized) density distributions of the estimated mutual information and asymptotic evolution for representative examples of pairs of features (g,h,i).
It can also be seen in that figure that the mass densities of each characteristic have different profiles, and they are strongly multimodal, see panels (a,b,c). In all of them there is a gap between what we could describe as being very close and being further away. However, the length of this gap varies in different variables, as it does the width of the single mode or several modes corresponding to separated distances. It is of great practical interest to see that mass densities are estimated accurately for each variable without the need to asymptotically rise to extremely high values of N, which supports the feasibility of estimating these densities if needed by probabilistic methods.
On the other hand, and as a result of the different profiles of the mass densities, the entropy of the different variables also turns out to be sometimes multimodal, though not always, see panels (d,e,f)). Note that the entropy values are moderate (of the order of magnitude of some few bits, and less than one bit for the special case of categorical variables). Regarding mutual information, some bimodalities can be seen in panels (g,h,i), though they are not excessively relevant because in any case they have low values, usually less than a bit. This means that the information provided by each of these random variables about the others is low, therefore, this indicates a low reduction of uncertainty and low dependence among them. Again, the estimation consistency is in general quickly reached for the entropy and for the mutual information, meaning that more advanced probabilistic methods could be used in this scenario.

4.5. Distance Increase with Multidimensional Spaces

Another relevant aspect which needs to be quantitatively analyzed is how restrictive can a duplicated search be, as a function of the number of features that have to be below their fixed threshold d t m . Based on the results of the previous experiments, we propose the following approach. If we focus only on some more apparently attractive-to-use features we could be missing profiles from features a priori less interesting, but which could be revealed as relevant by an automatic search. Therefore, in order to characterize and measure the constrained character of the search, we measured the number of neighbors of C t obtained by screening the number of thresholds to be fulfilled, in the so-called X from M criterion, if the distance threshold is fixed d m between 1 and 5 characters. This is consistent with the gap asymptotically determined in Figure 3. From a practical point of view, we restricted ourselves to search up to 5 characters because a larger number was observed to give an extremely large number of candidates to be post-analyzed by an expert.
Accordingly to the previous considerations, Figure 4 shows the results of calculating the averaged neighborhood of a specific customer C t , and repeating the experiment randomly for 100 independent realizations in the same available database. The vertical axis shows the mean size of the neighborhood as a joint function of the fixed threshold in distance d α m and the number of features required to fulfill its corresponding threshold, according to the previous setting of the choice of the value for X from M criterion. Note that any possible combination of X features has been scrutinized in this experiment, and in our database we used M = 18 features. As shown in the figure, and for the M = 18 considered features, if we try to allow more than 50 % of the features in X from M (i.e., X > 9 ) for threshold distances d t m < 5 it turns out to be too restrictive, and we do not get any candidate. On the other hand, for X = 3 from 18 we would get too many candidates. The graph represents the joint evolution of the distance and of the number of neighbor features in the scheme, and their inter-dependence. Note also that fixing the distance threshold to d t m = 1 , very few candidates are obtained, which is a case to avoid due to its extreme probability of loss. This allows us to establish both free parameters so that a very well defined search area is defined for them.

4.6. Case Study for Real-World Usefulness

The improvement of the quality of a CRM, and more specifically, the cleaning of duplicates, is a task that is usually carried out not only through distance-among-words criteria but also through business criteria. To our best understanding, and as explained so far in the present work, this is the reason why the application of techniques based on distances among words can be very useful for this purpose, as it can reduce the number of candidates to evaluate with business criteria at a later stage.
In order to further evaluate this usefulness in a simple yet real-case scenario, we applied the technique proposed in this paper on the data of a CRM from a big and international company. Specifically, it was been applied on 3753 duplicates in which the original assignment was known among a total of 363,960 data and in which the possibilities of detection were inside the scope of the techniques proposed here. Following the knowledge offered by the business area, it was applied to 5 of the most critical variables for the detection of duplicates (2 mandatory names, 1 optional name, 1 segmentation variable, and 1 binary variable). Given that it is interesting to offer the lowest number of candidates to a duplicate so that the business criteria can do their refining work, Table 1 shows both the average number of candidates that this technique is capable of saving (i.e., specificity, selectivity, or true negative rate, T N R ) and the percentage of times that the original entry to which the duplicate should be assigned has been lost (i.e., missing rate or false negative rate, F N R = 1 T P R , being T P R the true positive rate or sensitivity). These measures are given therein for different values of the maximum distance threshold to be set ( d α m { 1 , 5 } ) and the number of X from M variables (being M = 5 ) that would have to meet at least that maximum distance value. Figure 5 shows the trade-off between the missing rate or 100-Sensitivity, and between the number of candidates correctly discarded or Specificity (both in % in the figure).

5. Discussion and Conclusions

Several studies [62,63,64] have underlined the relevance of managing customer data at a high quality level.Experian revealed that 66 % of companies lack a coherent, centralized approach to data quality [33]. Loose quality data such as duplicates or missing data, inaccurate or outdated data, could lead managers and analysts to develop an incorrect image of customers and their purchase preferences [65]. This is specially more patent in multinational companies that operate through more channels than in those who operate in a single country. Moreover, the information collected across various channels is frequently exposed to human error as consumers and individual employees enter information manually, e.g., receptionist at the front office. Some authors point out the lack of vision of the organization to share, maintain, use, and improve customer information [66,67,68]. Despite its relevance [69], there is no recognized and efficient data management model to handle big-data (large volume, complex, growing datasets with multiple autonomous sources) since the traditional data models are not able to manage its increasing complexity [70].
In this scenario, our proposal is a big-data system to address the problem statement of duplicated customer identification as a large-scale data analysis in a multinational hotel chain. Duplicated records slow down client’s search engines, issues of the invoice configuration, rules of validation of fields, among others, which suppose that the time lapses of reservation management and invoice issuance (at checkout) substantially increase. In many cases, there are duplicated records but the CRM does not detect them. When detected, asking for these data from the customer could overpass the Data Protection Laws. In addition, the identification rules may be different depending on the country. Therefore, the fundamental problem is not the elimination of duplicates itself but their detection. This is a common problem in the hotel industry and it is worthwhile for them to invest in research in order to identify these duplicates, hence gaining in data quality, information homogeneity, improvement in the measurement of bonus of the commercial area (which is critical), and reduction of errors in the application of tariffs, to name only some few advantages. Even with validated data and improved searching, duplicates will inevitably be created due to the nature of human error. Databases should be checked at regular intervals to ensure that no duplicate accounts have been created and to consolidate their conveyed information. Our approach has consisted of first scrutinizing this specific data quality problem, then providing a statistical problem statement, and finally using statistical simplifications on the data structure to be handled. Even so, the number of operations is large, as far as a CRM in a big company will typically include hundreds of thousands of recordings, hence big-data simple solutions (map-reduce or data stores) can be applied when and for what they are actually needed.
Our method has implemented several simple, yet powerful elements. First, a generic feature representation for the customer fields is given by a table data structure, which can be handled as a heterogeneous data matrix, which has simplified the issue of data heterogeneity. Second, homogeneity in the comparison procedure has been done by working with the different data types in terms of the same distance, namely the Edit Distance, which has been shown in the experiments to be operative enough. The differences in the intrinsic neighbourhoods of potentially duplicated recordings tend to be similar among features of the same nature, and somehow different with respect to features with somehow different nature. Nevertheless, the characteristic distances for all of them are always small (we could say less than 10, often much less), in terms of the statistical distribution tail. Third, the use of big-data techniques has been tailored to the needs of this search, hence, the advantage of big data does not come from complex algorithms being parallelized for a given task, but rather it comes from the aggregation effect of a large number of recordings, as shown by the smooth confidence intervals obtained for the estimated distance distributions in our experiments. The estimation consistency was in general quickly reached for the mass density functions of the variables, for their entropy and for their mutual information, meaning that more advanced probabilistic methods could be used in this scenario. Also, the mutual information among variables was extremely moderate, though not null, which needs to be taken into account to further explore statistically principled methods in this scenario. Finally, the previous over-simplification could be expected to preclude the system to efficiently find diverse neighbours, for only working with close individual features. However, it has been shown how the use of a X from M simple strategy brings back to the algorithm the power and flexibility of searching for neighbours which could be at unexpected features.
It is very complex to create a completely automatic system, so that a reasonable choice in this kind of scenarios is to create a semi-supervised approach. On the one hand, this strategy can be checked in the literature to be a suitable approach to big data problems in emerging areas, in which data quality is revealed in those early stages to have an even more-than-suspected potential high impact on the company [71,72,73,74,75,76,77,78,79]. On the other hand, the Information Technology areas or similar ones often have staff partially devoted and responsible for the data quality aspects. Nevertheless, we only addressed one single aspect of data quality, which was the customer duplication in a multinational hotel chain CRM, considering that duplicated data can be situated among the top three data quality errors for 30% of organizations [33]. Although the usefulness of high data quality is desirable and clear, in the case of very large databases, this objective may be in conflict with high associated costs. A cost-benefit model [80] confirmed the assumption that the optimal level of data quality, in terms of maximizing net benefits, is not necessarily the highest technically possible. Nevertheless, the study by [18] showed that there is a clear similarity between the data quality factors that affect small and medium enterprises and large organizations from the point of view of CRM implementation. This desirable implantation often shocks [24] with the absence of planning in the stages of data cleaning, normalization, integration or migration, hence data governance has been proposed, which consists of defining a set of policies and procedures to supervise data use and management, and their turning into a strategic active.
Our method is being used in this setting, and the design has been done to provide the manager with increased sensitivity at the expense of moderately reduced specificity when identifying candidates to be duplicated. However, our experiments have shown that the workload for an expert supervisor after our system output is affordable. There are two scenarios that can be established for the proposed system. On the one hand, the manager can be willing to identify the possible duplicated recordings of a given customer. This can be due to this customer being strategically or economically relevant, or just the inclusion of the new customers of every month in the system. Here, a data store solution is enough to perform the analysis and to provide an operative candidate list. On the other hand, a large-scale exploration of the database can be required, for instance, for migration purposes. In this case, the use of more computationally intensive implementations (such as map-reduce, but also any improved state-of-the-art platform) can help to overcome the algorithmic load, while still being operative and usable.
We can conclude that the use of big-data technology can provide hospitality management, and hence many other sectors, with a tool for data quality control and improvement of their CRM systems. Specifically, duplicated customers can be identified, with high sensitivity and operative specificity, with the help of a particularly designed system. The use of simple algorithms and statistical models is compensated by the strength of using large numbers of recordings, which has been shown to provide stable statistical descriptions. Further open issues of big-data and data quality in CRM can be addressed starting from the results obtained here, and their use is expected to grow naturally when the specific challenges are identified through the cooperation between academia and the hospitality industry in this area.

Author Contributions

L.G.-S. and P.T.-B. conceptualized the problem, elaborated the state of art, and developed the discussion and conclusions. S.M.-R. and J.L.R.-Á. conducted the experiments. C.S.-R., L.G.-S. and P.T.-B. supported the experimental design. All authors discussed the results and contributed to write the final manuscript.

Funding

This work has been partly supported by the Institute of Health Carlos III, Spain (grant DTS 17/00158), and by the Spanish Projects FINALE (TEC2016-75161-C2-1-4), TEC2016-75361-R, and ECO2016-75379-R from Spanish Government.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krishna, G.J.; Ravi, V. Evolutionary computing applied to customer relationship management: A survey. Eng. Appl. Artif. Intell. 2016, 56, 30–59. [Google Scholar] [CrossRef]
  2. Kumar, V.; George, M. Measuring and maximizing customer equity: A critical analysis. J. Acad. Mark. Sci. 2007, 35, 157–171. [Google Scholar] [CrossRef]
  3. Ramani, G.; Kumar, V. Interaction orientation and firm performance. J. Mark. 2008, 72, 27–45. [Google Scholar] [CrossRef]
  4. Keramati, A.; Mehrabi, H.; Mojir, N. A process-oriented perspective on customer relationship management and organizational performance: An empirical investigation. Ind. Mark. Manag. 2010, 39, 1170–1185. [Google Scholar] [CrossRef]
  5. Kim, B.C.; Choi, J.P. Customer information sharing: Strategic incentives and new implications. J. Econ. Manag. Strategy 2010, 19, 403–433. [Google Scholar] [CrossRef]
  6. Sigala, M. Integrating customer relationship management in hotel operations: Managerial and operational implications. Int. J. Hosp. Manag. 2005, 24, 391–413. [Google Scholar] [CrossRef]
  7. Wu, L.W. Satisfaction, inertia, and customer loyalty in the varying levels of the zone of tolerance and alternative attractiveness. J. Serv. Mark. 2011, 25, 310–322. [Google Scholar] [CrossRef]
  8. Kasim, A.; Minai, B. Linking CRM strategy, customer performance measures and performance in the hotel industry. Int. J. Econ. Manag. 2009, 3, 297–316. [Google Scholar]
  9. Chadha, A. Case Study of Hotel Taj in the Context of CRM and Customer Retention. Kuwait Chapter Arab. J. Bus. Manag. Rev. 2015, 4, 1–8. [Google Scholar] [CrossRef]
  10. Dev, C.S.; Olsen, M.D. Marketing challenges for the next decade. Cornell Hotel Restaur. Adm. Q. 2000, 41, 41–47. [Google Scholar] [CrossRef]
  11. Kotler, P. When to use CRM and When to forget it. Paper Presented at the Academy of Marketing Science, Sanibel Harbour Resort and Spa, Fort Myers, FL, USA, 30 May 2002. [Google Scholar]
  12. Lin, Y.; Su, H.Y. Strategic analysis of customer relationship management-a field study on hotel enterprises. Total Qual. Manag. Bus. Excell. 2003, 14, 715–731. [Google Scholar] [CrossRef]
  13. Nasution, H.N.; Mavondo, F.T. Organisational capabilities: Antecedents and implications for customer value. Eur. J. Mark. 2008, 42, 477–501. [Google Scholar] [CrossRef]
  14. Nguyen, T.H.; Sherif, J.S.; Newby, M. Strategies for successful CRM implementation. Inf. Manag. Comput. Secur. 2007, 15, 102–115. [Google Scholar] [CrossRef]
  15. Padilla-Meléndez, A.; Garrido-Moreno, A. Customer relationship management in hotels: Examining critical success factors. Curr. Issues Tour. 2014, 17, 387–396. [Google Scholar] [CrossRef]
  16. Reimann, M.; Schilke, O.; Thomas, J.S. Customer relationship management and firm performance: The mediating role of business strategy. J. Acad. Mark. Sci. 2010, 38, 326–346. [Google Scholar] [CrossRef]
  17. Beg, J.; Hussain, S. Data Quality—A Problem and An Approach; White paper; Wipro Technologies: Bangalore, India, 2003. [Google Scholar]
  18. Alshawi, S.; Missi, F.; Irani, Z. Organisational, technical and data quality factors in CRM adoption-SMEs perspective. Ind. Mark. Manag. 2011, 40, 376–383. [Google Scholar] [CrossRef]
  19. Moore, C. How to Create a Business Case for Data Quality Improvement. Available online: http://www.gartner.com/smarterwithgartner/howto-create-a-business-case-for-data-quality-improvement/ (accessed on 4 April 2019).
  20. Turban, E.; Leidner, D.; McLean, E.; Wetherbe, J. Information Technology for Management; John Wiley & Sons: Hoboken, NJ, USA, 2008. [Google Scholar]
  21. Soltani, Z.; Navimipour, N.J. Customer relationship management mechanisms: A systematic review of the state of the art literature and recommendations for future research. Comput. Hum. Behav. 2016, 61, 667–688. [Google Scholar] [CrossRef]
  22. Akoka1a, J.; Berti-Equille, L.; Boucelma, O.; Bouzeghoub, M.; Comyn-Wattiau, I.; Cosquer, M.; Goasdoué-Thion, V.; Kedad, Z.; Nugier, S.; Peralta, V.; et al. A framework for quality evaluation in data integration systems. In Proceedings of the 9th International Conference on Entreprise Information Systems, Madeira, Portugal, 12–16 June 2007; p. 10. [Google Scholar]
  23. Thompson, E.; Sarner, A. Key Issues for CRM Strategy and Implementations; Technical Report; Gartner Research: Stamford, CT, USA, 2009. [Google Scholar]
  24. Alonso, Ó.; Delgado, A.; Pedrosa, P. Las Soluciones CRM en España; Technical Report; Penteo, ESADE Business School: Madrid, Spain, 2008. [Google Scholar]
  25. Eckerson, W.W. Data Quality and Bottom Line: Achieving Business Success through High Quality Data (TDWI Report Series); The Data Warehousing Institute: Seattle, WA, USA, 2002. [Google Scholar]
  26. Missi, F.; Alshawi, S.; Fitzgerald, G. Why CRM efforts fail? A study of the impact of data quality and data integration. In Proceedings of the 38th Annual Hawaii International Conference on System Sciences, Big Island, HI, USA, 3–6 January 2005; p. 216c. [Google Scholar]
  27. Xu, H.; Horn Nord, J.; Brown, N.; Daryl Nord, G. Data quality issues in implementing an ERP. Ind. Manag. Data Syst. 2002, 102, 47–58. [Google Scholar] [CrossRef]
  28. Moss, L.; Abai, M.; Adelman, S. How to improve data quality. In Data Strategy; Addison-Wesley Professional: Boston, MA, USA, 2005. [Google Scholar]
  29. Goga, O. Matching User Accounts Across Online Social Networks: Methods and Spplications. Ph.D. Thesis, LIP6-Laboratoire d’Informatique de Paris 6, Paris, France, 2014. [Google Scholar]
  30. Elmagarmid, A.K.; Ipeirotis, P.G.; Verykios, V.S. Duplicate record detection: A survey. IEEE Trans. Knowl. Data Eng. 2007, 19, 1–16. [Google Scholar] [CrossRef]
  31. Saberi, M.; Theobald, M.; Hussain, O.K.; Chang, E.; Hussain, F.K. Interactive feature selection for efficient customer recognition in contact centers: Dealing with common names. Expert Syst. Appl. 2018, 113, 356–376. [Google Scholar] [CrossRef]
  32. Helander, D. Solving the Hotel Data Management Problem in 3 Steps-Revinate. Available online: https://www.revinate.com/es/blog/solving-hotel-data-management-problem-3-steps/ (accessed on 12 February 2019).
  33. Schutz, T. The State of Data Quality. An Experian Data Quality White Paper. Available online: https://www.experian.com/assets/decision-analytics/white-papers/the%20state%20of%20data%20quality.pdf (accessed on 1 April 2019).
  34. Pinto, F.; Santos, M.F.; Cortez, P.; Quintela, H. Data pre-processing for database marketing. In Data Gadgets; Workshop: Malaga, Spain, 2004; pp. 76–84. [Google Scholar]
  35. Yujian, L.; Bo, L. A normalized Levenshtein distance metric. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1091–1095. [Google Scholar] [CrossRef] [PubMed]
  36. Waterman, M.S.; Smith, T.F.; Beyer, W.A. Some biological sequence metrics. Adv. Math. 1976, 20, 367–387. [Google Scholar] [CrossRef]
  37. Smith, T.F.; Waterman, M.S. Comparison of biosequences. Adv. Appl. Math. 1981, 2, 482–489. [Google Scholar] [CrossRef]
  38. Jaro, M.A. Advances in record linkage-methodoly as applied to matching the 1985 census of Tampa, Florida. J. Am. Stat. Assoc. 1989, 84, 414–420. [Google Scholar] [CrossRef]
  39. Bernstein, P.A.; Haas, L.M. Information integration in the enterprise. Commun. ACM 2008, 51, 72–79. [Google Scholar] [CrossRef]
  40. Villaverde, A.F.; Ross, J.; Moran, F.; Banga, J.R. MIDER: Network inference with mutual information distance and entropy reduction. PLoS ONE 2014, 9, e96732. [Google Scholar] [CrossRef] [PubMed]
  41. Macedo, F.; Oliveira, M.R.; Pacheco, A.; Valadas, R. Theoretical foundations of forward feature selection methods based on mutual information. Neurocomputing 2019, 325, 67–89. [Google Scholar] [CrossRef]
  42. Enríquez, J.G.; Domínguez-Mayo, F.J.; Escalona, M.; Ross, M.; Staples, G. Entity reconciliation in big data sources: A systematic mapping study. Expert Syst. Appl. 2017, 80, 14–27. [Google Scholar] [CrossRef]
  43. Bahmani, Z.; Bertossi, L.; Vasiloglou, N. ERBlox: Combining matching dependencies with machine learning for entity resolution. Int. J. Approx. Reason. 2017, 83, 118–141. [Google Scholar] [CrossRef]
  44. Maddodi, S.; Attigeri, G.V.; Karunakar, A. Data deduplication techniques and analysis. In Proceedings of the Third International Conference on Emerging Trends in Engineering and Technology, Goa, India, 19–21 November 2010; pp. 664–668. [Google Scholar]
  45. Gaikwad, S.; Bogiri, N. A survey analysis on duplicate detection in hierarchical data. In Proceedings of the 2015 International Conference on Pervasive Computing (ICPC), Pune, India, 8–10 January 2015; pp. 1–6. [Google Scholar]
  46. Beheshti, S.M.R.; Benatallah, B.; Venugopal, S.; Ryu, S.H.; Motahari-Nezhad, H.R.; Wang, W. A systematic review and comparative analysis of cross-document coreference resolution methods and tools. Computing 2017, 99, 313–349. [Google Scholar] [CrossRef]
  47. Papadakis, G.; Svirsky, J.; Gal, A.; Palpanas, T. Comparative analysis of approximate blocking techniques for entity resolution. Proc. VLDB Endow. 2016, 9, 684–695. [Google Scholar] [CrossRef]
  48. Lin, M.J.; Yang, C.Z.; Lee, C.Y.; Chen, C.C. Enhancements for duplication detection in bug reports with manifold correlation features. J. Syst. Softw. 2016, 121, 223–233. [Google Scholar] [CrossRef]
  49. Daniel, C.; Serre, P.; Orlova, N.; Bréant, S.; Paris, N.; Griffon, N. Initializing a hospital-wide data quality program. The AP-HP experience. Comput. Methods Prog. Biomed. 2018. [Google Scholar] [CrossRef] [PubMed]
  50. Faed, A. An Intelligent Customer Complaint Management System with Application to the Transport and Logistics Industry; Springer Science & Business Media: Cham, Switzerland, 2013. [Google Scholar]
  51. Lykourentzou, I.; Giannoukos, I.; Nikolopoulos, V.; Mpardis, G.; Loumos, V. Dropout prediction in e-learning courses through the combination of machine learning techniques. Comput. Educ. 2009, 53, 950–965. [Google Scholar] [CrossRef]
  52. Chandran, K.; Veeraraghavan, K.; Tb, A. Inquire management for hospital websystem using SaaS. Int. J. Adv. Res. Comput. Sci. 2016, 7. [Google Scholar] [CrossRef]
  53. Farhan, M.S.; Abed, A.H.; Ellatif, M.A. A systematic review for the determination and classification of the CRM critical success factors supporting with their metrics. Future Comput. Inform. J. 2018, 3, 398–416. [Google Scholar] [CrossRef]
  54. Reid, A.; Catterall, M. Hidden data quality problems in CRM implementation. In Marketing, Technology and Customer Commitment in the New Economy; Springer: Berlin, Germany, 2015; pp. 184–189. [Google Scholar]
  55. Anshari, M.; Almunawar, M.N.; Lim, S.A.; Al-mudimigh, A. Customer Relationship Management and Big Data Enabled: Personalization & Customization of Services. Appl. Comput. Inform. 2018. [Google Scholar] [CrossRef]
  56. Maguire, E. The Data Differentiator. How Improving Data Quality Improves Business. Available online: https://www.forbes.com/forbes-insights/our-work/data-differentiator-report/ (accessed on 8 February 2019).
  57. Isele, R.; Bizer, C. Active learning of expressive linkage rules using genetic programming. Web Semant. 2013, 23, 2–15. [Google Scholar] [CrossRef]
  58. Dean, J.; Ghemawat, S. MapReduce: Simplified Data Processing on Large Clusters. Commun. ACM 2008, 51, 107–113. [Google Scholar] [CrossRef]
  59. Levenshtein, V. Binary Codes Capable of Correcting Deletions, Insertions and Reversals. Sov. Phys. Doklady 1966, 10, 707. [Google Scholar]
  60. Navarro, G. A Guided Tour to Approximate String Matching. ACM Comput. Surv. 2001, 33, 31–88. [Google Scholar] [CrossRef]
  61. Wagner, R.A.; Fischer, M.J. The String-to-String Correction Problem. J. ACM 1974, 21, 168–173. [Google Scholar] [CrossRef]
  62. Courtheoux, R.J. Marketing data analysis and data quality management. J. Target. Meas. Anal. Mark. 2003, 11, 299–313. [Google Scholar] [CrossRef]
  63. Foss, B.; Henderson, I.; Johnson, P.; Murray, D.; Stone, M. Managing the quality and completeness of customer data. J. Database Mark. Cust. Strategy Manag. 2002, 10, 139–158. [Google Scholar] [CrossRef]
  64. Khalil, O.E.; Harcar, T.D. Relationship marketing and data quality management. SAM Adv. Manag. J. 1999, 64, 26–33. [Google Scholar]
  65. Talón-Ballestero, P.; González-Serrano, L.; Soguero-Ruiz, C.; Muñoz-Romero, S.; Rojo-Álvarez, J.L. Using big data from Customer Relationship Management information systems to determine the client profile in the hotel sector. Tour. Manag. 2018, 68, 187–197. [Google Scholar] [CrossRef]
  66. Rust, R.T.; Moorman, C.; Bhalla, G. Rethinking marketing. Harv. Bus. Rev. 2010, 88, 94–101. [Google Scholar]
  67. Seddon, P.B.; Calvert, C.; Yang, S. A multi-project model of key factors affecting organizational benefits from enterprise systems. MIS Q. 2010, 34, 305–328. [Google Scholar] [CrossRef]
  68. Zahay, D.; Griffin, A.; Fredericks, E. Sources, uses, and forms of data in the new product development process. Ind. Mark. Manag. 2004, 33, 657–666. [Google Scholar] [CrossRef]
  69. Aloini, D.; Dulmin, R.; Mininno, V.; Zerbino, P. Big Data: A proposal for enabling factors in Customer Relationship Management. In Proceedings of the 11th International Forum on Knowledge Asset Dynamics, Dresden, Germany, 15–17 June 2016; pp. 1858–1874. [Google Scholar]
  70. Wu, X.; Zhu, X.; Wu, G.Q.; Ding, W. Data mining with big data. IEEE Trans. Knowl. Data Eng. 2014, 26, 97–107. [Google Scholar]
  71. Hussain, A.; Cambria, E. Semi-supervised learning for big social data analysis. Neurocomputing 2018, 275, 1662–1673. [Google Scholar] [CrossRef]
  72. Huh, J.H. Big data analysis for personalized health activities: Machine learning processing for automatic keyword extraction approach. Symmetry 2018, 10, 93. [Google Scholar] [CrossRef]
  73. Oliver, A.; Odena, A.; Raffel, C.A.; Cubuk, E.D.; Goodfellow, I. Realistic evaluation of deep semi-supervised learning algorithms. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montréal, QC, Canada, 2–8 December 2018; pp. 3235–3246. [Google Scholar]
  74. Ross, T.; Zimmerer, D.; Vemuri, A.; Isensee, F.; Wiesenfarth, M.; Bodenstedt, S.; Both, F.; Kessler, P.; Wagner, M.; Müller, B.; et al. Exploiting the potential of unlabeled endoscopic video data with self-supervised learning. Int. J. Comput. Assisted Radiol. Surg. 2018, 13, 925–933. [Google Scholar] [CrossRef] [PubMed]
  75. Zhang, Q.; Sun, J.; Zhong, G.; Dong, J. Random multi-graphs: A semi-supervised learning framework for classification of high dimensional data. Image Vision Comput. 2017, 60, 30–37. [Google Scholar] [CrossRef]
  76. Charalampakis, B.; Spathis, D.; Kouslis, E.; Kermanidis, K. A comparison between semi-supervised and supervised text mining techniques on detecting irony in greek political tweets. Eng. Appl. Artif. Intell. 2016, 51, 50–57. [Google Scholar] [CrossRef]
  77. Jin, J.; Liu, Y.; Ji, P.; Liu, H. Understanding big consumer opinion data for market-driven product design. Int. J. Prod. Res. 2016, 54, 3019–3041. [Google Scholar] [CrossRef]
  78. Parihar, L.S.; Tiwari, A. Survey on intrusion detection using data mining methods. Int. J. Sci. Adv. Res. Technol. 2016, 3, 342–347. [Google Scholar]
  79. De Bruijne, M. Machine Learning Approaches in Medical Image Analysis: From Detection to Diagnosis. Med. Image Anal. 2016, 33, 94–97. [Google Scholar] [CrossRef]
  80. Even, A.; Shankaranarayanan, G.; Berger, P.D. Evaluating a model for cost-effective data quality management in a real-world CRM setting. Decis. Support Syst. 2010, 50, 152–163. [Google Scholar] [CrossRef]
Figure 1. Normalized f d m ( d m ) for the different features in the experimental database. Features have been sorted in descending order of the position of the maximum of each distribution, for visualization purposes, which is strongly consistent with their nature, as seen.
Figure 1. Normalized f d m ( d m ) for the different features in the experimental database. Features have been sorted in descending order of the position of the maximum of each distribution, for visualization purposes, which is strongly consistent with their nature, as seen.
Entropy 21 00419 g001
Figure 2. Distance in number of characters obtained for d α m when screening percentile α m in the statistical distributions of each of the m features. Panels show mean and 95% confidence interval (shaded in gray) for the 100 independent realizations. Panel down is a zoom for the range α m < 1 , given that this region turns to be the most interesting one for neighborhood purposes of distance d α m .
Figure 2. Distance in number of characters obtained for d α m when screening percentile α m in the statistical distributions of each of the m features. Panels show mean and 95% confidence interval (shaded in gray) for the 100 independent realizations. Panel down is a zoom for the range α m < 1 , given that this region turns to be the most interesting one for neighborhood purposes of distance d α m .
Entropy 21 00419 g002
Figure 3. Density mass function estimations and their asymptotic CI for examples of features with different nature (ac). Estimated entropies and asymptotic evolution for representative example features (df). Estimated mutual information and asymptotic evolution for representative examples of pairs of features (gi).
Figure 3. Density mass function estimations and their asymptotic CI for examples of features with different nature (ac). Estimated entropies and asymptotic evolution for representative example features (df). Estimated mutual information and asymptotic evolution for representative examples of pairs of features (gi).
Entropy 21 00419 g003
Figure 4. Averaged neighborhood of a specific customer C t , obtained for 100 independent realizations, as a function of the fixed threshold in distance d α m and the number of features required to fulfill its corresponding threshold, according to the X from M criterion.
Figure 4. Averaged neighborhood of a specific customer C t , obtained for 100 independent realizations, as a function of the fixed threshold in distance d α m and the number of features required to fulfill its corresponding threshold, according to the X from M criterion.
Entropy 21 00419 g004
Figure 5. Trade-off between False Negative Rate (a) and True Negative Rate (b).
Figure 5. Trade-off between False Negative Rate (a) and True Negative Rate (b).
Entropy 21 00419 g005
Table 1. Trade-off between Sensitivity and Specificity.
Table 1. Trade-off between Sensitivity and Specificity.
Sensitivity, TPR (%)Specificity, TNR (%)
d α m 1 from 52 from 53 from 54 from 55 from 51 from 52 from 53 from 54 from 55 from 5
189.1691.6696.8699.99100.099.6077.4838.0890.5494.67
289.1691.6196.7799.9999.9999.7989.4864.1983.4081.69
389.1691.6196.7799.9999.9999.8490.0365.2875.2768.51
489.1691.3996.5099.9799.9999.8490.9967.3971.3060.59
589.1690.0093.7199.9099.9499.8495.6676.3774.3758.89

Share and Cite

MDPI and ACS Style

González-Serrano, L.; Talón-Ballestero, P.; Muñoz-Romero, S.; Soguero-Ruiz, C.; Rojo-Álvarez, J.L. Entropic Statistical Description of Big Data Quality in Hotel Customer Relationship Management. Entropy 2019, 21, 419. https://doi.org/10.3390/e21040419

AMA Style

González-Serrano L, Talón-Ballestero P, Muñoz-Romero S, Soguero-Ruiz C, Rojo-Álvarez JL. Entropic Statistical Description of Big Data Quality in Hotel Customer Relationship Management. Entropy. 2019; 21(4):419. https://doi.org/10.3390/e21040419

Chicago/Turabian Style

González-Serrano, Lydia, Pilar Talón-Ballestero, Sergio Muñoz-Romero, Cristina Soguero-Ruiz, and José Luis Rojo-Álvarez. 2019. "Entropic Statistical Description of Big Data Quality in Hotel Customer Relationship Management" Entropy 21, no. 4: 419. https://doi.org/10.3390/e21040419

APA Style

González-Serrano, L., Talón-Ballestero, P., Muñoz-Romero, S., Soguero-Ruiz, C., & Rojo-Álvarez, J. L. (2019). Entropic Statistical Description of Big Data Quality in Hotel Customer Relationship Management. Entropy, 21(4), 419. https://doi.org/10.3390/e21040419

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop