Next Article in Journal
Kano Model Integration with Data Mining to Predict Customer Satisfaction
Previous Article in Journal
How Does Learning Analytics Contribute to Prevent Students’ Dropout in Higher Education: A Systematic Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Enhanced Parallelisation Model for Performance Prediction of Apache Spark on a Multinode Hadoop Cluster

1
School of Natural and Computational Sciences, Massey University, Auckland 0745, New Zealand
2
Department of Mechanical and Electrical Engineering, Massey University, Auckland 0745, New Zealand
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2021, 5(4), 65; https://doi.org/10.3390/bdcc5040065
Submission received: 22 September 2021 / Revised: 14 October 2021 / Accepted: 1 November 2021 / Published: 5 November 2021

Abstract

:
Big data frameworks play a vital role in storing, processing, and analysing large datasets. Apache Spark has been established as one of the most popular big data engines for its efficiency and reliability. However, one of the significant problems of the Spark system is performance prediction. Spark has more than 150 configurable parameters, and configuration of so many parameters is challenging task when determining the suitable parameters for the system. In this paper, we proposed two distinct parallelisation models for performance prediction. Our insight is that each node in a Hadoop cluster can communicate with identical nodes, and a certain function of the non-parallelisable runtime can be estimated accordingly. Both models use simple equations that allows us to predict the runtime when the size of the job and the number of executables are known. The proposed models were evaluated based on five HiBench workloads, Kmeans, PageRank, Graph (NWeight), SVM, and WordCount. The workload’s empirical data were fitted with one of the two models meeting the accuracy requirements. Finally, the experimental findings show that the model can be a handy and helpful tool for scheduling and planning system deployment.

1. Introduction

An increasing amount of data is coming from many applications, and it has become a challenging task to store and process big data efficiently [1]. In the last decade, researchers proposed and developed efficient distributed parallel file systems, such as MapReduce [2] and Spark [3], which provide various functions, including fault-tolerant, high scalability, open access [4], and simple application programming interfaces (APIs). Spark got prompt attention from professionals and researchers because of those features and fast data processing [5]. Spark can support of a wide range of data processing libraries, such as SQL spark for structured data processing; MLlib; and GraphX for machine learning, image processing, and streaming [6]. Besides, it can also store batch and streaming data and process this data using the applications and store the results in HDFS.
Spark introduced a new data abstraction technique called resilient distributed dataset (RDDs) [3] that improves multiple applications’ performances. Its application execution time is an essential factor in measuring real-time processing where the optimum execution time can be obtained based on accurate resource allocation. Spark’s performance expansively depends on the suitable selection of parameters, as this system has more than 150 parameters, and the selection and configuration of these parameters are challenging. The users need to adjust the configuration parameters as per the cluster resources; else, the cluster’s performance degrades significantly. Indeed, it is essential to select and configure the parameters that play an important role in system’s performance [7]. In the recent past, researchers proposed number of techniques, such as trial-and-error [8], cost-based (analytical) [9], and machine learning modelling [10,11]. However, all these techniques are either time-consuming or require large amounts of training and test data [12]. There are many issues practitioners may encounter when trying to model the performance of a cluster. One standard option to create a model is the use of machine learning algorithms, but this requires that enough sample runs are acquired, and this can take time. If not enough samples are acquired, capturing diverse data points, the accuracy of the model may suffer. Furthermore, machine learning can be a black box for the practitioner, and finding a simple model would be very useful because that would minimise the need to run workloads repeatedly too many times. Therefore, the following research question arises: “What parallelisation model for a Hadoop cluster can be found and implemented quickly and efficiently in order to improve the performance prediction of a job?”
Any algorithm can be parallelised, but not all algorithms can run efficiently in parallel machines such as a Hadoop cluster. It is a common phenomenon that the parallel performance depends mostly on how the algorithm operates and how nodes communicate to one another. In any parallel system, two of the most important parameters that will determine the runtime are the size of the job and the number of available executors (here executors can be interpreted as CPUs, or nodes of a cluster). Other parameters can drag the performance down, but they will not necessarily increase the performance. For example, if not enough memory is available to a job, this will increase runtime. If the minimum amount of memory is available, more memory will not make the job run faster. The number of available executors is very important, especially when the algorithm being parallelised requires communication that is not needed when the same algorithm is implemented and run on a single executor. For example, some algorithms are embarrassingly parallel (a term coined in the 90s), meaning that no extra work is needed when the job is parallelised. In this case, the speed-up is proportional to the number of processors available. In other cases, the speed-up can be super-linear, as in the case of searching algorithms running in parallel. Unfortunately, there are also groups of algorithms that do not present this optimistic speed-up [13]. One important factor that causes the degradation of performance is the fact that the algorithm may require extra communication and I/O operations that are inherently serial in nature.
The motivation for this paper was to extend our previous work, where we proposed a simple model to predict runtime as a function of number of executors [14]. The novelty of that work was the consideration of the importance of the amount of data in such performance prediction models. Accordingly, we extend the previous model and propose new parallelisation models that consider the number of executors and the the amount of data simultaneously. To the best of the authors’ knowledge, such models have not been published in the literature before. These new runtime performance prediction models rely on simple equations. They can potentially be as fast and as accurate as models created using machine learning. The authors have experimentally confirmed that the proposed ideas can be very useful for the runtime performance prediction of Spark jobs on the Hadoop cluster because they require minimum training data to achieve good predictions in less time.
The key contributions of this paper are as follows:
  • We introduced two distinct parallelisation models for performance prediction of Spark jobs on Hadoop cluster. Each model is based on a different communication pattern between the nodes of a Hadoop cluster.
  • We accomplished extensive experimental work. The authors analysed and verified the performance pattern based on two main parameters, the number of executors and the amount of data for each job. The data reliability was verified by running each workload at least three times.
  • We evaluated our models on five HiBench workloads in order to test the data fitting accuracy. Our results show that the experimental data fitted one of the models accurately, and the fitness was compared with Amdhal’s law, Gustafson’s law, and Ernest’s model. The data fitness was compared based on two criteria, Rsquared and RRSE.
The remainder of this paper is organised as follows: Section 2 provides a brief overview of Apache Spark. Section 3 presents some interesting Spark performance prediction based on a recently published Hadoop cluster-related study. In Section 4, we discuss existing models for runtime prediction for a Hadoop cluster. In Section 5, we describe a parallel model based on a fully connected network, and discuss the motivation for this model. In Section 6, we explain the experimental setup and present the workload execution and show the DAG of stages. Section 7 presents the results and analysis; in particular, it shows how the different equations fit the data. Finally, in Section 8 we present our conclusions with a discussion on the future developments for the model.

2. Apache Spark Platform

Matei Zahari developed Apache Spark at UC Berkely’s AMPLab in 2009 [3]. In 2010, Spark became an open-source project. Spark has since been very popular and serves as an alternative to the MapReduce model for open access, high-performance [4], and real-time data processing [15]. Spark presents a new way to process data faster, and its uses are in data analytics, big data processing, and machine learning. The major advantage of Apache Spark for machine learning is its end-to-end capabilities. As per the Datanyze market research [16], Apache Spark’s market share is almost 6.40% with more than 2770 companies globally. As per enlyft data [17], 59% of customers of Apache Spark are in the United State, 6% are in the United Kingdom, and 6% are in India.
Many programming languages, such as Python, Scala, Java, and SQL APIs, are embedded within this technology for use and development purposes. Compared to Hadoop, Spark offers a hundred times faster memory and ten times faster performance on disk. Due to its memory, Spark increases the performance of the application. Spark is an ecosystem which consists of various components, such as Spark SQL, Spark Streaming, Mllib, GraphX, and core API components. These components are designed to work closely to the core, and an application can be developed based on their libraries. Apache Spark offers well-defined architecture. In this architecture, the two main abstractions are Resilient Distributed Datasets (RDDs) and Directed Acyclic Graph (DAG). Generally, in the Spark cluster, an RDD collects the data and splits the data into partitions; then, this partitioned data are stored in the memory on worker nodes and parallel operations are performed. Spark’s RDD supports two types of operations, transformations and actions. A transformation uses the existing data to create a new dataset, and actions perform computations on the dataset and return their values to the driver program [18]. In Apache Spark, DAG consists of sequences of vertices and edges. Any job submitted in Spark creates a DAG and forwards the job into the stage level, where every stage is comprised of tasks based on input data and the RDD partition.
Apache Spark architecture uses master–slave systems with driver programs. The driver program runs as a master node, and the executors run as slave nodes. The executors start their processes once they receive the input file and continue until the job is completed. In this case, the executors keep themselves active the entire time and use multiple CPU threads for the task in parallel. The driver program creates the SparkContext and stores all the components. Spark driver and SparkContext look after the job execution in the cluster. In Spark, a job is executed in one or multiple physical units, and the jobs are divided into smaller sets of tasks at this stage. A single spark job can trigger many jobs that are dependent on the parent stage. Thus, the submitted job can be executed in parallel. Spark runs submitted jobs in two stages: ShuffleMapStage and ResultStages. ShuffleMapStage is an intermediate stage where the output data are stored for the following stages in the DAG. The ResultStages is the final stage of this process that applies a function to one or multiple partitions of the target RDD.
For any given work, the number of executors, the amount of data, and the number of threads play vital roles in the performance [19]. The block manager acts as a cache storage for a user’s program when executors allocate memory storage for the RDDs. Spark runs on a Hadoop cluster with Apache YARN (Yet Another Resource Negotiator) [20] as a framework for resource management and job scheduling or monitoring, in separate domains; and Apache Ambari manages, monitors, and profiles the individual workloads running the Hadoop cluster. Figure 1 shows a typical Spark cluster architecture.

3. Related Work

The current state-of-art in Spark performance predictions of big data has received widespread attention from researchers. Researchers have proposed several exciting works based on trial-and-error [8], grey-box modelling [21], black-box modelling [12], and machine learning approaches [22,23]. In this section, we present a review of the literature published in the recent past.
Petridis et al. [8] presented a trial-and-error methodology to predict the execution time of a Spark job. This work highlighted how the number of cores and parallelism play significant roles in the performance. There were twelve parameters considered with three benchmark applications—sort-by-key, shuffling, and Kmeans. They obtained significant performance improvements by using KryoSerializer rather than using the default Java Serializer, and the speed-up achievement was 10-fold. In their second work [24], they proposed an alternative systematic methodology for parameter tuning which can be applied to any computing infrastructure. They identified that the number of cores of the Spark executor has most impact; and the level of parallelism—for example, the number of partitions per participating core—plays a significant role in maximising the performance improvement.
Muhammad Usama Javaid et al. [25] proposed a robust performance model based on a machine learning (ML) algorithm. In order to train the ML algorithm, they used various amounts of input data, sets of spark parameters, and features.
A complex data-driven workflow application was proposed by Gulino et al. [26] where they combined ML and an analytical model to predict the execution times of arbitrary complex workflow applications.
Cheng et al. [27] proposed a machine-learning-based, efficient, performance prediction model for Apache Spark. This technique was capable of predicting the execution times accurately for the given application and configurations. At the stage level, Adaboost was used to build the model. They used projective sampling and data mining techniques to mitigate the modelling overhead. They claimed that the proposed model offers three advantages: no prior assumption of configuration parameters, stand-out robustness and stability, and less overall cost for the modelling process. They found that the average prediction error of the model was only 9% as compared to the other techniques.
In their second work, Cheng [11] proposed combined multi-object optimization (MOP) and an Adaboost algorithm to find the optimal configuration of parameters and predict the model’s performance. They evaluated the system with six Spark benchmarks. Five different datasets were used to analyse the performance. They claimed that the model can find the appropriate configuration setup and minimise the time and computational cost. The average improvement in computational cost was about 35% to 40%.
Aziz et al. [28] presented the resource management and data processing, the system processing time and speed-up, and the impact of persistence of resilient distributed datasets (RDDs) in Spark based on machine learning algorithms. In this analysis, the appropriate storage level of execution time was presented for Spark using a machine learning algorithm in RDD. They found that the speed-up does not improve by adding additional nodes, and the performance is degraded; and the total processing time increased significantly. There were many factors behind this degradation: among them, the most significant reason was the 100% allocation of cores to executors.
Boden [29] proposed a representative set of machine learning algorithms (supervised and unsupervised) to investigate large-scale datasets. The mathematical variation and appropriate system parameters were tuned for the amount of data and dimensionality of the data. The author reported that machine learning algorithm problems exhibit very high dimensionality due to data scaling and model size scaling. Therefore, they focused on the aspects likely affecting scaling the data and scaling the model’s dimensionality. Their study found that as the the amount of data was increased, the system exhibited linearly increments in time consumed.
A cost–benefit Spark performance prediction model based on a machine learning algorithm was proposed by Maros [30]. They have proposed both black-box and grey-box models based on four machine learning algorithms. They considered three different aspects: the amount of training data, platform configurations, and workloads. They compared their model with Ernest [31]. They found that the performance estimation error was better than Ernest when the dataset extrapolation was required. Mustafa [10] proposed a new platform to predict the execution time for SQL queries and machine learning applications. This technique is very similar to the grey-box model. They applied three different approaches which used existing methods to predict the execution times of the queries. The authors claimed that the SQL query workload produced less than 10% error, whereas the machine learning workload produced less than 25%.
An exciting system was proposed by Amannejad et al. [32], which can predict the execution time in a short time. In this method, minimum resource settings are considered, which do not have complex dependencies and parallel stages. An application is used to analyse the work log files. This method requires two reference files, and the files are relatively small. This method had excellent accuracy regarding execution time, where the average prediction error of the workloads was about 4.8%. Unlike this work, they considered only a single node cluster, not a real cluster environment.
In a related but alternative model, PERIDOT was presented by Amannejad et al. in their second paper [33]. A small subset of input data and fixed limited cluster resources settings were considered to get quick execution time. They analysed the logs from both the executions and checked the internal dependencies between the internal stages. There were eight HiBench workloads used this experiment. They reported that the data partitions and the number of executors had significant impacts on execution time. This method had an overall mean prediction error of 6.6%, except for naive prediction techniques.
We summarise the different approaches in Table 1.
Unlike our approach, other approaches described in the literature may require time to modify several default parameters, which are very complex and tedious to work with. Apart from this, machine learning models usually require a large number of experiments in order to generate enough data for model training. Our proposed models need very few experiments, fitting the data obtained into simple equations. The equations can also give some insights into the pattern of communication between the nodes when running Spark jobs.

4. Parallelisation Models

4.1. Amdahl’s Law and Gustafson’s Law

If no communication between the various executors is needed to run a job, the job is called “embarrassingly parallel” [13]. The implication of having no need to communicate between different executors is that the speed-up is proportional to the number of executors; i.e., if one executor takes time t, then n executors will take time t n . However, any small portion of the job that is not parallelisable can bring major consequences for parallel performance. In this case, the linear speed-up achieved by adding more executors (in the form of CPUs or cores, or separate nodes) may decline sharply.
Amdahl came up with a generic equation to predict the speed-up factor of a parallel application as a function of the number of processors [36]. The equation considers that parts of the application (or job, or workload) are inherently serial in nature and would not be parallelisable.
S ( n e x e c ) = n e x e c 1 + ( n e x e c 1 ) f n p
where S ( n e x e c ) is a function that represents the speed-up as a function of the number of executors, n e x e c is the number of executors (often interpreted and nodes or CPUs available in the infrastructure), and f n p is the factor of non-parallelisable portions of a job. f n p = 0 represents a perfectly parallelisable job that will yield full speed-up (e.g., if there are 10 executors available, the job will run 10 times faster, or S ( 10 ) = 10 ).
From Equation (1), and considering that a single processor takes time t to run a certain workload, the predicted runtime running on multiple processors would be:
r u n t i m e = ( 1 f n p ) t n e x e c + f n p t
where t is a hypothetical runtime needed to run a job in a single executor.
If we consider the size of the job, we can modify Equation (2) to:
r u n t i m e = a f ( S i z e ) ( 1 f n p ) n e x e c + f n p
where a is a constant coefficient, and f ( S i z e ) is a function that reflects the growth of the runtime with increasing sizes (an approximation of the algorithm complexity). As most of the workloads implemented in HiBench are either linear or quadratic, f ( S i z e ) can be replaced by either S i z e or S i z e 2 .
An example of what Amdahl’s law means for different serial factors and numbers of executors is shown in Figure 2. It shows the influence of both parameters on the runtime of a simulated job with a fixed dataset size.
A few years after Amdahl’s publication, Gustafson argued that the percentage of the serial part of a job is rarely fixed for different problem sizes [37]. In Amdahl’s law even a small percentage of serial work can be detrimental to the potential speed-up after adding more executors. Gustafson noticed that for many practical problems the serial portion would not grow with an increase in problem size. Gustafson’s speed-up equation is:
S ( n e x e c ) = n e x e c + ( 1 n e x e c ) f n p
Additionally, the runtime equation as a function of S i z e and n e x e c can be written as:
r u n t i m e = a f ( S i z e ) n e x e c + ( 1 n e x e c ) f n p
Both Amdahl’s and Gustafson’s equations show that runtime will always go down as the number of executors increases. However, often in practice the communication can impose an overhead, so runtime might increase after a certain limit on the number of executors. We compare Equations (3) and (5) to our own model of parallelisation, as discussed in the next section.
An example of what Gustafson’s Equation (5) means for different serial factors and numbers of executors is shown in Figure 3. It shows the influences of both parameters on the runtime of a simulated job with a fixed dataset size.

4.2. A Model Using a 2D Plate Communication Pattern

In our previous work [14] a model using a 2D plate communication pattern based on a description by [13] was proposed and tested. In that work, we used the following equation, which is a function of n e x e c only (fixed problem sizes):
r u n t i m e = a n e x e c + b n e x e c
where a and b are coefficients, and n e x e c is the number of executors.
The equations were based on a communication model where each node has to exchange information with certain neighbours, but not all. Figure 4 shows an example of the boundaries of communication between nodes for this model.
The second part of Equation (6) is a function of s q r t ( n e x e c ) . We have also experimented with a different function, n e x e c c , where c < = 1 , and found that for some data it worked even better. Furthermore, based on [31], we added a constant term d to the equation to improve the fitting. Therefore, Equation (6) can be rewritten as:
r u n t i m e = a f ( S i z e ) n e x e c + b g ( S i z e ) n e x e c c + d
where f ( S i z e ) is a function representing the time complexity of the workload when it runs on one executor, and g ( S i z e ) is a function that indicates the growth of the communication and overhead when parallelising the job. After preliminary experiments, we found that g ( S i z e ) = S i z e works well.
For linear algorithms, f ( S i z e ) = S i z e , so we can rewrite Equation (7) as:
r u n t i m e = a S i z e n e x e c + b S i z e n e x e c c + d
If the workload is quadratic, f ( S i z e ) = S i z e 2 , and Equation (7) can be rewritten as:
r u n t i m e = a S i z e 2 n e x e c + b S i z e n e x e c c + d

5. An Enhanced Model for Runtime Prediction

The model described in Section 4 has its limitations, as it assumes that each node would only communicate with a small number of neighbour nodes. We observed that although the model fits the data well for some workloads, it still may not reflect the communication that may be required when using HDFS, where copies of the data may be anywhere in the cluster. It is a known issue that for different algorithms, different communication patterns emerge [13], as the algorithm itself may require data located elsewhere or computations carried out by other nodes. In the proposed models, we took communication as a single factor, as this simplifies the model.
In order to expand the model to include both the number of executors and job sizes, we decided to reformulate the model. Thus, we considered that a node (where the executors have CPU resources) can communicate with any other node in the cluster. Although the communication pattern is not known for a black box implementation, we can infer what is happening through the empirical data acquired by running the same workload with many sizes and numbers of executors.
The communication pattern is assumed to be of the shape of a fully connected graph (see Figure 5). The assumption is that a function (unknown) of the size and number of executors drives the extra runtime needed to complete the communication between nodes. The extra runtime is, of course, compensated by the extra nodes involved in the job. Therefore, two components of the equation drive the runtime in opposite directions: the extra nodes will divide the processing to run the job, but communication between them requires extra time. The basic parallel equations start as:
r u n t i m e = t n e x e c + t s e r i a l
where t is the runtime for a job of a certain size to run in a single executor (no communications involved), n e x e c is the number of executors, and t s e r i a l is the serial portion of the job that cannot be parallelised, here considered to be communication overheads and any other overheads required to run the job in parallel.
If the size is added to the model, we need to know the algorithmic complexity of the implemented code.
r u n t i m e = f ( S i z e ) n e x e c + t s e r i a l
For simplicity, we assume that each node will communicate with every other node, and that the communication (be it HDFS or partial computations being exchanged between nodes) is symmetric and homogeneous. This makes the growth of t s e r i a l a function of both the size and number of executors. We hypothesise that a good approximation for t s e r i a l depends on the number of links between the nodes. This is the same as the number of edges in a fully connected graph:
n l i n k s = n e x e c ( n e x e c 1 ) 2
Furthermore, the serial portion becomes:
t s e r i a l = g ( S i z e ) n e x e c ( n e x e c 1 ) 2
Equation (11) can be rewritten as:
r u n t i m e = f ( S i z e ) n e x e c + g ( S i z e ) n e x e c ( n e x e c 1 ) 2
In the parallelisable part of the runtime, the function f ( S i z e ) can be simplified to the complexity of the algorithm implemented for the workload. However, for the serial part, the g ( S i z e ) function is unknown. The data can fit well considering that g ( S i z e ) = b S i z e c , where c is a constant exponent less than 1. For better fitting, we added another coefficient, d, representing a constant term for a given dataset (similarly to [31]).
For linear algorithms, f ( S i z e ) = a S i z e , so we can rewrite Equation (14) as:
r u n t i m e = a S i z e n e x e c + b S i z e c n e x e c ( n e x e c 1 ) 2 + d
If the workload is quadratic, f ( S i z e ) = a S i z e 2 , and Equation (14) can be rewritten as:
r u n t i m e = a S i z e 2 n e x e c + b S i z e c n e x e c ( n e x e c 1 ) 2 + d

6. Experiments

6.1. Experimental Setup

The experimental big data cluster used in this work was designed and developed by a group of academics at Massey University, Auckland campus [38]. The hardware for this experimental big data cluster is similar to a Beowful cluster. The cluster runs on dedicated network infrastructure with dedicated switches. All other network machines are kept away from this infrastructure to reduce the network latency and unwanted network resource utilisation. The cluster was designed and developed with one master node and nine slave/worker nodes. The Hadoop cluster server and node configuration is presented in Table 2. The schematic diagram of the cluster is presented in Figure 6.

6.2. Experiment Performance Evaluation

HiBench [39] is a popular big data benchmark suite that helps researchers and professionals to evaluate big data frameworks’ performances. HiBench offers various characteristics and evaluates cluster deployment through comprehensive benchmarking [40]. It consists of various Hadoop programs, namely, synthetic micro-benchmarking and real-world applications. This experiment used five workloads from four different benchmark categories: Micro-Benchmark (WordCount), Machine Learning (Kmeans and SVM), Web Search (PageRank), and Graph (NWeight). The statistics of the experimental workloads are presented in Table 3 and the workload of Spark HiBench’s characteristics are presented in Table 4. Our target was to predict the Spark execution time considering the above workloads, and show how the execution time will be fitted with the proposed models. The individual workload DAG of stages and their execution are presented in the following section.
WordCount (WC): The WC workload performs the operation based on the Map function, which transforms the data into various representations. In HiBench, the WC input data are produced based on RandomTextWriter, which is contained in the Hadoop distribution. It counts the occurrences of separate words from the text or sequence file. An example of a job execution plan and its DAG of stage is presented in Figure 7. As shown in the figure, WC performed the operation in two stages; five tasks were involved in this operation.
Kmeans: K-means is a well-known clustering algorithm that is commonly used for knowledge discovery and data mining. The Kmeans input data are a group of samples, generated by GenKMeansDataset, which is based on uniform and Gaussian distribution. We used various amounts of input data, such as tiny, small, and large, with the dimensions of 3 and 20; “no of cluster”, 5; “max-iteration”, 5; centroid, 10; and converged, 0.5. The job was executed in 19 different stages, and a sample of job DAG stages is shown in Figure 8.
SVM: Support vector machines (SVMs) are used for large-scale data classification tasks. It is considered one of the standard methods of big data classification. In Spark, MLlib is used for SVM workload implementation. Its input data are generated by the SVM DataGenerator. We selected the SVM parameters such as number of iterations, stepSize, and regParam and modified their values to 100, 1.0, and 0.01. The system required 213 stages to complete the task. Figure 9 shows a sample DAG of SVM.
NWeight: NWeight is implemented by Spark GraphX library and pregel, and it works as an iterative parallel algorithm. It enhances the Spark RDD with a directed multigraph, which consists of properties enclosed with vertices and edges. The input files consist of millions of edges. It required eight different stages to complete the task. The workload DAG of stages is shown in Figure 10.
PageRank: PageRank is a well-known page search algorithm where every page has a unique number, and an individual page is ranked as per the vote. The vote is counted when the pages are connected with the other pages. Generally, when a page is linked with several different pages, it is considered as a higher PageRank. In PageRank, the data source is generated from Web data. The hyperlinks of those data follow the Zipfian distribution. Various sets of input samples (from thousands to millions) were used in this experiment. The job was executed in four different stages. Figure 11 shows the workload DAG of stages.

6.3. Configuration of Parameters

One of the challenging jobs of Spark cluster deployment is the parameter selection. There are more than 150 configurable parameters [8] in the Spark system, and each parameter plays a vital role in improving system performance. Spark’s cluster performance relies on hardware infrastructure and accurate parameter selection. Performance improvements can be achieved by tuning the values of the parameters. The configuration of these parameters needs to be investigated according to the applications, amount of data, and cluster architecture. However, some influential parameters—executors, executors core, executor memory, driver memory, etc.—massively influence the system’s performance. Besides, normally the number of parameters used is the default value. We have performed extensive experiments, selected the impactful parameters, and tuned their crucial factors to validate our cluster to get optimum performance. In the recent past, some studies [35,41] were carried out which illustrate the impacts and importance of the parameters. The chosen parameters for our study are listed in Table 5.
The default column in Table 5 presents the system’s default configuration, the range column presents the tuned values used in this experiment, and the description column presents parameter information. There were two reasons to choose these parameters: firstly, Spark’s runtime performance heavily depends on these parameters; secondly, these parameters control pivotal resources: CPU, disk read and write, and memory [42].

7. Results and Analysis

In this part, we present the experimental findings. We have considered various amounts of data and systematically increased the number of executors to study the system’s behaviour. For the results’ reproducibility, each experiment was repeated at least three times, and the average execution time was taken into consideration in the final graph. We have collected the log files from the history server and calculated the job execution time using a Python script. We found that there is some fraction of the time difference between Ambari and our Python script. We have considered the most realistic time.

7.1. Procedure to Fit Equations

The procedure for the experiments for the HiBench workloads used for this work is summarised as follows. Firstly, for a certain workload, we ran a few jobs with different sizes using only one executor (Section 6.2). We then estimated the function f ( S i z e ) , which reflects the time complexity of the implemented algorithm (Section 7.2).
We ran more jobs with similar sizes as above, while varying the number of executors (Section 6.2). For each workload, we used up to 14 executors and five different sizes, chosen appropriately for each workload (Table 3).
Using the multi-parameter fitting function available in Gnuplot [43], each dataset was fitted to the following Equations: (8) (linear) or (9) (quadratic), (15) (linear) or (16) (quadratic), Amdahl (3), Gustafson (5), and Ernest [31] (Section 7.3).
The best fit for each equation above was chosen considering Rsquared and R S E μ as a criterion, as discussed in Section 7.4.

7.2. Finding the Approximate Algorithm Complexity ( f ( S i z e ) )

Regarding the nominal time complexity of the algorithms used in HiBench for these experiments, some are linear and some are quadratic. SVM is typically O ( N 2 ) or even O ( N 3 ) [44]. K-means is usually quadratic [45]. The PageRank algorithm can be O ( n m ) , where n is the number of nodes and m is the number of arcs [46]. WordCount is usually linear O ( N ) [47]. The Graph (NWeight) algorithm can be either linear or quadratic, depending on the graph representation. Using an edge list, it is quadratic O ( N 2 ) [48].
In order to find the function f ( S i z e ) for Equation (11), we ran several jobs using a single executor. The results are shown in Figure 12.
Based on the residual standard error (given in Gnuplot [43] as the r m s value) of the fitting to linear or quadratic trends, it was found that only SVM and NWeight had quadratic trends. The other methods produced linear trends. The appropriate equations were fitted to the complete data. Dataset size and number of executors were used as parameters for the model.
The possible complexity for the workloads and the actual data fittings for single executors are summarised in Table 6.

7.3. The Full Model Fitting

After running several jobs with different HiBench workloads, we collected runtime data for various sizes and numbers of executors. The data were fitted using equations for the fully connected model: Equation (14) (replaced by Equation (15) for linear complexity or (16) for quadratic complexity); and for the special case where c = 1 , Amdahl’s Equation (3), Gustafson’s Equation (5), Ernest’s equation [31], and the 2D plate model using Equation (7) (replaced by Equation (8) for linear complexity or (9) for quadratic complexity). The best fit is shown in the figures below, considering the fitting criteria described in Section 7.4. Table 7 presents the Rsquared results for each equation, and Table 8 presents the RRSE results for each equation.
Figure 13 shows the graph presentation of the WordCount workload for the amounts of data between 0.3 and 19 GB. The best fit used Equation (8) and there was a draw with Amdahl’s equation, yielding an Rsquared of 0.997 and an RRSE of 0.074.
Figure 14 shows the SVM workload for sizes between 0.034 and 2 GB. Equations (9) and (16) were the best fit for the data, with an Rsquared of 0.917 and an RRSE of 0.271. The relatively high RRSE indicates that the data may be dependent on other factors, which may be investigated in future works.
Figure 15 shows the PageRank workload for sizes between 0.057 and 5 GB. Equation (8) and Amdahl’s were the best fit for the data, with an Rsquared of 0.990 and an RRSE of 0.113.
Figure 16 shows the Kmeans workload for sizes between 19 and 168 GB. Equation (8) was the best fit for the data, with an Rsquared of 0.993 and an RRSE of 0.130.
Figure 17 shows the Graph (NWeight) workload for sizes between 37 and 211 MB. Equation (9) was the best fit for the data with an Rsquared of 0.966 and an RRSE of 0.189.

7.4. Evaluation of the Fitting Errors

The proposed models’ fitting results are shown in Section 7. The nominal time complexities of different implemented algorithms in the HiBench workloads are also presented. This section illustrates the accuracy error of the proposed models, and shows the comparison results among the two proposed models and Amdahl’s and Gustafson’s laws. In addition, the proposed models offer an improvement over those of Ernest [31]. Our results revealed that accuracy and effectiveness of our proposed models are better than those of the Ernest models. We used both the Rsquared ( R 2 ) and the relative residual standard error R M S μ as metrics for the quality of the fitting. Rsquared values (also known as coefficient of determination) are calculated by the following equation:
R 2 = 1 S S r e s S S t o t
where S S r e s is the sum of the squares of the residuals and S S t o t is the sum of the squares relative to the mean of the data. For a perfect fitting, S S r e s = 0 and R 2 = 1 , so the closer R 2 is to one, the better the fitting.
The residual standard error [49] is:
R S E = i = 1 n ( y i y i ¯ ) 2 d f
where ( y i y i ¯ ) is the difference between the observed data and the predicted value using the model, and d f is the degrees of freedom given by the number of samples minus the number of parameters being fitted.
The relative residual standard error is:
R R S E = R S E μ
The R R S E gives a metric for the error or distance between the observed points and the ones generated by the model. The smaller the R R S E , the better the fit accuracy.

7.5. Benefits of the Proposed Models

Our work considered three well-established equations, namely, those of Amdhal [36], Gustafson [37], and Ernest [31], as comparative models. We limited our analysis to these models while recognising the existence of alternative models in the published literature, which we deemed out of scope for the purposes of this study. As shown in Table 7 and Table 8, for every workload, both proposed models (Equations (8) or (9), and (15) or (16)) had better fitting results than Ernest. Only in two cases did Amdahl’s model Rsquared tie with our models, Wordcount and PageRank. Only in the case of Kmeans workload did Gustafson’s model tie with one of the proposed models (Equation (8)). The results show that the two proposed models either tied with or performed better than the previously published models.
Considering the above results, the proposed models can be used as very effective tools for performance prediction, as they can offer several benefits for Spark job run time prediction using the Hadoop cluster. Several key benefits differentiate the proposed models from existing approaches. One of the crucial benefits is that using one of the proposed equations, it is possible to estimate the runtime. This can be achieved with a small number of experiments given the amount of data for the job and the chosen number of executors. The major advantage of the proposed models is that they do not require any trial-and-error approach, nor do they require large amounts of training or test data that are usually needed by machine learning models. Both models can capture the performance characteristics of a large number of complex workloads and are capable of predicting the runtime with good accuracy. Finally, the results also show that the models are highly effective, generic, and platform agnostic. Based on these models, it is possible for the managerial teams of big-data-driven organisations to minimise the time of their systems’ configuring processes, plan and schedule large jobs by allocating critical resources for the clusters, and choose appropriate numbers of executors to maximise resource utilisation.

8. Conclusions

This paper proposed and investigated parallelisation models with enhanced capabilities for predicting the runtime performance of Apache Spark for several workloads running on Hadoop clusters. The configuration of Spark parameters is a complex and challenging task for the users. The system’s performance mainly depends on the user’s choice and targets.
To overcome this challenge, we proposed two models based on a function of the two most important parameters, the number of executors and the size of the job. A significant contribution of this work is the finding that with limited data points, one can fit the data into simple equations and understand the pattern of communication between the nodes when running Spark jobs in a Hadoop cluster. We have found that the communication patterns can vary wildly between different workloads. This is expected, as different algorithms have different requirements from the Hadoop cluster. However, it can be noted that all the workloads used in the experiments could fit one of the two proposed models. The experimental results show that all the workloads could be fitted very accurately using the models proposed and completely outperformed the Ernest model. For two of the workloads (SVM and NWeight), the Rsquared values produced were lower than those the other three, alongside relatively high residual standard error. The two models fit the data better or at least as well as other alternative models (Amdahl, Gustafson, and Ernest).
However, the proposed models should be evaluated with other benchmark workloads, such as SQL and streaming. Due to time constraints, we considered only five workloads and selected a limited number of suitable Spark parameters. As future work, we have a plan to test the proposed model on the latest version of Apache Spark. Besides, we aim to add more suitable Spark parameters and workloads, and compare the proposed models with machine learning models. Furthermore, we intend to expand the experiments to other HiBench workloads to determine which equations are more suitable for which workloads.

Author Contributions

Conceptualization: N.A. and A.L.C.B.; methodology: N.A. and A.L.C.B.; resources: N.A.; validation: N.A. and A.L.C.B.; formal analysis: N.A. and A.L.C.B.; investigation: N.A. and A.L.C.B.; Data curation: N.A. and A.L.C.B.; writing—original draft preparation: N.A. and A.L.C.B.; writing—review and editing: N.A., A.L.C.B., M.A.R. and T.S.; visualization: N.A. and A.L.C.B.; supervision: A.L.C.B., M.A.R. and T.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the REaDI funding [Project code: 96670].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article. However, the correspondence author can be contacted for more details.

Acknowledgments

This work was supported in part by the Massey University Doctoral Scholarship.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SVMSupport vector machines
APIApplication programming interface
SQLStructured query language
HDFSHadoop distributed file system
RDDResilient distributed datasets
MLlibMachine learning library
CPUCentral processing unit
I/Oinput/output
UCUniversity of california
AMPAlgorithms, machines and people
DAGDirected acyclic graph
YARNYet another resource negotiator
PERIDOTPerformance predIction moDel fOr Spark applicaTions
NEXECNumber of executor
SQRTSquare root
2DTwo dimensional
GHzGigahertz
TBTerabyte
RAMRandom access memory
DDRDouble data rate
GBGigabyte
MBMegabyte
WCWordCount
ExecExecutor
MOPMulti-object optimization

References

  1. Katal, A.; Wazid, M.; Goudar, R.H. Big data: Issues, challenges, tools and good practices. In Proceedings of the 2013 Sixth international conference on contemporary computing (IC3), Nodia, India, 8–10 August 2013; pp. 404–409. [Google Scholar]
  2. Dean, J.; Ghemawat, S. Mapreduce: Simplified data processing on large clusters. Commun. ACM 2008, 1, 107–113. [Google Scholar] [CrossRef]
  3. Zaharia, M.; Chowdhury, M.; Das, T.; Dave, A.; Ma, J.; McCauly, M.; Franklin, M.J.; Shenker, S.; Stoica, I. Resilient distributed datasets: A fault-tolerant abstraction for in-memory cluster computing. In Proceedings of the 9th Symposium on Networked Systems Design and Implementation(NSDI), San Jose, CA, USA, 25 April 2012; pp. 15–28. [Google Scholar]
  4. Mazhar Javed, A.; Rafia Asad, K.; Haitham, N.; Awais, Y.; Syed Muhammad, A.; Usman, N.; Vishwa Pratap, S. A Recommendation Engine for Predicting Movie Ratings Using a Big Data Approach. Electronics 2021, 10, 1215. [Google Scholar]
  5. Zaharia, M.; Xin, R.S.; Wendell, P.; Das, T.; Armbrust, M.; Dave, A.; Meng, X.; Rosen, J.; Venkataraman, S.; Franklin, M.J.; et al. Apache spark: A unified engine for big data processing. Commun. ACM 2016, 59, 56–65. [Google Scholar] [CrossRef]
  6. Meng, X.; Bradley, J.; Yavuz, B.; Sparks, E.; Venkataraman, S.; Liu, D.; Freeman, J.; Tsai, D.B.; Amde, M.; Owen, S.; et al. Mllib: Machine learning in apache spark. J. Mach. Learn. Res. 2016, 17, 1235–1241. [Google Scholar]
  7. Kroß, J.; Krcmar, H. PerTract: Model Extraction and Specification of Big Data Systems for Performance Prediction by the Example of Apache Spark and Hadoop. Big Data Cogn. Comput. 2019, 3, 47. [Google Scholar] [CrossRef] [Green Version]
  8. Petridis, P.; Gounaris, A.; Torres, J. Spark Parameter Tuning via Trial-and-Error. In Proceedings of the INNS Conference on Big Data; Springer: Cham, Switzerland, 2016; pp. 226–237. [Google Scholar]
  9. Herodotou, H.; Lim, H.; Luo, G.; Borisov, N.; Dong, L.; Cetin, F.B.; Babu, S. Starfish: A self-tuning system for big data analytics. In Proceedings of the 5th Biennial Conference on Innovative Data Systems Research (CIDR), Asilomar, CA, USA, 9–12 January 2011; pp. 261–272. [Google Scholar]
  10. Mustafa, S.; Elghandour, I.; Ismail, M.A. A machine learning approach for predicting execution time of spark jobs. Alex. Eng. J. 2018, 57, 3767–3778. [Google Scholar] [CrossRef]
  11. Cheng, G.; Ying, S.; Wang, B. Tuning configuration of apache spark on public clouds by combining multi-objective optimization and performance prediction model. J. Syst. Softw. 2021, 180, 111028. [Google Scholar] [CrossRef]
  12. Wang, G.; Xu, J.; He, B. A novel method for tuning configuration parameters of spark based on machine learning. In Proceedings of the 2016 IEEE 18th International Conference on High Performance Computing and Communications; IEEE 14th International Conference on Smart City; IEEE 2nd International Conference on Data Science and Systems (HPCC/SmartCity/DSS), Sydney, NSW, Australia, 12–14 December 2016; pp. 586–593. [Google Scholar]
  13. Wilkinson, B.; Allen, M. Parallel Programming, 2nd ed.; Prentice Hall: Hoboken, NJ, USA, 1999; p. 268. [Google Scholar]
  14. Ahmed, N.; Barczak, A.L.; Rashid, M.A.; Susnjak, T. A Parallelization Model for Performance Characterization of Spark Big Data Jobs on Hadoop Clusters. J. Big Data 2021, 8, 1–28. [Google Scholar] [CrossRef]
  15. Mavridis, I.; Karatza, H. Performance evaluation of cloud-based log file analysis with apache hadoop and apache spark. J. Syst. Softw. 2017, 125, 133–151. [Google Scholar] [CrossRef]
  16. Apache Spark Market Share. Available online: https://www.datanyze.com/market-share/big-data-processing–204/apache-spark-market-share (accessed on 9 October 2021).
  17. Companies using Apache Spark. Available online: https://enlyft.com/tech/products/apache-spark (accessed on 9 October 2021).
  18. Apache Spark Overview 2.4.4. RDD Programming Guide. Available online: https://spark.apache.org/docs/2.4.4/ (accessed on 7 August 2020).
  19. Chen, Y.; Goetsch, P.; Hoque, M.A.; Lu, J.; Tarkoma, S. d-simplexed: Adaptive delaunay triangulation for performance modeling and prediction on big data analytics. IEEE Trans. Big Data 2019, 1–12. [Google Scholar] [CrossRef]
  20. Vavilapalli, V.K.; Murthy, A.C.; Douglas, C.; Agarwal, S.; Konar, M.; Evans, R.; Graves, T.; Lowe, J.; Shah, H.; Seth, S. Apache hadoop yarn: Yet another resource negotiator. In Proceedings of the 4th Annual Symposium on Cloud Computing, Santa Clara, CA, USA, 1–3 October 2013; pp. 1–16. [Google Scholar]
  21. Al-Sayeh, H.; Hagedorn, S.; Sattler, K.U. A gray-box modeling methodology for runtime prediction of apache spark jobs. Distrib. Parallel Databases 2020, 38, 1–21. [Google Scholar] [CrossRef] [Green Version]
  22. Assefi, M.; Behravesh, E.; Liu, G.; Tafti, A.P. Big data machine learning using apache spark mllib. In Proceedings of the 2017 IEEE International Conference on Big Data (Big Data), Boston, MA, USA, 11–14 December 2017; pp. 3492–3498. [Google Scholar]
  23. Taneja, R.; Krishnamurthy, R.B.; Liu, G. Optimization of machine learning on apache spark. In Proceedings of the International Conference on Parallel and Distributed Processing Techniques and Applications (PDPTA), Las Vegas, NV, USA, 25–28 July 2016; pp. 163–167. [Google Scholar]
  24. Gounaris, A.; Torres, J. A methodology for spark parameter tuning. Big Data Res. 2018, 11, 22–32. [Google Scholar] [CrossRef] [Green Version]
  25. Javaid, M.U.; Kanoun, A.A.; Demesmaeker, F.; Ghrab, A.; Skhiri, S. A Performance Prediction Model for Spark Applications. In Proceedings of the International Conference on Big Data, Honolulu, HI, USA, 18–20 September 2020; pp. 13–22. [Google Scholar]
  26. Gulino, A.; Canakoglu, A.; Ceri, S.; Ardagna, D. Performance Prediction for Data-driven Workflows on Apache Spark. In Proceedings of the 2020 28th International Symposium on Modeling, Analysis, and Simulation of Computer and Telecommunication Systems (MASCOT.S.), Nice, France, 17–19 November 2020; pp. 1–8. [Google Scholar]
  27. Cheng, G.; Ying, S.; Wang, B.; Li, Y. Efficient performance prediction for apache spark. J. Parallel Distrib. Comput. 2021, 149, 40–51. [Google Scholar] [CrossRef]
  28. Aziz, K.; Zaidouni, D.; Bellafkih, M. Leveraging resource management for efficient performance of apache spark. J. Big Data 2019, 6, 1–23. [Google Scholar] [CrossRef] [Green Version]
  29. Boden, C.; Spina, A.; Rabl, T.; Markl, V. Benchmarking data flow systems for scalable machine learning. In Proceedings of the 4th ACM SIGMOD Workshop on Algorithms and Systems for MapReduce and Beyond, Chicago, IL, USA, 19 May 2017; pp. 1–10. [Google Scholar]
  30. Maros, A.; Murai, F.; da Silva, A.P.C.; Almeida, M.J.; Lattuada, M.; Gianniti, E.; Hosseini, M.; Ardagna, D. Machine learning for performance prediction of spark cloud applications. In Proceedings of the 2019 IEEE 12th International Conference on Cloud Computing (CLOUD), Milan, Italy, 8–13 July 2019; pp. 99–106. [Google Scholar]
  31. Venkataraman, S.; Yang, Z.; Franklin, M.; Recht, B.; Stoica, I. Ernest:Efficient performance prediction for large-scale advanced analytics. In Proceedings of the 13th Symposium on Networked Systems Design and Implementation (NSDI), Santa Clara, CA, USA, 16–18 March 2016; pp. 363–378. [Google Scholar]
  32. Amannejad, Y.; Shah, S.; Krishnamurthy, D.; Wang, M. Fast and lightweight execution time predictions for spark applications. In Proceedings of the 2019 IEEE 12th International Conference on Cloud Computing (CLOUD), Milan, Italy, 8–13 July 2019; pp. 493–495. [Google Scholar]
  33. Shah, S.; Amannejad, Y.; Krishnamurthy, D.; Wang, M. Quick execution time predictions for spark applications. In Proceedings of the 2019 15th International Conference on Network and Service Management (CNSM), Halifax, NS, Canada, 21–25 October 2019. [Google Scholar]
  34. Ahmed, N.; Barczak, A.L.; Susnjak, T.; Rashid, M.A. A comprehensive performance analysis of apache hadoop and apache spark for large scale data sets using HiBench. J. Big Data 2020, 7, 1–18. [Google Scholar] [CrossRef]
  35. Chao, Z.; Shi, S.; Gao, H.; Luo, J.; Wang, H. A gray-box performance model for apache spark. Future Gener. Comput. Syst. 2018, 89, 58–67. [Google Scholar] [CrossRef]
  36. Amdahl, G.M. Validity of the single processor approach to achieving large scale computing capabilities. In Proceedings of the AFIPS ’67 (Spring): Spring Joint Computer Conference, Sunnyvale, CA, USA, 18–20 April 1967; pp. 483–485. [Google Scholar]
  37. Gustafson, J.L. Reevaluating amdahl’s law. Commun. ACM 1988, 31, 532–533. [Google Scholar] [CrossRef] [Green Version]
  38. Barczak, A.L.; Messom, C.H.; Johnson, M.J. Performance characteristics of a cost-effective medium-sized beowulf cluster supercomputer. In Proceedings of the International Conference on Computational Science, Melbourne, Australia, 2–4 June 2003; pp. 1050–1059. [Google Scholar]
  39. HiBench Suite. Available online: https://github.com/Intel-bigdata/HiBench (accessed on 4 June 2019).
  40. Huang, S.; Huang, J.; Dai, J.; Xie, T.; Huang, B. The HiBench benchmark suite: Characterization of the mapreduce-based data analysis. In Proceedings of the 2010 IEEE 26th International Conference on Data Engineering Workshops (ICDEW 2010), Long Beach, CA, USA, 1–6 March 2010; pp. 41–51. [Google Scholar]
  41. Zhao, Y.; Hu, F.; Chen, H. An adaptive tuning strategy on spark based on in-memory computation characteristics. In Proceedings of the 2016 18th International Conference on Advanced Communication Technology (ICACT), PyeongChang, Korea, 31 January–3 February 2016; pp. 484–488. [Google Scholar]
  42. Marcu, O.C.; Costan, A.; Antoniu, G.; Perez-Hernandez, M. Spark versus flink: Understanding performance in big data analytics frameworks. In Proceedings of the 2016 IEEE International Conference on Cluster Computing (CLUSTER), Taipei, Taiwan, 12–16 September 2016; pp. 433–442. [Google Scholar]
  43. Williams, T.; Kelley, C. Gnuplot 5.4: An Interactive Plotting Program. 2020. Available online: http://gnuplot.sourceforge.net/ (accessed on 7 July 2021).
  44. Bottou, L.; Lin, C.-J. Support vector machine solvers. Large Scale Kernel Mach. 2007, 3, 301–320. [Google Scholar]
  45. Li, X.; Fang, Z. Parallel clustering algorithms. Parallel Comput. 1989, 11, 275–290. [Google Scholar] [CrossRef]
  46. Chen, P.; Xie, H.; Maslov, S.; Redner, S. Finding scientific gems with Google’s PageRank algorithm. J. Inf. 2007, 1, 8–15. [Google Scholar] [CrossRef] [Green Version]
  47. Goel, A.; Munagala, K. Complexity measures for map-reduce, and comparison to parallel computing. arXiv 2012, arXiv:1211.6526. [Google Scholar]
  48. Tomita, E.; Tanaka, A.; Takahashi, H. The worst-case time complexity for generating all maximal cliques and computational experiments. Theor. Comput. Sci. 2006, 363, 28–42. [Google Scholar] [CrossRef] [Green Version]
  49. James, G.; Witten, D.; Hatie, T.; Tibshirani, R. An Introduction to Statistical Learning, 2nd ed.; Springer: New York, NY, USA, 2021. [Google Scholar]
Figure 1. A typical Spark cluster architecture.
Figure 1. A typical Spark cluster architecture.
Bdcc 05 00065 g001
Figure 2. Amdahl’s law for various serial factors and numbers of executors.
Figure 2. Amdahl’s law for various serial factors and numbers of executors.
Bdcc 05 00065 g002
Figure 3. Gustafson’s law for various percentages of serial work.
Figure 3. Gustafson’s law for various percentages of serial work.
Bdcc 05 00065 g003
Figure 4. A 2D plate’s homogeneous node communication.
Figure 4. A 2D plate’s homogeneous node communication.
Bdcc 05 00065 g004
Figure 5. Communication model based on fully connected graphs.
Figure 5. Communication model based on fully connected graphs.
Bdcc 05 00065 g005
Figure 6. Schematic diagram of the Hadoop cluster used in the experiment.
Figure 6. Schematic diagram of the Hadoop cluster used in the experiment.
Bdcc 05 00065 g006
Figure 7. Spark stages DAG of WC.
Figure 7. Spark stages DAG of WC.
Bdcc 05 00065 g007
Figure 8. Spark stages DAG of Kmeans.
Figure 8. Spark stages DAG of Kmeans.
Bdcc 05 00065 g008
Figure 9. Spark stages DAG of SVM.
Figure 9. Spark stages DAG of SVM.
Bdcc 05 00065 g009
Figure 10. Spark stages DAG of NWeight.
Figure 10. Spark stages DAG of NWeight.
Bdcc 05 00065 g010
Figure 11. Spark stages DAG of PageRank.
Figure 11. Spark stages DAG of PageRank.
Bdcc 05 00065 g011
Figure 12. Single executor runtime complexity with different sizes.
Figure 12. Single executor runtime complexity with different sizes.
Bdcc 05 00065 g012
Figure 13. Fitting the model to WordCount workload for amount of data.
Figure 13. Fitting the model to WordCount workload for amount of data.
Bdcc 05 00065 g013
Figure 14. Fitting the model to SVM workload with different dataset sizes.
Figure 14. Fitting the model to SVM workload with different dataset sizes.
Bdcc 05 00065 g014
Figure 15. Fitting the model to PageRank workload for amount of data.
Figure 15. Fitting the model to PageRank workload for amount of data.
Bdcc 05 00065 g015
Figure 16. Fitting the model to Kmeans workloads with different sizes.
Figure 16. Fitting the model to Kmeans workloads with different sizes.
Bdcc 05 00065 g016
Figure 17. Fitting the model to Graph (NWeight) workloads of different sizes.
Figure 17. Fitting the model to Graph (NWeight) workloads of different sizes.
Bdcc 05 00065 g017
Table 1. The recent approaches to Spark performance prediction.
Table 1. The recent approaches to Spark performance prediction.
ReferencesApproach/MethodSystem/Environments
Cheng et al. [27]Machine LearningEfficient performance prediction
for Apache Spark.
Ahmed et al. [34]Comprehensive
Trial-and-Error
Apache Hadoop and Apache Spark
for large scale datasets.
Al-Sayeh et al. [21]Gray-box modellingRuntime prediction of Spark jobs.
Shah et al. [33]PERIDOTQuick execution time predictions
for Spark applications.
Aziz et al. [28]Machine LearningResource management for
efficient performance of Apache Spark.
Gounaris et al. [24]Alternative
Systematic
Spark parameter tuning.
Mustafa et el. [10]Machine LearningPredicting execution time of Spark jobs.
Chao et al. [35]Gray-box modelling
(Machine Learning)
Spark performance model for
accuracy improvements.
Petridis et al. [8]Trial-and-ErrorSpark parameter tuning.
Table 2. Experimental configuration of the Hadoop cluster.
Table 2. Experimental configuration of the Hadoop cluster.
Server Configuration
Processor2.9 GHz
Main memory64 GB
Storage10 TB
Node Configuration
CPUIntel (R) Xeon (R) CPU E3-1231 [email protected] GHz
Main memory32 GB
Number of Nodes9
Storage6 TB each, 54 TB total
CPU cores8 each, 72 total
Software
Operating SystemUbuntu 16.04.2 (GNU/Linux 4.13.0-37-generic x86 64)
Hadoop2.4.0
Spark2.1.0
JDK1.7.0
Table 3. Spark HiBenchmark workload considered for this study.
Table 3. Spark HiBenchmark workload considered for this study.
Benchmark
Categories
Application                     Input Data SizeInput Samples
Micro
Benchmark
WordCountMultiple-Exec.Single-Exec.-
313 MB, 940 MB,
5.9 GB, 8.8 GB,
and 19.2 GB
3 GB, 5 GB, 7 GB,
10 GB, 12.8 GB,
14.4 GB, 16 GB,
18 GB, and 21.6 GB
Machine
Learning
Kmeans19 GB, 56 GB,
94 GB, 130 GB,
and 168 GB
1 GB, 38 GB, 75 GB,
113 GB, 149 GB,
and 187 GB
10, 30, 50,
70, and 90
(million samples)
SVM34 MB, 60 MB,
1.2 GB, 1.8 GB
and 2 GB
200 MB, 400 MB,
600 MB, 800 MB,
1.35 GB, 2 GB,
2.3 GB, and 2.5 GB
2100, 2600, 3600,
4100, and 5100
(samples)
Web SearchPageRank507 MB, 1.6 GB,
2.8 GB, 4 GB,
and 5 GB
100 MB, 250 MB,
750 MB, 6 GB,
7 GB, 8 GB, 9 GB,
and 10 GB
1, 3, 5, 7, and 9
(million of pages)
GraphNWeight37 MB, 70 MB,
129 MB, 155 MB,
and 211 MB
20 MB, 55 MB,
99 MB, 141 MB,
175 MB, 214 MB,
247 MB, 262 MB,
and 286 MB
1, 2, 4, 5, and 7
(million of edges)
Table 4. Workload application characteristics.
Table 4. Workload application characteristics.
WorkloadsStagesParallel StagesCollectSerializationDeserializationShuffleAggregate
WC2noyes--yes-
SVM209noyesnoyesyesyes
NWeight9yes-noyesyes-
Kmeans20noyesyesyesyes-
PageRank5no-noyesyes-
Table 5. Spark HiBenchmark parameters considered in this study.
Table 5. Spark HiBenchmark parameters considered in this study.
ParametersDefaultRangeDescription
Spark.executor.memory112Amount of memory to use per
executor process, in GB.
Spark.executor.cores12–14The number of cores to use on
each executor.
Spark.driver.memory14Amount of memory to use for
the driver process, in GB.
Spark.driver.cores13The Number of cores to use for
the driver process.
Spark.shuffle.file.buffer3248Size of the in-memory buffer for
each shuffle file output stream,
in KB.
Spark.reducer.maxSizeInFlight4896Maximum size of map outputs to
fetch simultaneously from each
reduce task, in MB.
Spark.memory.fraction0.60.1–0.4Fraction of heap space used for
execution and storage.
Spark.memory.storageFraction0.50.1-0.4Amount of storage memory immune
to eviction expressed as a fraction
of the size of the region.
Spark.task.maxFailures45Number of failures of any particular
the task before giving up on the job.
Spark.speculationFalseTrue/
False
If set to “true” performs speculative
execution of tasks.
Spark.rpc.message.maxSize128256Maximum message size to allow in
“control plane” communication,
in MB.
Spark.io.compression.codecsnappylz4/lzf/snappyCompress map output files.
Spark.io.compression.snappy.blockSize3232–128Block size in Snappy compression, in KB
Table 6. Time complexity for the workloads.
Table 6. Time complexity for the workloads.
WorkloadTheoretical Time ComplexitySingle Executor Best Fit f ( Size )
WordCount O ( N ) [47]linear
SVM O ( N 2 ) [44]quadratic
PageRank O ( n m ) [46]linear
Kmeans O ( N 2 ) or O ( N ) [45]linear
NWeight O ( N 2 ) or O ( N ) [48]quadratic
Table 7. Rsquared (Equation (17)) values for different models and workloads.
Table 7. Rsquared (Equation (17)) values for different models and workloads.
WorkloadBest FitEquations (15) or (16)Equations (15) or (16) (c = 1)Amdhal Equation (3)Gustafson Equation (5)Equations (8) or (9)Ernest [31]
Wordcountlinear0.9960.9960.9970.9960.9970.995
SVMquadrat.0.9170.9120.9060.8870.9170.847
PageRanklinear0.9900.9890.9900.9890.9900.988
Kmeanslinear0.9920.9920.9920.9930.9930.992
NWeightquadrat.0.9640.9640.9560.9650.9660.950
Table 8. Relative residual standard error (RRSE Equation (19)) values for different models and workloads.
Table 8. Relative residual standard error (RRSE Equation (19)) values for different models and workloads.
WorkloadBest FitEquations (15) or (16)Equations (15) or (16) (c = 1)Amdhal Equation (3)Gustafson Equation (5)Equations (8) or (9)Ernest [31]
Wordcountlinear0.0830.0830.0740.0820.0740.091
SVMquadrat.0.2710.2760.2850.3130.2710.367
PageRanklinear0.1160.1180.1130.1210.1130.127
Kmeanslinear0.1380.1370.1390.1310.1300.137
NWeightquadrat.0.1930.1930.2120.1900.1890.226
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ahmed, N.; Barczak, A.L.C.; Rashid, M.A.; Susnjak, T. An Enhanced Parallelisation Model for Performance Prediction of Apache Spark on a Multinode Hadoop Cluster. Big Data Cogn. Comput. 2021, 5, 65. https://doi.org/10.3390/bdcc5040065

AMA Style

Ahmed N, Barczak ALC, Rashid MA, Susnjak T. An Enhanced Parallelisation Model for Performance Prediction of Apache Spark on a Multinode Hadoop Cluster. Big Data and Cognitive Computing. 2021; 5(4):65. https://doi.org/10.3390/bdcc5040065

Chicago/Turabian Style

Ahmed, Nasim, Andre L. C. Barczak, Mohammad A. Rashid, and Teo Susnjak. 2021. "An Enhanced Parallelisation Model for Performance Prediction of Apache Spark on a Multinode Hadoop Cluster" Big Data and Cognitive Computing 5, no. 4: 65. https://doi.org/10.3390/bdcc5040065

APA Style

Ahmed, N., Barczak, A. L. C., Rashid, M. A., & Susnjak, T. (2021). An Enhanced Parallelisation Model for Performance Prediction of Apache Spark on a Multinode Hadoop Cluster. Big Data and Cognitive Computing, 5(4), 65. https://doi.org/10.3390/bdcc5040065

Article Metrics

Back to TopTop