Next Article in Journal
Fractional Dynamics and Pseudo-Phase Space of Country Economic Processes
Next Article in Special Issue
Gated Recurrent Unit with Genetic Algorithm for Product Demand Forecasting in Supply Chain Management
Previous Article in Journal
Geometric Dynamics on Riemannian Manifolds
Previous Article in Special Issue
Using Dynamic Adjusting NGHS-ANN for Predicting the Recidivism Rate of Commuted Prisoners
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Clustering and Dispatching Rule Selection Framework for Batch Scheduling

Department of Industrial and Management Engineering, Hanyang University, Ansan 15588, Korea
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(1), 80; https://doi.org/10.3390/math8010080
Submission received: 3 December 2019 / Revised: 27 December 2019 / Accepted: 29 December 2019 / Published: 3 January 2020

Abstract

:
In this study, a batch scheduling with job grouping and batch sequencing is considered. A clustering algorithm and dispatching rule selection model is developed to minimize total tardiness. The model and algorithm are based on the constrained k-means algorithm and neural network. We also develop a method to generate a training dataset from historical data to train the neural network. We use numerical examples to demonstrate that the proposed algorithm and model efficiently and effectively solve batch scheduling problems.

1. Introduction

The aim of this study involves developing a framework that schedules batch jobs with identical machines to minimize total tardiness. The objectives of this framework include solving the problem in a short period, solving various batch scheduling problems, and improving performance as more problems are solved. We design a framework based on clustering and classification models because the machine learning approach fits well with the aforementioned goals.
Typical approaches for batch scheduling problems are categorized into heuristic, meta-heuristic, and dispatching rule approaches. The heuristic approach develops an algorithm that is problem-specific. The meta-heuristic approach modifies a meta-heuristic algorithm (e.g., to solve problems). The dispatching rule approach calculates the priority of each job or batch of jobs via a set of predetermined dispatching rules. The jobs are then processed in order of descending priority.
Heuristic and meta-heuristic approaches are usually dependent on the types of problems that are considered [1,2]. Since they are designed for a specific problem, they may yield relatively poor results when applied to different problems, or a substantial modification to the model is necessary before application. Additionally, a relatively long run time is sometimes necessary to yield a schedule using these approaches.
The dispatching rule approach is suitable for the aforementioned goals because it is relatively robust in terms of problem types, and thus can be applied to various problems without substantial modification to yield a schedule quickly. However, this approach occasionally encounters issues in selecting the optimal dispatching rule because it cannot outperform other rules in every scheduling situation [3]. Thus, an optimal universal dispatching rule is absent. Additionally, the optimal rule can change as successive jobs are completed, thereby implying that a dynamic dispatching rule can work better. Several studies developed classification models to select dynamic dispatching rules. For example, Mouelhi-Chibani and Pierreval [3] trained a neural network with dozens of queue state variables to select dispatching rules for flow shop scheduling. Similarly, Shiue [4] developed a support vector machine-based dispatching rule selection model for shop floor control system scheduling. Meanwhile, a few studies developed machine learning model to determine scheduling parameters. Mönch et al. [5] trains a neural network and inductive decision tree to estimate the weight of the ATC (apparent tardiness cost) rule, which consists of two dispatching rules: weighted shortest processing time rule and the least slack rule.
However, there is a paucity of extant studies on developing a dispatching rule selection model for the batch scheduling problem. This requires grouping jobs into batches before applying a dispatching rule. We design a clustering algorithm to group similar jobs under machine capacity constraints and add it to the framework.
The major contents and contributions of this study are as follows. First, we develop a clustering algorithm based on the constrained k-means algorithm to group jobs into batches. The algorithm uses a set of features impacting total tardiness and yields clusters that contain jobs wherein sums of job sizes do not exceed the machine load capacity. Second, we develop a step-by-step method based on the Monte Carlo Markov chain (MCMC) approach to create a training dataset to work with the dispatching rule selection model. Each sample in the dataset represents the virtual queue state via queue state variables (e.g., the number of batches in the queue and maximum processing time of the batches). Subsequently, the samples are attached to the optimal dispatching rule for the particular state. Third, we design a dispatching rule selection model based on neural networks for batch sequencing. Whenever a machine is available, queue state variable values for the current queue state enter the model to yield the optimal dispatching rule.
Section 2 introduces the batch scheduling problem including major assumptions, notations, and an overview of the framework. Section 3 describes methods to generate the training dataset to transform the variables to the queue status variables and to design the dispatching rule selection model. Section 4 discusses the clustering algorithm and application procedure and explains the model deployment process. Section 5 provides illustrative examples to show the application of the proposed model. Section 6 summarizes our conclusions and proposes future research directions.

2. Problem Statement, Notations, and Framework Overview

There are n jobs to be processed by one of m identical machines. Each job is characterized by job characteristics, namely processing time ( p ) , size ( s ) , due date ( d ) , and release time ( r ) . Jobs are grouped into batches, and a process sequence is determined via a dispatching rule (Figure 1). The former and latter stages are termed as grouping jobs and sequencing batches, respectively [6]. The sum of job sizes in each batch must be equal to or less than the load capacity of a machine. Each batch is processed by a machine as soon as it is available.
The batch processing time denotes the longest processing time among all jobs in the batch [7]. The objective of scheduling involves minimizing total tardiness as follows:
j = 1 n T j = j = 1 n max ( 0 , c j d j ) ,
where T j , c j , and d j denote tardiness, completion time, and due date of job j , respectively.
The major assumptions are as follows:
  • All job characteristics, ( p j ,   s j ,   d j ,   r j ) , of job j are known in advance.
  • Every job is ready for processing at its release time.
  • Neither setup nor rework is assumed.
  • Each machine only processes one batch at a time.
  • The characteristics data of several historical jobs are available and are sufficient to create a training dataset for the dispatching rule selection model.
Notations used in this study are as follows.
Indices
i
Batch index ( i = 1 ,   2 ,   ,   b )
j
Job index ( j = 1 ,   2 ,   ,   n )
t
Time index ( t = 1 ,   2 ,   )
k
Training sample index ( k = 1 ,   2 ,   ,   K )
Job characteristics
p j
Processing time of job j
s j
Size of job j
d j
Due date of job j
r j
Release time of job j
ξ j
Slack time of job j , i.e., the remaining time until due date after finishing the process, at its release time, ξ j = d j r j p j
ξ j t
Slack time of job j at time t , ξ j t = d j t p j
Batch characteristics
J ( i )
Set of job indices in batch i
p ( i )
Processing time of batch i , p ( i ) = max j J ( i ) ( p j )
Notations for clustering
δ
Load capacity of a machine
π j
Vector of job j ’s characteristics for grouping purposes, π j = ( p j ,   d j ,   r j )
C i
Centroid of batch i
E i , j
Distance between C i and π j
Notations for dispatching rule selection
y k
Dispatching rule for sample k ( k = 1 ,   2 ,   ,   K )
W k
Number of batches waiting in the queue in sample k , W k W m a x for all k
t k
System’s clock time with respect to sample k
f c
Probability mass function of c , where c either corresponds to processing time ( p ), size (s), slack time ( ξ ), or release time (r)
Ψ q
Candidate queue state variable ( q = 1 ,   2 ,   ,   Q )
S q
Selected queue state variable ( q = 1 ,   2 ,   ,   Q ) , Q Q
Figure 2 shows the step-by-step process that is used to obtain the set of job batches and the batch sequence for the machines to process. The framework consists of two tracks. Track I is termed as the model training track and is used to build a dispatching rule selection model via a training dataset of batches that is artificially created from historic data. Track II is termed as the scheduling track and is used to group the actual jobs into batches and to determine the processing sequence using the model provided by Track I.
A more detailed explanation of Track I is presented in Section 3, and Section 4 introduces Track II. Numbers inside the circle at the upper-left corner of the blocks denote the corresponding steps and subsections where the corresponding description are provided.

3. Model Training for Dispatching Rule Selection

3.1. Generating the Training Dataset

In this subsection, we introduce the step-by-step method used to create a dataset from historic data of job characteristics. The probability distributions of each job characteristic are calculated, and samples are then created to form a virtual batch dataset. This dataset is then transformed to a more adequate format for training.
Step 3.0: Prepare the probability distributions of job characteristics.
First, the pmfs (probability mass functions) f p , f s , f ξ , and f r of processing time ( p ), size (s), slack time ( ξ ) , and release time (r), respectively, are estimated from the historic data based on the relative frequencies as follows:
f c ( c ^ ) = Pr ( c = c ^ ) = N ( c = c ^ ) c ˜ N ( c = c ˜ ) ,
where c   { p ,   s ,   ξ ,   r } and N ( c = c ^ ) denote the number of jobs that satisfy c = c ^ .
Step 3.1: Create batches in the queue dataset.
With the pmfs, total K samples are generated where sample k   ( k = 1 , , K ) exhibits W k batches and each batch consists of one or more jobs characterized by job characteristics ( p ,   s ,   d ,   r ) . Specifically, W k is arbitrarily picked in the interval [ 1 ,   W m a x ] , where W m a x = j = 1 n s j δ . Next, the system’s clock time t k of sample k should exceed any release times of jobs in the queue ( max j Sample   k ( r j ) ). It also must be smaller than the maximum remaining processing time, and this is as follows. There are currently W k batches that remain in the queue, and thus it is inferred that W m a x W k batches are processed and the mean processing time of existing batches in the queue corresponds to i S a m p l e   k p ( i ) / W k . Therefore, t k is arbitrarily picked and satisfies max j S a m p l e   k ( r j ) t k ( W m a x W k ) × i S a m p l e   k p ( i ) W k .
Figure 3 shows a typical example of a set of K samples generated from estimated pmfs. The set of artificial samples is termed as the queue dataset because the batches in each sample wait in the queue to be processed.
Note: It is necessary to exercise caution when more than one job is created in a batch. When a job is sampled, then the next job should be “similar” in terms of its job characteristics ( p ,   s ,   d ,   r ) to that previously sampled because batches contain “similar” jobs. This is performed by using the following conditional pmfs as follows:
f c | c p r e ( c ^ | c ^ p r e ) = Pr ( c = c ^ | c p r e = c ^ p r e ) = N ( c ^   a n d   c ^ p r e   i n   c l u s t e r s ) N ( c ^ p r e   i n   c l u s t e r s ) ,
where c denotes one of the variables { p ,   ξ ,   r } , c p r e denotes the previously generated job’s characteristic, N ( c ^   a n d   c ^ p r e   i n   c l u s t e r s ) denotes the number of clusters containing jobs satisfying c = c ^ and c = c ^ p r e , and N ( c ^ p r e   i n   c l u s t e r s ) denotes the number of clusters that contain at least one job wherein c corresponds to c ^ p r e .
Algorithm 1 shows the generation process of the queue dataset explained above.
Algorithm 1. Queue dataset generation algorithm
Input K , W m a x , f p , f s , f ξ , f r , f p | p p r e , f ξ | ξ p r e , f r | r p r e
ProcedureFor ( k = 1 to K ) Do {
  Select W k from { 1 ,   2 ,   ,   W m a x } at random
  Initialize batch i as an empty set for i = 1 ,   2 ,   ,   W k
    For ( i = 1 to W k ) Do {
      job_size_sum = 0
      /* First job generation */
       p 1 f p , s 1 f s , ξ 1 f ξ , r 1 f r
       d 1 = r 1 + p 1 + ξ 1
      Append job 1 = ( p 1 , s 1 , d 1 , r 1 ) to batch i
      Increase job_size_sum by s 1
       j = 2
      /* job j 2 generation */
      While (True) Do { # while capacity constraint is satisfied
          p j f p | p p r e = p j 1 , s j f s , ξ j f ξ | ξ p r e = ξ j 1 , r j f r | r p r e = r j 1
          d j = r j + p j + ξ j
         If (job_size_sum + s j > δ ) Do {
         break}
      Increase job_size_sum by s j
      Append job j = ( p j , s j , d j , r j ) to batch i
      Increase j by 1}
    Append batch i to sample k in queue dataset}
   t k is randomly selected from [ max j S a m p l e   k ( r j ) , ( W m a x W k ) × i S a m p l e   k p ( i ) W k ]}
OutputQueue dataset
Step 3.2: Transform the variables.
Each sample of the queue dataset contains the values of system time ( t ) and job characteristics ( p , d , s , r ). It is necessary to transform them into the values representing the status of batches in the queue, as this facilitates improved training for the machine learning model. We introduce queue state variables, Ψ q   ( q = 1 , , Q ) , with which the current status of the queue is well defined and contributes to determining the optimal dispatching rule [3]. Queue state variables are defined by in extant studies [3,4,8] as listed in Table 1.
Variables Ψ 1 to Ψ 9 represent the batch state (e.g., number of batches and processing times of the batch) while Ψ 10 to Ψ 21 describe the job state (e.g., slack time of jobs). Moreover, the queue dataset consists of 21 queue state variables along with system time.
Step 3.3: Assign the optimal dispatching rule to each training sample.
This step determines the optimal dispatching rule y k for sample k   ( k = 1 ,   2 ,   , K ) of the batches in the queue dataset. The dispatching rules in Table 2 are considered.
We apply each rule to the batches in each queue dataset sample and then determine the optimal rule to minimize total tardiness. This is appended to each queue dataset sample and is used as a class label of the dispatching rule selection model.
Step 3.4: Select relevant queue state variables.
This step involves examining the queue state variables and removing those that are not relevant to the dispatching rule, y k . The relevance of a queue state variable is calculated, and variables with high relevance are selected as inputs in the training. We adopt ANOVA’s (analysis of variance) F-statistics for class-relevance, as suggested by [9]:
F = b e t w e e n g r o u p   v a r i a n c e w i t h i n g r o u p   v a r i a n c e ,
where the group denotes a set of samples with the same optimal dispatching rule.
Table 3 shows the structure of the queue dataset that is used as the training dataset for the dispatching rule, where { S 1 ,   S 2 ,   , S Q } { Ψ 1 ,   Ψ 2 ,   , Ψ 21 } denotes a set of (selected) relevant queue state variables.

3.2. Model Training

We develop a machine learning model that uses the values of the queue state variables to determine the dispatching rule, from which the processing order of actual job batches is derived. We select the artificial neural network because it is widely used and has yielded reliable results in previous research [10]. It should be noted that any type of classification model can be used in place of the neural network model.
Recall that, according to the steps of Section 3.1, values of the queue state variables Ψ i   ( i = 1 , 2 , , 21 ) are calculated from the generated dataset, and the optimal dispatching rule y is derived, which gives the minimum tardiness. We use the ANOVA to select the relevant queue state variables S 1 ,   S 2 ,   , S Q , and then ( S 1 ,   S 2 ,   , S Q ,   y ) composes one sample of the training dataset that will be an input to the neural network. Therefore, the neural network is trained against the dispatching rule for each given values of the queue state variables.
The neural network in our model exhibits an input layer, an output layer, and two hidden layers. Each node in the input layer corresponds to each queue state variable, and each node in the output layer matches the candidate dispatching rule. The number of nodes in the hidden layers are parameters that should be tuned. We test 45 models with nodes in each hidden layer ( h 1 , h 2 ) = ( 5 , 1 ) , ( 5 , 2 ) , , ( 5 , 5 ) , ( 6 , 1 ) , , ( 10 , 10 ) , where h 1 and h 2 denote the number of nodes in the first and second hidden layers, respectively, and we select the parameters exhibiting the optimal performance with five-fold cross validation.

4. Clustering Model to Group Jobs

Similar jobs are grouped together to form a group of batches when a set of jobs to be processed are newly entered. Jobs in each batch are simultaneously processed.

4.1. Clustering Model to Group Jobs

When jobs are processed in a batch and not individually, the completion time of the jobs are negatively impacted by the wasted processing time, j J ( i ) ( p ( i ) p j ) , and residual machine capacity of batch i , δ j J ( i ) ( s j ) . Therefore, minimizing both wasted processing time and residual machine capacity decreases the tardiness of the jobs [11]. Additionally, variance of the jobs’ due dates also increases the tardiness of the jobs. Thus, it is necessary to use the processing time ( p ) , due date ( d ) , and release time ( r ) while grouping the jobs into batches. Specifically, π j = ( p j ,   d j ,   r j ) denotes the vector of the job characteristics for clustering and is termed as clustering characteristics. Job size ( s j ) is considered as the clustering constraint.

4.2. Formations of Batches

Grouping jobs is based on the constrained k -means clustering algorithm. It consists of three steps, namely the centroid initialization step, job assignment step, and centroid update step as follows:
(i) Centroid initialization step:
The initial number of batches is obtained by dividing the total size of the jobs by the machine capacity (i.e., b = j = 1 n s j δ ), and the centroids C i for i = 1 ,   2 ,   ,   b are randomly initialized.
(ii) Job assignment step:
Each job is assigned to the nearest cluster. If there is a cluster that does not satisfy the machine capacity constraints, then one or more jobs in that cluster should be re-assigned. The algorithm identifies a job j * that is the farthest from the cluster’s centroid to which it belongs (e.g., cluster i ), albeit the closest to the next nearest cluster’s centroid, i.e., j * = argmax j ( E i , j / min i i ( E i , j ) ) ). It assigns job j * to the nearest cluster i * ( i ) among those satisfying the capacity constraint, i * = argmin i ( 1 E i , j * × s g n ( δ ( j J ( i ) ( s j ) + s j * ) ) ) .
(iii) Centroid update step:
The centroid of each cluster is updated.
The job assignment step and centroid update step are repeated until the clustering converges. If the clustering does not converge even after a certain predetermined number of times, then we increase the number of batches by 1, and the procedure is repeated until we have enough number of batches that accommodates all jobs. The following Algorithm 2 summarizes the clustering algorithm described above.
Algorithm 2. Clustering algorithm for grouping jobs
Input s j and π j = ( p j ,   d j ,   r j ) for all j , maximum iteration number M
Procedure/* Centroid initialization step */
b = j = 1 n s j δ # Initialize the number of batches
Select C i   ( i = 1 ,   2 ,   ,   b ) randomly
/* Job assignment step */
number of iterations = 1
Until (converged) Do {
  If (number of iterations > M ) Do {
    Iteration = 1
    Increase b by 1}
  For ( j = 1 to n ) Do {
    Assign job j to the nearest cluster i * = a r g m i n i ( E i , j ) }
  For ( i = 1 to b ) Do {
    While ( j J ( i ) s j > δ ) Do {
      Find job j * = argmax j ( E i , j min i i ( E i , j ) )
      Assign job j * to another cluster i * * , where
       i * * = argmax i ( 1 E i , j * × s g n ( δ ( j J ( i ) ( s j ) + s j * ) ) ) }}
  /* Centroid update step */
  Update C i as j J b π j | J ( i ) |
  Increase the number of iterations by 1}
OutputJob clusters (batches)
The following is an illustrative example of the clustering algorithm to improve the reader’s understanding.
Assume there are six jobs: π 1 = ( 1 ,   3 ) , π 2 = ( 2 ,   2 ) , π 3 = ( 4 ,   1 ) , π 4 = ( 5 ,   4 ) , π 5 = ( 3 ,   4 ) , and π 6 = ( 4 ,   2 ) as depicted in Figure 4. Each job is expressed as a two-dimensional vector for graphical convenience. The job sizes are s 1 = 1 , s 2 = 1 , s 3 = 1 , s 4 = 2 , s 5 = 2 , and s 6 = 3 , and the machine capacity is 4.
(i)
Centroid initialization step: Since j = 1 6 s j δ = 10 4 = 3 , we commence with three clusters (i.e., b = 3 ). Three centroids, namely C 1 = ( 3 ,   1 ) , C 2 = ( 5 ,   2 ) , and C 3 = ( 2 ,   3 ) (black squares in the figure), are randomly generated.
(ii)
Job assignment-1: every job is assigned to the nearest cluster. Cluster 1 = {job 3}, cluster 2 = {job 4, 6}, and cluster 3 = {job 1, 2, 5}.
(iii)
Job assignment-2: Cluster 2 does not satisfy the capacity constraint since s 4 + s 6 = 5 > 4 , and either job 4 or job 6 should be reassigned to another cluster. Given that E 2 , 6 min i 2 ( E i , 6 ) = 0.7071 > E 2 , 4 min i 2 ( E i , 4 ) = 0.6324 , we select j * = 6 . The total job sizes of cluster 1 and cluster 3 are 1 and 4, respectively; thus, job 6 is reallocated to cluster 1.
(iv)
Centroid update: each cluster centroid is updated as π 1 + π 2 + π 5 3 , π 4 , and π 3 + π 6 2 .
This process is repeated until the centroid of every batch is not updated any longer.

4.3. Model Deployment

After the jobs are grouped via the proposed clustering model, the rule selection model determines the batch to be processed whenever a machine becomes idle. First, the values of the selected queue state variables are calculated from batches, and the current time is set when the machine becomes idle. Next, the rule selection model selects the proper dispatching rule via the trained neural network. Finally, the selected rule calculates the priority of each batch, determines the highest priority, and processes it with the idle machine. Note that the selected rule may be different from the those selected by human experts because scheduling decisions of human and software models can be different from each other, and thus it is recommended not to rely only on the selected rule but to consider a human expert’s experience for practical usage as suggested in [12].

5. Experiment

This section illustrates the application procedure of the proposed framework to batch scheduling problems to minimize total tardiness where every job is released on the first day. We consider a test case with a small number of jobs and verify the proposed framework’s effectiveness by comparing it with the optimal schedule obtained via exhaustive search. Please refer to [13] for the algorithm to find the optimal solution of small batch scheduling problems.
Additionally, we show the validity of the proposed framework by comparing it with the schedules from single dispatching rules to solve large-scale realistic problems. Based on the experiment, we suggest that the proposed framework determines an optimal schedule for small problems and a better schedule than those obtained via a single dispatching rule for large problems.

5.1. Parameters

We assume that the pmfs and conditional pmfs (cpmfs) of job characteristics are estimated from the historical data listed in Table 4 and Table 5, respectively.
Each cell in Table 5 indicates Pr ( c = c ^ | c 0 = c ^ 0 ) . For example, the value 0.076 in Table 5 denotes the probability Pr ( p = 2 | p 0 = 1 ) = 0.076 , which denotes that the processing time of a newly generated job corresponds to 2 when the processing time of the previously generated sample corresponds to 1. It should be noted that we do not consider Pr ( r = r ^ | r 0 = r ^ 0 ) because every job is assumed to be released at once.
For the purpose of training, we first generated a training dataset with 100,000 samples as explained in Section 3.1. The optimal neural network was obtained by using this training dataset (which exhibits eight nodes in the first hidden layer and nine nodes in the second hidden layer), and the accuracy corresponds to 88.45%.

5.2. Comparison with Exhaustive Search

Table 6 illustrates the characteristics of 10 jobs to be processed via one of two identical machines wherein all capacities are 3.
We applied Algorithm 2 (Clustering algorithm for grouping jobs) to the jobs, and this resulted in five batches. A batch schedule via the trained neural network was established, and the total tardiness of this schedule was found to be 1.
Subsequently, this result was compared with that obtained by the exhaustive search. All possible schedules were generated by considering the processing order of batches as well as grouping jobs. This results in a total of 1,300,561,920 schedules. Each schedule’s total tardiness and maximum tardiness were calculated, and Figure 5 plots the distributions of them for all schedules.
There are 55,296 schedules (0.0042% among all the possible schedules) with a total tardiness of 1, and it is confirmed that the schedule obtained from the trained neural network is included in the 55,296 schedules.
We performed an additional experiment with problem sizes similar to those mentioned above to obtain the schedule with the minimum total tardiness via the proposed approach and then compare them with the optimal schedule. We randomly generated 1000 datasets that included 2, 3, and 4 machines with capacities of 3, 4, 5, and 6, and the numbers of jobs were 8, 9, 10, 11, 12, 13, 14, and 15. In the experiment, job characteristics were randomly generated based on Table 4. The results indicate that the proposed approach obtains the optimal schedules in 948 datasets out of 1000 datasets. Based on the result, we conclude that the proposed approach can establish the optimal schedule for small problems in most cases.

5.3. Comparison with Single Dispatching Rule

In this subsection, we compare the proposed framework with cases of single dispatching rule when there are hundreds of jobs, thereby resulting in a computationally intractable optimal batch schedule. The experiment involves demonstrating that the proposed method can be applied to a real-sized problem and exhibits an excellent performance.
We generated 100 test datasets via Table 4 and Table 5 where each dataset contained 300 jobs and there were four identical machines wherein all the capacities were 6. Subsequently, we applied the clustering algorithm to each dataset. The dispatching rule selection model was then applied to the batches, and the average total tardiness was calculated. This was compared with the cases wherein single dispatching rules (Shortest processing time first (SPT), Longest processing time first (LPT), Earliest minimum due date first (EMIDD), Earliest maximum due date first (EMADD), Earliest minimum release time first (EMIST), and Earliest maximum release time first (EMAST)) were adopted to the same datasets. Figure 6 shows the average total tardiness over all the datasets.
As shown in Figure 6, the proposed model yields the least average total tardiness among all dispatching rules, and this is approximately 4–40% less than the other models. Therefore, it is concluded that the proposed dispatching model yields a more efficient schedule when compared to that of the single dispatching rule, even with respect to a real-sized batch scheduling problem. The proposed model shows only a slight improvement compared to EMIST. This is because this rule seems to be appropriate for the test datasets and the proposed model selects the most appropriate dispatching rule among the pre-defined rules (i.e., SPT, LPT, EMIDD, EMADD, EMIST, and EMAST). It implies that the proposed model relies on the pre-defined rules, and therefore, it is important to consider various rules as pre-defined rules for practical usage.

6. Conclusions

The batch scheduling problem is one of the most important scheduling problems in the manufacturing process, and consists of two subproblems: grouping jobs into batches and determining the sequence of batches. Heuristic approach is a typical way to solve these problems because they are NP-hard (Non-deterministic Polynomial-time hard). However, the heuristic algorithm has the evident limitation that it needs to be designed differently according to the specific problem. In addition, much previous research has addressed only one of those two subproblems.
Motivated by these issues, in this study, we developed a batch scheduling model that is not problem specific and can solve both subproblems. The model consists of a clustering model and dispatching rule selection model to solve the problems of grouping jobs and sequencing batches during batch scheduling, respectively. The clustering model is developed based on the constrained k-means algorithm using a set of features impacting total tardiness (e.g., due date, processing time, and arrival time) and yields clusters that contain jobs wherein the sums of job sizes do not exceed the machine load capacity. The dispatching rule selection model is a supervised model that uses queue state variables (e.g., the number of batches in the queue and maximum processing time of the batches) as features and the most appropriate dispatching rule as the class label. The results of the experiment indicated that the framework solves batch scheduling problems and yields a good batch schedule.
With respect to future work, the proposed framework can be extended to solve diverse batch scheduling problems, for example, modification of the clustering algorithm to include machine capacities that are not identical wherein different types of objective functions are introduced. Additionally, more queue state variables (which cover various batch scheduling problems) can be adopted to make problems more practical.

Author Contributions

Funding acquisition, S.H.; Investigation, G.A.; Methodology, G.A. and S.H.; Software, G.A.; Validation, G.A.; Writing—original draft, G.A.; Review and editing, S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (2019R1A2C1088255).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vanchipura, R.; Sridharan, R.; Babu, A.S. Improvement of constructive heuristics using variable neighbourhood descent for scheduling a flow shop with sequence dependent setup time. J. Manuf. Syst. 2014, 33, 65–75. [Google Scholar] [CrossRef]
  2. Ahn, G.; Park, M.; Park, Y.J.; Hur, S. Interactive Q-Learning Approach for Pick-and-Place Optimization of the Die Attach Process in the Semiconductor Industry. Math. Probl. Eng. 2019, 2019, 4602052. [Google Scholar] [CrossRef]
  3. Mouelhi-Chibani, W.; Pierreval, H. Training a neural network to select dispatching rules in real time. Comput. Ind. Eng. 2010, 58, 249–256. [Google Scholar] [CrossRef]
  4. Shiue, Y.R. Data-mining-based dynamic dispatching rule selection mechanism for shop floor control systems using a support vector machine approach. Int. J. Prod. Res. 2009, 47, 3669–3690. [Google Scholar] [CrossRef]
  5. Mönch, L.; Zimmermann, J.; Otto, P. Machine learning techniques for scheduling jobs with incompatible families and unequal ready times on parallel batch machines. Eng. Appl. Artif. Intell. 2006, 19, 235–245. [Google Scholar] [CrossRef]
  6. Xu, R.; Chen, H.; Li, X. A bi-objective scheduling problem on batch machines via a Pareto-based ant colony system. Int. J. Prod. Econ. 2013, 145, 371–386. [Google Scholar] [CrossRef]
  7. Damodaran, P.; Srihari, K. Mixed integer formulation to minimize makespan in a flow shop with batch processing machines. Math. Comput. Model. 2004, 40, 1465–1472. [Google Scholar] [CrossRef]
  8. Shiue, Y.R.; Guh, R.S.; Tseng, T.Y. Study on shop floor control system in semiconductor fabrication by self-organizing map-based intelligent multi-controller. Comput. Ind. Eng. 2012, 62, 1119–1129. [Google Scholar] [CrossRef]
  9. Tsai, C.F. Feature selection in bankruptcy prediction. Knowl. Based Syst. 2009, 22, 120–127. [Google Scholar] [CrossRef]
  10. El-Bouri, A.; Shah, P. A neural network for dispatching rule selection in a job shop. Int. J. Adv. Manuf. Technol. 2006, 31, 342–349. [Google Scholar] [CrossRef]
  11. Chen, H.; Du, B.; Huang, G.Q. Scheduling a batch processing machine with non-identical job sizes: A clustering perspective. Int. J. Prod. Res. 2011, 49, 5755–5778. [Google Scholar] [CrossRef]
  12. Rodger, J.A. An expert system gap analysis and empirical triangulation of individual differences, interventions, and information technology applications in alertness of railroad workers. Expert. Syst. Appl. 2020, 144, 113081. [Google Scholar] [CrossRef]
  13. Bilyk, A.; Mönch, L.; Almeder, C. Scheduling jobs with ready times and precedence constraints on parallel batch machines using metaheuristics. Comput. Ind. Eng. 2014, 78, 175–185. [Google Scholar] [CrossRef]
Figure 1. Batch scheduling problem.
Figure 1. Batch scheduling problem.
Mathematics 08 00080 g001
Figure 2. Step-by-step process.
Figure 2. Step-by-step process.
Mathematics 08 00080 g002
Figure 3. Structure of samples of batches in the queue dataset.
Figure 3. Structure of samples of batches in the queue dataset.
Mathematics 08 00080 g003
Figure 4. Clustering algorithm example.
Figure 4. Clustering algorithm example.
Mathematics 08 00080 g004
Figure 5. Total and max tardiness distribution.
Figure 5. Total and max tardiness distribution.
Mathematics 08 00080 g005
Figure 6. Comparison results of the proposed model and single dispatching rule.
Figure 6. Comparison results of the proposed model and single dispatching rule.
Mathematics 08 00080 g006
Table 1. Queue state variables.
Table 1. Queue state variables.
VariableDescription
Ψ 1 Number of batches in the queue.
Ψ 2 , Ψ 3 , Ψ 4 , Ψ 5 Maximum, minimum, average, and standard deviation of the number of jobs in batches in the queue, respectively.
Ψ 6 , Ψ 7 , Ψ 8 , Ψ 9 Maximum, minimum, average, and standard deviation of the processing time of batches in the queue, respectively.
Ψ 10 , Ψ 11 , Ψ 12 , Ψ 13 Maximum, minimum, average, and standard deviation of the slack time at current time of jobs in the queue, respectively.
Ψ 14 , Ψ 15 , Ψ 16 , Ψ 17 Maximum, minimum, average, and standard deviation of the time remaining before the due date of the jobs, respectively.
Ψ 18 , Ψ 19 , Ψ 20 , Ψ 21 Maximum, minimum, average, and standard deviation of waiting times of the jobs, respectively.
Table 2. Considered candidate dispatching rules.
Table 2. Considered candidate dispatching rules.
Dispatching RuleBatch with the Highest Priority
Shortest processing time first (SPT) i * = argmin i = 1 , 2 , , n p ( i )
Longest processing time first (LPT) i * = argmax i = 1 , 2 , , n p ( i )
Earliest minimum due date first (EMIDD) i * = argmin i = 1 , 2 , , n ( min j J ( i ) d j )
Earliest maximum due date first (EMADD) i * = argmin i = 1 , 2 , , n ( max j J ( i ) d j )
Earliest minimum release time first (EMIRT) i * = argmin i = 1 , 2 , , n ( min j J ( i ) r j )
Earliest maximum release time first (EMART) i * = argmin i = 1 , 2 , , n ( max j J ( i ) r j )
Earliest minimum slack time first (EMIST) i * = argmin i = 1 , 2 , , n ( min j J ( i ) ξ j )
Earliest maximum slack time first (EMAST) i * = argmin i = 1 , 2 , , n ( max j J ( i ) ξ j )
Table 3. Structure of the training dataset for the dispatching rule selection model.
Table 3. Structure of the training dataset for the dispatching rule selection model.
Time S 1 S 2 S 3 S Q Optimal Rule
t 1 x 1 , 1 x 1 , 2 x 1 , 3 x 1 , Q y 1
t 2 x 2 , 1 x 2 , 2 x 2 , 3 x 2 , Q y 2
t K x K , 1 x K , 2 x K , 3 x K , Q y K
Table 4. Pmfs (probability mass functions) of the job characteristics.
Table 4. Pmfs (probability mass functions) of the job characteristics.
Value Pr ( p = Value ) Pr ( s = Value ) Pr ( ξ = Value )
10.300.300.25
20.400.400.25
30.300.300.25
40.000.000.25
Table 5. Cpmfs (conditional probability mass functions) of job characteristics.
Table 5. Cpmfs (conditional probability mass functions) of job characteristics.
(a) Processing Time(b) Slack Time
p ξ
1231234
0.9240.0760.001 ξ 0 0.8360.1460.0190.000
p 0 0.0480.9020.0500.1530.7280.0970.022
0.0010.0710.9280.0140.0980.8150.073
0.0030.0090.0750.913
Table 6. Job characteristics.
Table 6. Job characteristics.
j p j s j d j
1122
2315
3324
4123
5336
6214
7325
8324
9212
10336

Share and Cite

MDPI and ACS Style

Ahn, G.; Hur, S. Clustering and Dispatching Rule Selection Framework for Batch Scheduling. Mathematics 2020, 8, 80. https://doi.org/10.3390/math8010080

AMA Style

Ahn G, Hur S. Clustering and Dispatching Rule Selection Framework for Batch Scheduling. Mathematics. 2020; 8(1):80. https://doi.org/10.3390/math8010080

Chicago/Turabian Style

Ahn, Gilseung, and Sun Hur. 2020. "Clustering and Dispatching Rule Selection Framework for Batch Scheduling" Mathematics 8, no. 1: 80. https://doi.org/10.3390/math8010080

APA Style

Ahn, G., & Hur, S. (2020). Clustering and Dispatching Rule Selection Framework for Batch Scheduling. Mathematics, 8(1), 80. https://doi.org/10.3390/math8010080

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop