The validation of the surrogate models is performed according to different situations that may be encountered in a real industrial plant. Then, we have several validation scenarios to compare the performance of the two surrogate models and determine which is the best model able to generalize in various regions within the range of the parameter space. It must be noticed that some of the validation scenarios are built using the same simulation conditions, rather than Training Set A or B. In these cases, it is unfair to compare the SModA performing in a validation set with the conditions of Training Set A, and the same happens for SModB and Training Set B conditions. However, this issue has been introduced with the purpose of validating the SModA in the training conditions of the SModB and vice versa. In this way, the generalization capability of a surrogate model in the prediction of unseen scenarios compared with an unfair prediction is remarked upon.
3.2.1. Single-Cycle Prediction
The final metal sheet quality and the die state are the most significant features after a press hardening process. The resulting temperatures of the sheet and the die provide this information. Additionally, the control of the maximum temperature of the die during the process ensures that the die has not exceeded its operational window. The simulation of a single cycle of a forming process provides these target variables as outputs. Hence, we expect the surrogate model to accurately predict the same target variables after a process without the need of the simulation, under different input conditions. Each of the three validation sets consists of 500 samples, which gives a total ratio of ∼1:8 with respect to the training sets.
Validation Scenario 1: = 10 s, 15 s and 20 s.
The validation set is formed by 500 randomly input samples obtained from simulations under the conditions of Training Set A. We do not include samples from the first three cycles in the set, since the surrogate model fed with Training Set A requires information about the three previous cycles to work. In this scenario, the initial die temperature varies between 80 and 150 °C, the cooling time have values of = 10 s, 15 s, and 20 s, and the forming times have a random value with the restriction of the cycle time s.
Figure 4 presents the surrogate models’ prediction values for the
versus the simulated values obtained from the simulation using the same inputs. The axis of the plot are divided into 50 bins to build a histogram of the distribution of
for both the predicted and simulated values, which act as our ground truth. This divides the space in the cells that are colored according the relative density of the samples compared to the cell with the maximum number of samples. For instance,
of relative density in a cell means that there are the same number of samples, rather than in the cell with the maximum number of samples. The figure additionally shows the empirical distributions of the simulated values (at the top of the figure) and the predicted ones (at the right side of the figure). In the ideal case, both distributions should be the same.
In
Figure 4a, we appreciate a narrow line following the diagonal of the plot, implying an almost perfect prediction from the SModA. Observing the empirical distributions, we see that the zones with more density correspond to the values of the
in the stationary regime for the cases
= 10 s, 15 s, and 20 s, as can be seen in
Figure 3a. The high prediction capability of the SModA in this validation scenario was expected, as training and validation sets share the same
values. On the other hand, despite the randomness in its training, the SModB is able to approach the diagonal line and also correctly captures the zones with more density, as it is shown in
Figure 4b. Nevertheless, the dispersion of the points indicates that the predictive power is lower than in the other case.
These features are repeated for the other target variables
and
. The respective figures of these variables are displayed in the
Appendix A. The results are condensed in
Table 3, where it is evidenced how the SModA outperforms the SModB in this particular validation case for all the target variables, although the SModB does not show very high values of the MAE. An error of about 2 °C is not unfeasible in experimental conditions, and can be often present due to systematic errors or calibration issues.
Validation Scenario 2: Intermediate values.
In this case, the validation set contains 500 randomly selected input points from simulations with intermediate values of , rather than the ones in Training Set A. The same as before, we do not add points from the first three cycles in the set, taking into account the limitation of the surrogate model trained on Training Set A. Then, the initial die temperature ranges between 85 and 145 °C, the forming times have values of = 11 s, 12 s, 13 s, 14 s, 16 s, 17 s, 18 s, and 19 s, and the forming time has a random value with the restriction of the cycle time s.
In the current validation scenario, the SModA does not perform as well as in the previous case. The intermediate values of
force the model to make an interpolation. In
Figure 5a, the points are distributed around the diagonal, although they form a line with a significant width, meaning more prediction error and SD. The SModB presents a narrower line around the diagonal, as it can be observed in
Figure 5b. We notice that the intermediate values of
cause a more uniform density distribution along the range of temperatures.
Table 4 shows the commented results of the SModB for the variable
. It must be noted that for the rest of the target variables, SModA has a lower MAE. Nonetheless, also taking into account
Figure 5, we consider that SModB is better in the prediction of the
of a next cycle than SModA in this parameter interpolation case, but observing the SD of both models, we observe that the overlap in the results makes it difficult to establish a clear option.
Validation Scenario 3: Random.
The validation set consists of 500 randomly sampled input points obtained from simulations under the conditions of Training Set B. Again, for the same reason as before, the first three cycles are not included in the set. In this case, the initial die temperature of the samples ranges between 50 and 165 °C, while the forming and cooling times range between s, with the restriction of the cycle time s.
In
Figure 6a, we see that in a random scenario the SModA performs poorly due to its lack of information about some regions of the parameter space. We observe a high dispersion of the points and the diagonal has nearly disappeared. Otherwise, as expected,
Figure 6b shows that the SModB maintains its good performance. With a few exceptions, almost all the points are condensed around the diagonal, meaning that the predictions are very close to the simulation values. The training under random conditions results in a high adaptability to any value of the input variables. Checking the other target variables in
Table 5, we confirm that SModB outperforms SModA in this more general scenario.
Summarizing, we identify that the SModA is able to carry out good predictions of the next cycle target variables in the exactly same training regimes, specifically, cases when s. However, the SModB achieves reasonably good performances in all the validation scenarios, showing a constant and controlled behavior.
3.2.2. Batch Prediction
Usually, in industrial manufacturing, the demand requires several press hardening processes to obtain a batch consisting of a specific number of parts. The simulation of this sequence of cycles is even more time demanding. Therefore, we evaluate the surrogate models in the prediction of the target variables for all the cycles in a batch. Since the objective is to effectively substitute the simulations, the surrogate model performs a sequence where the prediction of the next cycle is performed by taking as input the previous predictions.
For the reasons explained in
Section 2.1, the validation sets have
along the whole batch, corresponding to real experimental cases where the transference of the sheet into the die is automatized and the forming time can be changed within the range of values given by the total cycle time. The validation sets consist of 14 batches for each one of the values of
, which gives a total ratio of ∼1:2 with respect to the training sets.
Validation Scenario 4: Batches of = 10 s, 15 s, and 20 s.
The validation set consists of batches of 50 cycles, where the cooling time is kept constant within the entire batch and it has values of = 10 s, 15 s, and 20 s, the same ones as Training Set A. The forming time has a random value for each cycle, with the restriction of the cycle time s. For each value of , we have 14 batches for validation.
In
Figure 7, we compare how both surrogate models predict the target variables
, which defines the state of our system. The diagonal line acts as a reference of the perfect prediction. We can also observe the distribution of the simulated values and the predicted values in the histograms. Moreover, since we are evaluating batches of 50 cycles, the colors indicate the cycle of the prediction. Notice that the SModA is able to have a very good performance in this scenario. The reason is that it has been trained and finely tuned to those particular settings. The SModB predictions are shifted to higher values of
than our ground truth simulations, although the histograms are similar. The deviation from the diagonal becomes more evident in higher temperatures. In both
Figure 7a and
Figure 7b, the batches with different values of
can be identified, as higher values of
imply lower values of
. Quantitatively, the MAE between the predictions and the simulated values for all the data of this validation scenario is presented
Table 6, where the rest of target variables also have a very low value of MAE with the SModA.
Figure 8 represents the MAE and the cumulative MAE of the predictions of
for each cycle within the batches in the validation set. We observe how the SModA has a nearly perfect prediction for the previously commented reasons. Furthermore, the error of SModB is accumulated in the first cycles and after that it remains constant or even decreases. Additionally, we confirm that this model works better for higher values of
, i.e., for lower temperatures. The explanation can be found in the training sets. In
Figure 3a (where the parameters are the same as in the current validation set), the stationary region of the curves of
= 10 s is not reached until cycle 15, in which the stationary region achieves temperatures around 140 °C.
Figure 8 shows the important accumulation of error in the transient region for the SModB, and when
= 10 s, the transient region lasts more cycles. Besides, looking at Training Set B, the interval of temperatures around 140 °C in
Figure 3b is not very populated. These are the main causes of the loss in the predictive power of the SModB for low values of
. Summarizing, the transient region is the main source of error for the SModB, since the mean absolute error increases in the first cycles, while in the stationary region it is kept constant.
Validation Scenario 5: Batches of Intermediate.
In this case, the validation set is formed by batches of 50 cycles that have intermediate values of = 11 s, 12 s, 13 s, 14 s, 16 s, 17 s, 18 s, and 19 s, and that are kept constant along the cycles. Therefore, the SModA is not trained with the same values of cooling time. The forming time has a random value for each cycle, but it is restricted by s. For each value of , we have 14 batches.
In this scenario,
Figure 9a evidences the lack of generalization of the SModA. We notice that the distribution of the predictions displayed in the vertical histogram has peaks in the same ranges of temperatures as the ones of
Figure 7a. These ranges correspond to the stationary region of when
= 10 s, 15 s, and 20 s, implying that the SModA is not able to interpolate for intermediate values. In opposition, the predictions of the SModB present a similar temperature distribution to the simulated values. Additionally, comparing with
Figure 7b of the previous validation scenario, we found an analogous behavior of the SModB in this case, as shown in the distribution of the points of
Figure 9b. Focusing on
Table 7, the comparison of the two surrogate models shows the lower values of MAE of the SModB with respect to the SModA for all the target variables for the intermediate values of
. We notice that the values of MAE for the SModB are close to the ones in
Table 6, which implies a comparable performance in both validation scenarios. Then, despite the small shift with respect to the diagonal line and the dispersion observed in high temperatures, the SModB is more convenient if we are seeking a model capable of generalizing within the defined range of the operational parameters.
The better performance of the SModB and its generalization potential are verified in
Figure 10. Although the error increases in the first cycles, coinciding with the transient region, the SModB approaches the simulated values after that. On the contrary, the SModA error in the transient zone remains during the rest of the batch. As discussed, notice how the SModB works better for higher values of
. After the evaluation of the model performance in the different scenarios, we choose the SModB over the SModA because it has shown to be a more general model. In spite of the remarkable generalization capability of the SModB, the model is far from being perfect, especially if we focus on
Figure 10a. In the next section, we try to optimize the model performance and to reduce the computational time spent in the generation of the training set.