Artificial Intelligence for Ensemble of surrogates and cross-validation Engineering Design, Analysis and Manufacturing for rapid and accurate predictions using small data sets cambridge.org/aie Reza Alizadeh1, Liangyue Jia2, Anand Balu Nellippallil3, Guoxin Wang2 , Jia Hao2, Janet K. Allen1 and Farrokh Mistree1 Research Article 1 Systems Realization Laboratory, University of Oklahoma, Norman, OK, USA; 2School of Mechanical Engineering, Institute for Industrial Engineering, Beijing Institute of Technology, Beijing, China and 3Center for Advanced Cite this article: Alizadeh R, Jia L, Nellippallil Vehicular Systems, School of Mechanical Engineering, Mississippi State University, Starkville, MS, USA AB, Wang G, Hao J, Allen JK, Mistree F (2019). Ensemble of surrogates and cross-validation for rapid and accurate predictions using small Abstract data sets. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 33, In engineering design, surrogate models are often used instead of costly computer simulations. 484–501. https://doi.org/10.1017/ Typically, a single surrogate model is selected based on the previous experience. We observe, S089006041900026X based on an analysis of the published literature, that fitting an ensemble of surrogates (EoS) based on cross-validation errors is more accurate but requires more computational time. In Received: 19 January 2019 Revised: 22 August 2019 this paper, we propose a method to build an EoS that is both accurate and less computation- Accepted: 25 August 2019 ally expensive. In the proposed method, the EoS is a weighted average surrogate of response First published online: 18 October 2019 surface models, kriging, and radial basis functions based on overall cross-validation error. We demonstrate that created EoS is accurate than individual surrogates even when fewer data Key words: Ensemble of surrogates; kriging; response points are used, so computationally efficient with relatively insensitive predictions. We dem- surface modeling; small data sets; surrogate onstrate the use of an EoS using hot rod rolling as an example. Finally, we include a rule-based models template which can be used for other problems with similar requirements, for example, the computational time, required accuracy, and the size of the data. Author for correspondence: Guoxin Wang, E-mail:

[email protected]

Frame of reference Computer simulations are commonly used to replace experiments with physical models. Often these simulations are computationally expensive. However, many model-based engineering design problems require numerous simulations to reach an acceptable solution. This can be computationally prohibitive. Often a single surrogate model – or metamodel – is used to replace a detailed simulation in design problems which require repeated calculations. This surrogate is obtained using infor- mation derived from the physical model. Many types of surrogate models have been proposed. Here, we demonstrate the advantages of computing ensemble of surrogates (EoS) from a single set of data and then averaging these surrogates to make use of the good features of each type of surrogate. We term these assemblies of surrogates, ensembles of surrogates. The EoS is built using the weighted average of different individual surrogates. These weights can be calculated randomly or in a systematic way. For instance, Viana and Haftka (2008) use a systematic weighted average surrogate (WAS) to utilize the advantage of n surrogates to cancel the errors in estimation by appropriate weight selection in the linear mix of the models. They use an ensemble of metamodels to minimize the root-mean-square error (RMSE) in surrogate mod- eling. They discuss using the lowest predicted residual error sum of squares (PRESS) solution or just an average surrogate when an individual surrogate is required. They also propose the optimization of the integrated squared error (SE) as an approach to calculate the weights of the WAS. They found that it is worthwhile to create a broad set of various surrogates and after that apply PRESS as the selection criterion. Cross-validation is utilized broadly to allocate the weights to individual surrogates in build- ing an EoS in a systematic way. Viana et al. (2009) use cross-validation to approximate the necessary safety border for a particular favorite conservativeness degree (safe approximations percentage). They also check how well they can reduce the loss of accuracy caused by a con- servative estimator1 by choosing among alternative surrogate models. They show that cross- validation enables to choose the best conservative surrogate model with the lowest accuracy loss. Also, they found that it is efficient in determining safety is effective for selecting the safety edge. Also, Viana et al. compare using the lowest PRESS with a weighted surrogate when an © Cambridge University Press 2019 individual surrogate is required. 1 The estimator is the PRESS and obtained by dividing the N data points into k subsets in cross-validation process. It is esti- mated by using a subset of points in building the surrogate and computing the errors at these left out points. Then, this process is replicated with various sets of left-out points to obtain PRESS which is statistically significant. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 485 They propose optimizing the incorporated SE as an approach find that using an EoS through weighted averaged surrogates pro- to calculate the weights of the WAS model. They find that it is bet- vides more robust estimation than single ones. ter to create a big set of various surrogate models and choose the Some authors analyze the fidelity of the estimation functions best based on the PRESS, and that the error of cross-validation modeling in surrogate-based optimization in engineering design. provides a great approximation of the RMSE if enough data points Bellary and Samad (2017) address this issue using the EoS to sug- are used. However, in high dimensions, the advantages of using gest estimations from alternate modeling schemes. Also, Habib the cross-validation error and weighted surrogates are decreased et al. (2017) use an EoS-assisted optimization method and evalu- considerably. Also, Goel et al. (2007) create a systematic heuristic ate it at various levels of fidelity. Yin et al. (2018) propose assem- process for computation of the weights as the PRESS weighted bling an EoS by dividing the design space into several subspaces average surrogate (PWS). Using a combination of neural net- such that each is allocated a collection of optimized weights. works, Bishop (1995) create a systematic WAS gained by estimat- Acar (2015) argues for giving greater importance to maximum ing the covariance between surrogates from residuals at test or error than RMSE by assigning weights of the individual surrogates training data sets. Following Bishop’s method, Acar and in the EoS. In this paper, weights for the surrogate models in the Rais-Rohani (2009) develop another approach to optimizing the EoS are selected to minimize the root-mean-square cross- mean square (MS) error. validation error (RMSE-CV) in a hope to minimize the original As shown in Table 1, we critically evaluate the surrogate mod- RMSE. Additionally, some studies are specifically focused on eling literature based on the type of research, the number of the the EoS of just one type of metamodel. For instance, Shi et al. combined surrogates, weight assignment process, and comparison (2016) introduce a combination of radial basis functions (RBFs) criteria. to determine the weights by solving a quadratic programming Experimental ensembles of surrogates calculated from a single (QP) subproblem. The results show that an ensemble of multiple set of data points can be used to overcome the weaknesses of every RBFs can remarkably enhance the modeling efficacy compared to single type of surrogate. For instance, Song et al. (2018) study the single RBF models. efficiency of using an EoS in improving accuracy and the robust- Many authors show the use of the EoS in evolutionary algo- ness for several problems. Robustness is the ability of the model to rithms. Lim et al. (2007) study the search efficiency of various have low fluctuation in accuracy in different situations (e.g., with a surrogate modeling methods and the EoS in a memetic surrogate- low and high amount of data). They use an integrated ensemble based technique. Bellary et al. (2016) use EoS integrated into a GA surrogate model to (i) filter out the individual models with low to acquire POFs. They realize that the WAS EoS has better perfor- performance and retain the higher-performing ones using cross- mance for both the goals than a single metamodel. Bhat et al. validation errors and (ii) calculate the appropriate weighting for (2010) use EoS and nondominated sorting genetic algorithm II each surrogate model included in the ensemble based on the ref- (NSGA-II) to optimize a simultaneous structure control design erence model and the approximated mean square error (MSE). Xu strategy. They found that by introducing a new weighting and Zeger (2001) use an EoS and introduce two independent pro- approach as a frequency-dependent function, it is possible to cesses to highlight their advantages instead of individual surro- minimize closed-loop measures for optimal performance by gates. On the other hand, Zhou et al. (2018) study the searching over the design space encompassing the open-loop drawbacks of compound and ensemble surrogates and their dynamic and controller variables while keeping constraints. inadequacy based on quasi-concavity. Samad et al. (2006) analyze Arias-Montano et al. (2012) use ensembles of surrogates com- the use of an EoS and performance approximation bined with an evolutionary method to obtain the benefit of simultaneously. their advantages for solving expensive multiobjective optimization Ensembles of surrogates are often used in the surrogate- problems. assisted design. For instance, Viana et al. (2013) utilizes an Bhattacharjee et al. (2016) introduce a multiobjective evolu- ensemble of surrogate modeling approach when adding more tionary algorithm embedded with different surrogates which are than one point in each optimization iteration. Existing global spatially distributed. They use a nondominated sorting GA as optimization algorithms may be revised to find multiple alterna- the underlying optimizer. They extract the best features of differ- tive designs; however, parallel computation is the key to increasing ent strategies and show that the multiple surrogates-assisted mul- optimization efficiency (Villanueva et al., 2013; Chaudhuri tiobjective optimization with local surrogates with improved and Haftka, 2014). Bhattacharjee et al. (2018) implement an pre-selection offers better performance than individual surro- EoS-assisted multiobjective optimization for engineering design gates. The same group, in another study, compares ensembles problems which are computationally costly. Also, Chaudhuri of surrogates in surrogate-assisted multiobjective optimization and Haftka (2014) use an EoS to compute Pareto optimal fronts algorithm (SAMO) with NSGA-II. They find that SAMO consis- (POFs). Adhav et al. (2015) utilize an EoS to decrease uncertainty tently outperforms NSGA-II (Bhattacharjee et al., 2016). Lv et al. in searching for an optimal point. Liu et al. (2015) utilize an EoS (2018) use an integrated framework of an EoS in particle swarm with a genetic algorithm (GA). Badhurshah and Samad (2015) optimization (PSO), which includes inside and outside optimiza- find that using an ensemble of surrogate-assisted optimization tion loops. In the outside optimization loop, a PSO algorithm is methods and computational fluid dynamics analysis, the optimal- used for both the sampling and the optimization approaches. In ity, efficiency of the optimization process, and the robustness of the inside optimization loop, an EoS-assisted parallel optimization the optimum solutions can be improved. Wang et al. (2016) use approach is implemented. They show that their framework can an EoS for global optimization to enhance the convergence converge to a good solution for nonconvex, multimodal, and low- ratio of an uncertainty predictor. Samad et al. (2006) evaluate dimensional problems. Khademi et al. (2013) use it to find the the performances of ensembles of surrogates in a turbomachinery best location for speed bump installation using experimental blade-shape optimization. They use response surface models design methodology. Ezhilsabareesh et al. (2018) use an (RSMs), kriging (KRG), radial basis neural network (RBNN), EoS-based multiobjective optimization approach using RMS, and weighted average models to test shape optimization. They KRG, RBNN, and NSGA-II. They find that among the surrogates, 486 Reza Alizadeh et al. Table 1. Critical evaluation of the EoS literature Paper Features Type of research Number of Method of specifying weights for the Criteria used to compare the (theoretical/ combined ensemble surrogates (systematic/ methods (time/size/accuracy) experimental) surrogates random) Accuracy Time Size Mack et al., 2005 Experimental 3 Random Samad et al., 2006 Theoretical 4 Random Samad et al., 2006 Experimental 3 Random Goel et al., 2007 Theoretical 3 Systematic Lim et al., 2007 Theoretical 4 Random Samad et al., 2007 Experimental 3 Systematic Viana and Haftka, Theoretical 4 Systematic 2008 Viana et al., 2009 Theoretical 4 Systematic Bhat et al., 2010 Experimental 4 Random Viana et al., 2010 Theoretical 4 Systematic Arias-Montano et al., Theoretical 5 Systematic 2012 Basudhar, 2012 Theoretical 5 Random Viana et al., 2013 Theoretical 10 Systematic Villanueva et al., 2013 Theoretical 4 Systematic Chaudhuri and Theoretical 4 Systematic Haftka, 2014 Chaudhuri and Experimental 4 Systematic Haftka, 2014 Acar, 2015 Theoretical 4 Systematic Adhav et al., 2015 Experimental 3 Systematic Badhurshah and Theoretical 2 Systematic Samad, 2015 Chaudhuri et al., 2015 Theoretical 2 Systematic Liu et al., 2015 Theoretical 24 Systematic Babaei and Pan, 2016 Experimental 9 Systematic Alizadeh et al., 2016 Experimental 4 Systematic Bellary et al., 2016 Experimental 2 Systematic Beynaghi et al., 2016 Theoretical 4 Systematic Bhattacharjee et al., Experimental 4 Random 2016 Qiu et al., 2016 Experimental 4 Random Shankar Experimental 7 Random Bhattacharjee et al., 2016 Shi et al., 2016 Theoretical 4 Systematic Kaleibari et al., 2016 Theoretical 4 Systematic Wang et al., 2016 Theoretical 4 Systematic Bellary and Samad, Theoretical 3 Systematic 2017 Habib et al., 2017 Theoretical 5 Systematic Theoretical 4 Systematic (Continued ) Artificial Intelligence for Engineering Design, Analysis and Manufacturing 487 Table 1. (Continued.) Paper Features Type of research Number of Method of specifying weights for the Criteria used to compare the (theoretical/ combined ensemble surrogates (systematic/ methods (time/size/accuracy) experimental) surrogates random) Accuracy Time Size Bhattacharjee et al., 2018 Zamani Sabzi et al., Theoretical 4 Systematic 2018 Ezhilsabareesh et al., Experimental 6 Systematic 2018 Lv et al., 2018 Theoretical 5 Systematic Song et al., 2018 Theoretical 4 Systematic Viana et al., 2008 Theoretical 10 Systematic Song et al., 2018 Theoretical 3 Systematic Yin et al., 2018 Theoretical 3 Systematic Table 2. Trade-offs among three criteria Situation Detail trade-off If computational time The larger the size (complexity) of the is fixed problem, the lower the accuracy If the problem size is The higher the accuracy, the greater the fixed computational time If the desired accuracy The larger the size (complexity) of the is fixed problem, the greater the computational time RSM delivers lower PRESS and has the best PRESS for all of the objectives. They also find that RMS produces the lowest RMS error after evaluation by NSGA-II. In summary, building an EoS results in higher accuracy in many cases, but it is more computationally intensive than using individual surrogates. In this paper, we address this gap in the published literature on creating a less computationally intensive EoS. Fig. 1. Triangle showing the relationships among three design criteria (Alizadeh et al., 2019). Ensembles of surrogates and cross-validation As shown in Table 1, two types of studies have been done, namely is defined as the amount of the sample data required, and the experimental and theoretical. Also, four surrogates are often com- desired accuracy is evaluated by the deviation of the predicted bined in order to build an EoS. Additionally, the weight assign- response of the surrogates from the response of the simulation. ment process has changed from a random to a systematic The interactions among the three criteria are shown in Figure 1. procedure. Finally, almost all studies use the only accuracy as the criteria for comparing the performance of the EoS with Problem description each individual surrogate model. However, more appropriate comparison criteria are computational time, size, and accuracy To test our hypothesis that an EoS is more accurate than those and an understanding of a method for making trade-offs determined using individual RSM, KRG, RBF surrogates, we among these attributes (Alizadeh et al., 2019). A qualitative choose a hot rod rolling problem as an example. In this example description of the trade-offs among these three criteria is shown problem, our interest is to accurately predict the microstructure of in Table 2. the final rod product using surrogate models such that the forma- The computational time represents the sum of the program tion of banded microstructure (the microstructure with alternate running time, which is used to construct the surrogate model, layers of ferrite and pearlite occurring due to the presence of and the simulation time used for sampling. So, in this work, microsegregates, thereby affecting the mechanical properties) the time is T = Tprogram + Tsimulation. Also, here the problem size can be managed. The accuracy of the surrogate modeling process 488 Reza Alizadeh et al. Fig. 2. Hot rod rolling process (Nellippallil et al., 2017). In prior studies, authors report the effects of banded micro- structures on the mechanical properties of final products (see Spitzig, 1983; Tomita, 1995; Krauss, 2003; Korda et al., 2006). Hence, it is critical to properly predict the microstructure after phase transformation to manage the banded microstructure such that the mechanical properties of the product can be con- trolled. The ferrite–pearlite banded microstructure is primarily caused due to microsegregates in the form of alloying elements of manganese, sulfur, etc., that are embedded into the steel dur- ing the solidification process after casting. The initial AGS, cool- ing rate, carbon concentration, and manganese concentration are selected in this study based on literature review as the major fac- tors influencing the final microstructure phase formed after roll- ing and cooling. The program STRUCTURE developed by Jones and Bhadeshia (2017) is used as the simulation program to pre- dict the microstructure phases. The other input fixed parameters for the simulation program like the austenite–ferrite interfacial Fig. 3. An example of a banded microstructure in 1020 steel consisting of ferrite (light) and pearlite (dark) (Jägle, 2007). energy, activation energy for atomic transfer, aspect ratio for nucleation of ferrite, fraction of effective boundary sites, total volume fractions of inclusions, nucleation factor for pearlite is hence important to properly estimate the final microstructure and aspect ratios of growing allotriomorphic ferrite, produced. As shown in Figure 2, in the hot rod rolling process Widmanstatten ferrite, and pearlite are selected and defined problem that we are addressing, a billet of square cross-section based on the values reported by Jägle (2007). The control factors having an initial austenitic microstructure is rolled using rollers thus considered in this work are described in Table 3. The out- and further cooled at the run-out table to change the shape, put of the simulation includes volume fractions of pearlite and microstructure, and mechanical properties of the rod. two types of ferrites, namely allotriomorphic ferrite and In this problem, we are only considering the phase transforma- Widmanstätten ferrite for the different values of each of the tion of austenite to the two phases, ferrite and pearlite. We four input variables. Therefore, the simulation addressed in assume in this problem that the phase transformation of austenite this problem involves four input variables and three output to ferrite and pearlite occurs only during the cooling stage after variables. hot rolling. Two types of ferrite phases, namely allotriomorphic A fractional factorial design of experiments (DOE) to generate ferrite and Widmanstatten ferrite, are considered in this problem. response data sets is carried out by Nellippallil et al. (2018) using A slow cooling rate favors the formation of banded ferrite/pearlite the simulation program, STRUCTURE developed by Jones and microstructure as there is enough time for carbon diffusion and Bhadeshia (1997). Polynomial RSMs are fit for each of the ferrite (allotriomorphic ferrite mostly) nucleation. responses. Nellippallil et al. explain how different values of the Suppressing banding is possible via fasting cooling rate, but the four input variables, cooling rate, carbon concentration, manga- elimination of microsegregates, the source for banding, is not pos- nese concentration, and AGS affect the final microstructural sible. In addition to the cooling rate, the initial austenite grain size phases (pearlite, allotriomorphic ferrite, and Widmanstätten fer- (AGS), percentages of carbon and manganese are important vari- rite) based on the polynomial RSM developed (Nellippallil et al., ables in the phase transformation process. For example, a small 2018). Nellippallil et al. verify the model predictions by compar- AGS facilitates the phase transformation to ferrite. A small AGS ing them with experimentally measured data reported by Bodnar supports the increase of grain boundary area per volume available and Hansen (1994). We use the fractional factorial DOE data set for nucleation resulting in more allotriomorphic ferrite nuclei and by Nellippallil et al. in this work. In Figure 4, we show the com- smaller ferrite grain sizes. This occurs at regions of low manganese parison of polynomial RSM predictions by Nellippallil et al. concentrations. The rest of austenite transforms to pearlite phase. A (2017) with the phase fractions reported in the literature by high AGS, however, results in an increase in Widmanstatten ferrite. Bodnar and Hansen (1994). We observe from Figure 4 that the A banded microstructure of ferrite and pearlite formed due to the model predictions lie in the vicinity of the straight line depicting presence of microsegregates is shown in Figure 3. the measured values. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 489 Table 3. The design variables Design variable Definition Manganese (Mn) concentration Jones and Bhadeshia (1997) point out that manganese is an austenite stabilizing agent. Therefore, transformation to after rolling ferrite occurs at low manganese regions. Due to this, the high manganese region gets enriched with carbon leading to the formation of pearlite. Final AGS after rolling This parameter has an inbuilt effect on grain boundary area per unit volume and thus on nucleation itself. Because of this effect, and the simultaneous phase transformations, the average grain size (neglecting the length scale) have a major bearing on the final microstructure. Cooling rate Banding is usually suppressed by high cooling rates. Lower cooling rates favor carbon diffusion leading to the development of banded microstructure. Carbon content The carbon content changes the physical properties of commercially available steel, and hence determine which component is formed first during the initial stages of cooling. Fig. 4. Comparison of polynomial RSM predictions (Nellippallil et al., 2017). Method prediction data and the test data (in the Prediction metric: root-mean-square error section). As shown in Figure 5, the procedure consists of several steps. The first step is to find some criteria to choose the most appropriate surrogate model. Based on a comprehensive critical literature DOE and cross-validation review, a balanced triangle of three important characteristics of In this work, we use the data generated by Nellippallil et al. (2018) the experiment (accuracy, size, and time) are presented in using a fractional factorial design (Montgomery, 2017) to orga- Figure 1 (Alizadeh et al., 2019). The second step is to build sur- nize the experiments. Considering the trade-off between the rogate models, including RSM, KRG, RBF, and an ensemble of cost of evaluation of the original simulation and the increase of them. In this step, we compare the performance of these models fidelity associated with the increasing number of sampling data, for the hot rod rolling problem. In the third step, the outcome of three levels including the upper, lower, and middle points of the second step is summarized in a sort of database to be used in the range in each factor are selected to manage the sampling pro- the future. cess. The factors and factor levels are shown in Table 4. Now, in order to estimate the performance of the predictive model, K-fold cross-validation is used. K−1 equally sized ran- Surrogate modeling process domly selected subsamples are used as training data and the The method for selecting the surrogate model based on time, size, remaining single subsample is retained to test the model. The and accuracy are illustrated in Figure 5. After identifying the prob- training process for each fold is repeated K−1 times, with each lem in the Problem description section, the next step is to gener- subsample being used exactly once as test data. Thus, all observa- ate sampling data. The DOE is used to obtain the sampling data tions are fully used for both training and testing. We repeat the over the desired range of input variables. In the second step, whole cross-validation process ten times, and the result is the cross-validation (CV) is used to divide the design data into “train- average value of ten runs of cross-validation. In this way, the ing data” and “testing data” (Fig. 6). The “training data” is used to impact of noise can be minimized. The next step is to use the develop different surrogate models, and the “testing data” is used training data to construct surrogates, in this case, RSM, KRG, as unknown data to estimate the performance of different surro- RBF, and an EoS and use the testing data for evaluation. gate models. Next, different surrogate modeling methods are used to fit the training design data generated and develop a predictive Function fitting model (in the Function fitting section). Finally, the surrogates are Several function approximation techniques are used as surrogates. evaluated using the RMSE of the response deviation between the Shyy et al. (2001) used RSM and RBF to rocket engine design and 490 Reza Alizadeh et al. Fig. 5. Method for selecting the surrogate model based on time, size, and accuracy. compare the prediction of alternative models. It turns out surro- number or a linear polynomial function. And g(x) is a localized gate models have good performance in prediction work. Zerpa deviation function or can be called the basic function. In this arti- et al. (2005) integrate RSM, KRG, and RBF as an ensemble of cle, the focus is on one type of KRG called Ordinary Kriging the surrogate model and apply it into an alkaline–surfactant– (Kleijnen, 2017). polymer flooding processes. Bellucci and Bauer (2017) use RSM, KRG, and RBF to make robust parameter design. So, in ŷ(x) = m + g(x), (3) this work, we choose RSM, KRG, and RBF, which are commonly used in the previous papers, and their combination as an ensem- where μ is an unknown constant, which represents the simulation ble of surrogate models to develop different prediction models. output averaged over the experimental area, and g(x) is a zero- Response surface method. The RSM is also known as the poly- mean stochastic process. nomial regression method which has the simplest of parameters Radial basis function. The RBF technique is based on a math- (i.e., coefficients in a polynomial function) and is calculated ematical function and its value is calculated based on the distance using least squares regression (Razavi et al., 2012). In this work, the between origin and each point (Montgomery, 2017). we use a second-order polynomial function as the surrogate Alternatively, the distance between the center point and each model. point can be used as shown in the following equation: If x is an independent vector of factors, y is the vector of responses, the impact of x on y and their relationship can be illus- trated as follows: ri, j = r(xi , xj ) =xi − xj, (4) n n where ri,j denotes the Euclidean distance between two different ŷ(x) = a0 + ai xi + aii xi2 + aii xi xj , (1) sample points. Radial functions are employed to connect the dis- i=1 i=1 i,j tance r with the outputs, then the integration of these functions is used to estimate complicated mathematical functions. These func- where α represents the coefficients of the polynomial function, tions are then used in constructing the surrogate models: and n is the number of independent factors. Kriging. Kriging is a type of interpolating technique which is a M polynomial model of an input vector of factors x, f (x), and loca- ŷ(xnew ) = ri Q(xnew − xi). (5) lized deviation of x, g(x), as follows: i=1 ŷ(x) = f (x) + g(x), (2) The surrogate function ŷ(xnew ) is an integration of M RBFs and each of them is linked to a distinct xi and has a weight of f (x) is the polynomial term, which is a global function over the ri (Mirjalili, 2019). In this work, the Gaussian function is selected entire input space (Razavi et al., 2012). Usually, f (x) is a constant as the RBF. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 491 Fig. 6. The experimental procedure. Table 4. Factors and factor levels for DOE NSM Ei Level Cooling rate (K/min) AGS (μm) C (%) Mn (%) Eavg = i , (7) NSM 1 11 30 0.18 0.7 2 55 55 0.24 1.1 where Ei denotes the GMS error, NSM is the number of sample points, and two constraints, α, β are chosen to be α = 0.05 and 3 100 100 0.3 1.5 β = −1, respectively (Samad et al., 2006). Ensemble of surrogates model. We use the weighted average model proposed by Goel et al. (2006) to create the EoS. As we Prediction metric: root-mean-square error just use three individual surrogate models, the predicted response In evaluating the process, the actual objective function value for of the ensemble of surrogate model is the testing data is known and regarded as the target value. This value is then used to calculate the error at all testing points. NSM The RMSE is used as the prediction metric to assess the accuracy ŷEoS = wi ŷi = wRSM ŷRSM + wKRG ŷKRG + wRBF ŷRBF , (6) of each prediction model generated by different surrogates. The i definition of RMSE is where wi is the weight of each individual surrogate, and ŷi is the predicted response of the ith individual surrogate. Ntest i=1 (yi − ŷ i ) 2 For the selection of weights, a strategy which is based on gen- RMSE = , (8) eralized mean square cross-validation error (GMS error) is pro- Ntest posed in Goel et al. (2006). where Ntest represents the number of testing points, yi, and ŷi are w∗ the actual response and the predicted value of the response from a wi ∗ = (Ei + aEavg )b , wi = i ∗ , surrogate. The lower the RMSE, the greater the accuracy of the wi i surrogate model, and vice versa. 492 Reza Alizadeh et al. Fig. 7. RMSE in the prediction of surrogates in ten runs. Table 5. Statistical analysis of RMSE value by various surrogates in ten runs The experimental results for three outputs from the hot rod roll- ing problem with RSM, KRG, RBF, and an EoS are shown in Surrogate RMS errors produced by the surrogates Figure 7. The result shows that RBF gives the highest RMS error for all three objectives, especially for output Y3 (Pearlite), the F1 F2 error of the RBF is almost five times larger than other three surro- Allotriomorphic Widmanstätten ferrite ferrite F3 Pearlite gates, which means RBF has the lowest accuracy in predicting all three objectives. Also, the difference between the EoS, KRG, and EoS 3.68 × 10−2 4.78 × 10−2 1.62 × 10−2 RSM is minimal, but the EoS shows subtle advantages in accuracy. As for the quantitative validation of the surrogates, a statistical KRG 4.55 × 10−2 6.35 × 10−2 1.95 × 10−2 analysis of RMS errors of surrogate predictions for ten runs in −2 −1 RBF 8.86 × 10 1.55 × 10 2.31 × 10−1 the design space is shown in Table 5. EoS has the best accuracy RSM 5.27 × 10−2 4.47 × 10−2 1.83 × 10−2 except for Y2 (Widmanstätten ferrite), and RBF is the least accurate. For Y2, EoS has the second-highest accuracy, which is The bold numbers are the lowest root mean square errors. second to RSM and the difference between these two is very small (0.0478 − 0.0447)/0.0478 ≈ 6.5%. This indicates that an Results and discussion EoS has the most appropriate performance for the hot rod rolling Comparison between ensembles of surrogates and individual problem because of its accuracy and relatively insensitive predic- surrogates tions to the number of data points for all three objectives and therefore ensembles of surrogates are more robust. Appendix A contains the sample points for the three objectives, allotriomorphic ferrite (Y1), Widmanstätten ferrite (Y2), and pearlite (Y3). In order to generate test data to validate the perfor- Trade-offs among accuracy, size, and time mance of these surrogate models, we repeat a ninefold cross- validation process ten times. In each run, all data sets are ran- To find a relationship among the size, accuracy, and computation domly partitioned into nine subsamples (groups). Of the nine time of surrogates for the hot rod rolling problem, we change the subsamples, one subsample is used as the testing data set and size of the problem in terms of the number of training data which the remaining eight subsamples are used to train the model. are used to train the prediction model. As the sample data has 27 Through nine repetitions, all observations are involved in training points and the K-fold cross-validation is used to generate the and testing. Three outputs are allotriomorphic ferrite (Y1), training data, we increase training data by the way of decreasing Widmanstätten ferrite (Y2), and pearlite (Y3) of steel. the fold value. The relationship between the fold number and In two upper plots of Figure 7, we indicate the RMSE for the the test data can be expressed as a function: prediction of output Y1 (allotriomorphic ferrite) and output Y2 (Widmanstätten ferrite), respectively. In the two lower plots of 27 Ntest = 27 − , (9) Figure 7, we represent the RMSE of output Y3 (pearlite). The dif- Nfold ference is that in the left one, we show the errors of four surro- gates, in the right one, we show the expanded version for the where Ntest is the number of test data, Nfold is the fold number. EoS, KRG, and RSMs. Thus, when the fold value decreases from ninefolds to twofolds, Artificial Intelligence for Engineering Design, Analysis and Manufacturing 493 Fig. 8. RMS errors in the prediction of surrogates with different folds of data from 9 to 2. Table 6. Program run time to compute values for four surrogates from nine- to Table 7. RMS errors generated by four different surrogates from nine- to twofold training data for output Y1 twofold training data for output Y1 No. of folds Program run time for different surrogate models No. RMS errors for different surrogate models of EoS KRG RSM RBF folds EoS KRG RSM RBF 9 115.5 114.2 0.21 0.098 9 3.52 × 10−2 4.65 × 10−2 4.91 × 10−2 8.63 × 10−2 −2 −2 −2 8 102.3 101.3 0.158 0.072 8 3.96 × 10 4.99 × 10 6.16 × 10 9.41 × 10−2 7 86.21 84.12 0.132 0.06 7 4.14 × 10−2 5.01 × 10−2 6.17 × 10−2 9.15 × 10−2 6 67.01 65.76 0.116 0.059 6 4.05 × 10−2 4.88 × 10−2 6.22 × 10−2 9.41 × 10−2 −2 −2 −2 5 46.69 45.29 0.092 0.051 5 4.62 × 10 5.18 × 10 6.66 × 10 9.71 × 10−2 4 44.21 40.31 0.082 0.039 4 4.36 × 10−2 5.12 × 10−2 7.77 × 10−2 9.78 × 10−2 3 36.49 35.21 0.072 0.031 3 5.50 × 10−2 5.46 × 10−2 9.74 × 10−2 1.64 × 10−1 2 23.69 21.56 0.027 0.022 2 6.86 × 10−2 6.70 × 10−2 1.62 × 10−1 1.87 × 10−1 The bold numbers are the lowest root mean square errors. The bold numbers are the lowest values of error for each number of folds. the training data will be reduced from 24 to 14. In order to make a have a smooth change from the ninefold to fourfold training data fair comparison between four different surrogates, the model but abruptly increase at the three- and twofold data, especially the training process is organized with the same training data. The RSM and RBF. That is because, with small numbers of training comparison between these four surrogates in three outputs of data (two- and threefold), surrogate models are more prone to the hot rod rolling problem is shown in Figure 8. have low fidelity compared to the associated physical problem. With the upper two plots in Figure 8, we indicate the RMS As for the comparison between these four surrogates, EoS, errors of predicting Y1 (allotriomorphic ferrite) and output Y2 KRG, and RMS have relatively lower RMS error and higher accu- (Widmanstätten ferrite). With the lower two plots of Figure 8, racy than RBF for all folds. EoS and KRG generally show very we represent the RMS errors of output Y3 (pearlite). The differ- robust behavior, and RSM also shows a robust behavior until ence is that with the left one, we show the errors of four surrogate two- and threefold data are used, where there is a sharp surge and with the right one, we only show three (except RBF). From in error, and the RSM’s accuracy goes down immediately. Figure 8, for each surrogate model, the RMS error gradually The program run time is the computational time required by increases as the amount of training data decreases, therefore accu- Python codes in a Lenovo computer i7-4720HQ 8G 128G SSD racy has a negative correlation with the number of samples used +1T GTX960M. in the training data, and this trend is in-line with our intuition. In As can be seen in Table 6, RBF has the lowest program run addition, it is interesting that the error curves of each surrogate all time among the four surrogates. However, from a practical 494 Reza Alizadeh et al. Table 8. RMS errors generated by four different surrogates from nine- to twofold training data for output Y2 No. of folds RMS errors of different surrogate models EoS KRG RSM RBF 9 4.59 × 10−2 6.38 × 10−2 4.26 × 10−2 1.51 × 10−1 8 5.07 × 10−2 6.23 × 10−2 5.07 × 10−2 1.54 × 10−1 −2 −2 −2 7 5.22 × 10 6.34 × 10 5.13 × 10 1.54 × 10−1 6 5.72 × 10−2 6.90 × 10−2 5.45 × 10−2 1.54 × 10−1 5 5.95 × 10−2 7.03 × 10−2 5.60 × 10−2 1.70 × 10−1 −2 −2 −2 4 5.91 × 10 5.99 × 10 6.66 × 10 1.65 × 10−1 3 6.66 × 10−2 6.67 × 10−2 9.11 × 10−2 1.68 × 10−1 2 7.92 × 10−2 8.42 × 10−2 1.71 × 10−1 2.03 × 10−1 The bold numbers are the lowest values of error for each number of folds. Table 9. RMS errors generated by four different surrogates from nine- to twofold training data for output Y3 No. of folds RMS errors of different surrogate models EoS KRG RSM RBF −2 −2 −2 9 1.75 × 10 2.05 × 10 1.75 × 10 2.19 × 10−1 8 1.66 × 10−2 2.02 × 10−2 2.00 × 10−2 2.37 × 10−1 7 1.74 × 10−2 1.94 × 10−2 1.95 × 10−2 2.31 × 10−1 6 1.81 × 10−2 2.14 × 10−2 1.78 × 10−2 2.33 × 10−1 −2 −2 −2 5 1.95 × 10 2.13 × 10 2.06 × 10 2.38 × 10−1 4 2.04 × 10−2 2.12 × 10−2 2.39 × 10−2 2.35 × 10−1 3 2.57 × 10−2 2.31 × 10−2 3.68 × 10−2 2.73 × 10−1 −2 −2 −2 2 3.52 × 10 2.66 × 10 9.39 × 10 3.18 × 10−1 The bold numbers are the lowest root mean square errors. perspective, the time should include not only the program run Results for output Y1 (allotriomorphic ferrite) time which is consumed to construct the surrogate model but The RMS errors, namely the accuracy, for each surrogate for dif- also the simulation time used for sampling. So, in this work, ferent numbers of design data from nine- to twofolds are shown the time is defined as follows. in Figure 8. Ensembles of surrogates have relatively lower RMS errors and therefore higher accuracy than three individual surro- T = Tpm + Tsm , (10) gates for the nine-, eight-, seven-, six-, five-, and fourfold data. In addition, ensembles of surrogates in two- and threefold data have almost equal accuracy (The difference is about 0.001.) and two- where Tpm is the program run time, and Tsm is the simulation and threefold KRG have much higher accuracy than RSM and time. RBF. Both KRG and an EoS have higher accuracy than RBF in In the hot rod rolling problem, the average sampling time for all K-fold cross-validation methods. Also, they have higher accu- generating each data point in the simulation is about 3.5 h racy even when we use twofold cross-validation for them and (12,600 s), which is at least two orders of magnitude larger than ninefold cross-validation for RBF. So, RBF is not a good choice the program run time (the longest one is about 150 s) for all regarding the accuracy and size measures. four surrogates. So, the program run time can be ignored in Eq. In Table 7, we bold the lowest values of error for each number (10), and the time we mention in the following discussion is of folds in bold. It shows that the RMS error gradually decreases the equivalent of the sampling time, as shown in the following: with decreasing training data for each surrogate, which means that the accuracy is negatively correlated with the amount of training T ≈ Tsample = 3.5 h × Nsample , (11) data, and this trend is in-line with our intuition. In addition, an EoS has much greater accuracy than the other where the Nsample is the number of sample points. Therefore, the three individual surrogates trained for ninefold data even when we time is proportional to the number of sample points. In order to use eight-, seven-, six-, five-, and fourfolds instead of ninefold generate selection rules, a detailed comparison of results among data. So, an EoS gives the most accurate predictions and needs the different surrogate models for each output is shown in the fol- the lowest time to obtain the prediction for allotriomorphic ferrite lowing three sections. when we use nine- to fourfold training data. Also, RBF is not a Artificial Intelligence for Engineering Design, Analysis and Manufacturing 495 good choice because time, accuracy, and size measures for nine- 2.19 × 10−1 2.37 × 10−1 2.31 × 10−1 2.33 × 10−1 2.38 × 10−1 2.35 × 10−1 2.73 × 10−1 3.18 × 10−1 fold cross-validation for RBF is not greater than the twofold cross- RBF validation for the ensembles of surrogates and KRG. As for the comparison between RSM and EoS, EoS has greater accuracy than RSM in four- to ninefold cross-validation. When we use threefold data to train the EoS, we obtain higher accuracy than 1.78 × 10−2 1.75 × 10−2 2.00 × 10−2 −2 2.06 × 10−2 2.39 × 10−2 −2 9.39 × 10−2 when we use ninefold of data to train RSM. Also, we obtain higher 1.95 × 10 3.68 × 10 RSM accuracy by using five- to eightfold of RSM rather than using two- Y3 (Pearlite) fold of data to train the EoS, while RSMs trained with two- to fourfold of data have lower accuracy than EoS trained with two- fold data. −2 2.66 × 10−2 2.05 × 10−2 2.02 × 10−2 −2 2.14 × 10−2 2.13 × 10−2 2.12 × 10−2 2.31 × 10 1.94 × 10 Based on this analysis, some recommendations for time and KRG size of the data set along with the accuracy of output 1 (Y1) can be developed. If we consider the EoS individually, by using twofold instead of ninefold, accuracy will increase about (0.07 − 0.035)/0.07 = 50%. On the other hand, by using twofold 1.75 × 10−2 1.66 × 10−2 −2 1.95 × 10−2 2.04 × 10−2 1.81 × 10−2 −2 3.52 × 10−2 1.74 × 10 2.57 × 10 instead of ninefold, we need (27–27/2 = 13)*3.5 = 45.5 h instead of EoS 27*3.5 = 94.5 h, and so, there are a savings of 35 h. As for the comparison between ensembles of surrogates and individual sur- rogates, ensembles of surrogates are the recommended because of 1.51 × 10−1 1.54 × 10−1 −1 1.54 × 10−1 1.70 × 10−1 1.65 × 10−1 −1 2.03 × 10−1 their higher accuracy and low time-consumption for four- to 1.54 × 10 1.68 × 10 RBF ninefold (19–24) training data, whereas RBF is always the worst choice. Results for output Y2 (Widmanstätten ferrite) 4.26 × 10−2 5.07 × 10−2 −2 5.45 × 10−2 5.60 × 10−2 6.66 × 10−2 −2 1.71 × 10−1 Y2 (Widmanstätten ferrite) For Widmanstätten ferrite, the EoS has relatively lower RMS 5.13 × 10 9.11 × 10 RSM errors, and so, higher accuracy than KRG, and RBF even when we use nine-, eight-, seven-, six-, five-, four-, three-, and twofolds of the data. Although RSM has a relatively higher accuracy for Table 10. RMS errors generated by four surrogates from nine- to twofold training data for output Y1, Y2, and Y3 Bold values mentioned the highlighted values denote to the lowest RMS error obtained from the corresponding Surrogate model. nine-, eight-, seven-, six-, and fivefolds of data, again it has 6.38 × 10−2 6.23 × 10−2 −2 6.90 × 10−2 7.03 × 10−2 5.99 × 10−2 −2 8.42 × 10−2 6.34 × 10 6.67 × 10 lower accuracy in two- and threefolds in comparison to the KRG EoS. RBF also again has relatively low accuracy in comparison to other surrogate modeling methods. KRG also has relatively less accuracy for five- to ninefolds. RSM and has only greater 5.91 × 10−2 −2 7.92 × 10−2 accuracy than RSM in two- to fourfolds. Interestingly, RSM 4.59 × 10−2 5.07 × 10−2 −2 5.72 × 10−2 5.95 × 10−2 6.66 × 10 5.22 × 10 EoS with fewer data points, including five-, six-, seven-, and eightfolds has better accuracy than ninefold KRG. From the analysis in Table 8, some recommendations consid- ering the time and size of the data set along with the accuracy for 8.63 × 10−2 9.41 × 10−2 −2 9.41 × 10−2 9.71 × 10−2 9.78 × 10−2 −1 1.87 × 10−1 Widmanstätten ferrite (Y2) are developed. Entries in bold show 9.15 × 10 1.64 × 10 RBF the lowest values of error for each number of folds. If we consider the EoS by itself, by using twofold instead of ninefold, accuracy will increase about (0.08 − 0.035)/0.08 = 0.43%. On the other hand, by using twofold instead of ninefold data, we need 45.5 h 4.91 × 10−2 6.16 × 10−2 −2 6.22 × 10−2 6.66 × 10−2 7.77 × 10−2 −2 1.62 × 10−1 Y1 (Allotriomorphic ferrite) 6.17 × 10 9.74 × 10 instead of 94.5 h, and so, we can save 35 h. Also, to compare RSM ensembles of surrogates with other single surrogates, by using a fourfold EoS, we need 70.87 h instead of 94.5 h when we choose to use ninefold RSM, and so, we can save 23.63 h (23.63 h faster surrogate) and have (0.055 − 0.04)/0.055 = 27% less accuracy. −2 6.70 × 10−2 4.65 × 10−2 4.99 × 10−2 −2 4.88 × 10−2 5.18 × 10−2 5.12 × 10−2 5.46 × 10 5.01 × 10 RMS errors for different surrogate models KRG Results for output Y3 (pearlite) As shown in Table 9, there is a slight difference between the accu- racy of EoS, KRG, and RSM except for two- and threefold RSM 3.52 × 10−2 3.96 × 10−2 −2 4.05 × 10−2 4.62 × 10−2 4.36 × 10−2 −2 6.86 × 10−2 which has much lower accuracy. On the other hand, there is a 4.14 × 10 5.50 × 10 EoS huge gap among these three surrogates and RBF, which has very low accuracy and can hardly be compared with others. If we consider surrogates individually, there is almost no difference among using four-, five-, six-, eight-, ninefolds of KRG and EoS. No. of folds Therefore, it is possible to save 23.63 h by using fourfold KRG or and an EoS instead of ninefold KRG or and an EoS. Also, for RSM, there is a very slight difference between using fourfold 9 8 7 6 5 4 3 2 instead of ninefold. While using threefold RSM, the accuracy 496 Reza Alizadeh et al. Table 11. Specific guidance for the selection of surrogate models for output Y1 (allotriomorphic ferrite) First selection criterion based on First selection Second selection criterion based on accuracy (rule 1) result time (s) (rule 2) Second selection result RMSE < 0.03 EoS None None 0.03 ≤ RMSE < 0.05 EoS Time < 12 None 12 ≤ Time < 24 None 24 ≤ Time < 36 None 36 ≤ Time < 48 None 48 ≤ Time < 60 None 60 ≤ Time < 72 None 72 ≤ Time < 84 Five- to eightfold EoS 84 ≤ Time Ninefold EoS 0.05 ≤ RMSE < 0.07 EoS, KRG, RSM Time < 12 None 12 ≤ Time < 24 None 24 ≤ Time < 36 None 36 ≤ Time < 48 Twofold EoS and twofold KRG 48 ≤ Time < 60 None 60 ≤ Time < 72 Three- to fourfold EoS and three- to fourfold KRG 72 ≤ Time < 84 Five- to eightfold EoS, five- to eightfold KRG, and five- to ninefold RSM 84 ≤ Time Ninefold EoS, ninefold KRG, and ninefold RSM 0.07 ≤ RMSE < 0.09 RSM, RBF Time < 12 None 12 ≤ Time < 24 None 24 ≤ Time < 36 None 36 ≤ Time < 48 None 48 ≤ Time < 60 None 60 ≤ Time < 72 Fourfold RSM 72 ≤ Time < 84 None 84 ≤ Time Ninefold RBF 0.09 ≤ RMSE < 0.11 RSM, RBF Time < 12 None 12 ≤ Time < 24 None 24 ≤ Time < 36 None 36 ≤ Time < 48 None 48 ≤ Time < 60 None 60 ≤ Time < 72 Threefold RSM and fourfold RBF 72 ≤ Time < 84 Five- to eightfold RBF 84 ≤ Time None 0.11 ≤ RMSE RSM, RBF Time < 12 None 12 ≤ Time < 24 None 24 ≤ Time < 36 None 36 ≤ Time < 48 Twofold RBF and twofold RSM 48 ≤ Time < 60 None 60 ≤ Time < 72 Threefold RBF 72 ≤ Time < 84 None 84 ≤ Time None Artificial Intelligence for Engineering Design, Analysis and Manufacturing 497 will reduce (0.035 − 0.03)/0.035 = 14%, but the process will be to find an EoS which is created by the least possible number of much faster. In other words, by using threefold instead of nine- data points. The resulting ensemble surrogate has higher accuracy fold, 63 h instead of 94.5 h are needed, which can save 31.5 h. than each individual surrogate and is less computationally inten- It also means that we can use only 9 data points instead of 27 sive. To achieve this ensemble surrogate, we compare it with indi- data points which saves 67% of the simulation time which we vidual surrogate models based on computation time, size, and need to spend to generate the simulation data. desired accuracy. For this purpose, we use RMSE as the accuracy The summary of the results is given in Table 10. The high- measure, time of simulation as the computation performance lighted values denote to the lowest RMS error obtained from measure, and the number of data points as the dimension mea- the corresponding surrogate model. As illustrated in Table 10, sure. In summary, we find that (1) it is effective to use cross- for the first response variable, Y1 (allotriomorphic ferrite), in validation to study the impact of the size of the sample data set; four- to ninefolds of data, an EoS has lower RMSE than individual (2) the highest accuracy with least required data and less compu- surrogates and only for two- and threefolds of data, KRG has tation time is achievable using the right number of samples; and lower RMSE than others. Also, for the second response variable, (3) an example of surrogates is relatively insensitive to the size of Y2 (Widmanstätten ferrite), in two- to fourfolds of data, EoS the sample data or number of data points. We create rule-based has lower RMSE than other surrogates. Finally, for the third guidance for selection of surrogate models for one of the response response variable, Y3 (pearlite), in four- to ninefolds of data, variables in Table 11. Creating rule-based templates for each EoS has lower RMSE than other surrogates response variable and putting them together into a knowledge- In comparison to some other studies, such as Viana et al. based platform and ontology is possible future research. As (2013), Chaudhuri and Haftka (2014), Badhurshah and Samad another possible research direction, these knowledge-based plat- (2015), and Bhattacharjee et al. (2018) in ensemble surrogates forms and ontology can be used in automating the surrogate related issues, there is a big difference in the method. Authors modeling process. in these papers applied multiple surrogate methods for multiob- This paper is based on several assumptions: (1) hot rod rolling jective optimization, while in this paper, we used an EoS for pre- process is a single independent engineering process and (2) we are diction. Also, Goel et al. (2007) and Bishop (1995) create a WAS able to create a surrogate model for hot rod rolling considering it by estimating the covariance between surrogates from residuals at as an independent engineering process. Removing these assump- test or training data sets and using the PWS, while we create the tions and considering the casting and reheating as preprocesses weights based on cross-validation errors. Basudhar (2012), Viana along with cooling and forging as post processes is another pos- et al. (2013), Bhattacharjee et al. (2016), and Ezhilsabareesh et al. sible future research direction that we can take. Furthermore, we (2018) used an EoS in a relatively high number of data, while we create rule-based guidance for selection of surrogate models for use an EoS in small data size. one of the response variables in Table 11. Creating rule-based In addition, the ensembles created by Chaudhuri and Haftka templates for each and every response variable and putting (2014), Wang et al. (2016), Bhattacharjee et al. (2018), Lv et al. them together into a knowledge-based platform and ontology is (2018), Song et al. (2018), and Yin et al. (2018) are more time possible future research. As another possible research direction, consuming than each individual surrogate they used to create these knowledge-based platforms and ontology can be used in the ensemble. Whereas, we create ensembles which are less time automating the surrogate modeling process. consuming than individual surrogates and have less inconsistency. Acknowledgments. Reza Alizadeh, Anand Balu Nellippallil, Janet K. Allen, In order to give specific guidance for surrogate selection, a and Farrokh Mistree acknowledge financial support from the John and Mary rule-based template is given in Table 11. For the hot rod rolling Moore Chair and the L.A. Comp Chair at the University of Oklahoma. Guoxin problem, three tables as decision trees corresponding to the Wang, Liangyue Jia, and Jia Hao also acknowledge financial support from the three outputs are developed to manage the selection among the National Ministries of China (JCKY2016602B007). This paper is an outcome four types of surrogate. As we have three characteristics for the of the International Systems Realization Partnership between the Institute for problem, accuracy, size, and time, and time is directly propor- Industrial Engineering at The Beijing Institute of Technology, The Systems tional to size, the selection of appropriate surrogates can be Realization Laboratory at The University of Oklahoma and the Design based on accuracy and size. In this work, accuracy is used as Engineering Laboratory at Purdue. the criterion for the first selection, and size is the criterion for the second selection. Specific guidance for the selection of surro- gate models for output Y1 is shown in Table 11 as an example. References There is no surrogate model which gives us RMSE < 0.03, while Acar E (2015) Effect of error metrics on optimum weight factor selection for EoS can be used if accuracy 0.03 ≤ RMSE < 0.05 is needed, and ensemble of metamodels. Expert Systems with Applications 42, 2703–2709. here, EoS is the choice based on the first selection criteria doi:10.1016/j.eswa.2014.11.020 (RMSE). Also, for the second selection criteria (time), five- to Acar E and Rais-Rohani M (2009) Ensemble of metamodels with optimized eightfold can be chosen when we need to spend no more than weight factors. Structural and Multidisciplinary Optimization 37, 279–294. 84 s to run the code for the surrogate model. Adhav R, Samad A and Kenyery F (2015) Design optimization of electric cen- trifugal pump by multiple surrogate models. Paper Presented at the SPE Middle East Oil & Gas Show and Conference, Manama, Bahrain. Closing remarks Alizadeh R, Lund PD, Beynaghi A, Abolghasemi M and Maknoon R (2016) An integrated scenario-based robust planning approach for foresight and Based on the published literature, creating an EoS using cross- strategic management with application to energy industry. Technological validation errors results is higher accurate, but it is more compu- Forecasting and Social Change 104, 162–171. doi:10.1016/ tationally intensive than using individual surrogates. Our contri- j.techfore.2015.11.030 bution in this paper is to propose a method to build an EoS Alizadeh R, Allen JK and Mistree F (2019) Managing computational com- that is both accurate and less computationally expensive. The nov- plexity using surrogate models: a critical review. Research in Engineering elty in this paper is to propose a method based on cross-validation Design. Under Review. 498 Reza Alizadeh et al. Arias-Montano A, Coello CAC and Mezura-Montes E (2012) Multi-objective Jägle E (2007) Modelling of Microstructural Banding During Transformations airfoil shape optimization using a multiple-surrogate approach. Paper in Steel (Master of Philosophy Dissertation). University of Cambridge, Presented at the IEEE Congress on Evolutionary Computation (CEC), Cambridge, UK. Brisbane, Australia. Jones S and Bhadeshia H (1997) Kinetics of the simultaneous decomposition Babaei M and Pan I (2016) Performance comparison of several response sur- of austenite into several transformation products. Acta Materialia 45, 2911– face surrogate models and ensemble methods for water injection optimiza- 2920. tion under uncertainty. Computers & Geosciences 91, 19–32. doi:10.1016/ Jones SJ and Bhadeshia HKDH (2017) Program STRUCTURE on the j.cageo.2016.02.022 Materials Algorithm Project Web Site. Available at http://www.msm.cam. Badhurshah R and Samad A (2015) Multiple surrogate based optimization of ac.uk/map/steel/programs/structure.html. a bidirectional impulse turbine for wave energy conversion. Renewable Kaleibari SS, Beiragh RG, Alizadeh R and Solimanpur M (2016) A frame- Energy 74, 749–760. doi:10.1016/j.renene.2014.09.001 work for performance evaluation of energy supply chain by a compatible Basudhar A (2012) Selection of anisotropic kernel parameters using multiple network data envelopment analysis model. Scientia Iranica. Transaction surrogate information. Paper Presented at the 12th AIAA Aviation E, Industrial Engineering 23, 1904–1917. Technology, Integration, and Operations (ATIO) Conference and 14th Khademi A, Ghorbani Renani N, Mofarrahi M, Rangraz Jeddi A and Mohd AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference, Yusof N (2013) The best location for speed bump installation using exper- Indianapolis, Indiana. imental design methodology. Promet - Traffic and Transportation 25, 565– Bellary SAI and Samad A (2017) An alternative approach to surrogate averaging 574. doi:10.7307/ptt.v25i6.1188 for a centrifugal impeller shape optimisation. International Journal of Computer Kleijnen JP (2017) Regression and Kriging metamodels with their experi- Aided Engineering and Technology 9, 62–83. doi:10.1504/ijcaet.2017.080769 mental designs in simulation: a review. European Journal of Operational Bellary SAI, Adhav R, Siddique MH, Chon B-H, Kenyery F and Samad A Research 256, 1–16. (2016) Application of computational fluid dynamics and surrogate-coupled Korda AA, Mutoh Y, Miyashita Y, Sadasue T and Mannan S (2006) In situ evolutionary computing to enhance centrifugal-pump performance. observation of fatigue crack retardation in banded ferrite–pearlite micro- Engineering Applications of Computational Fluid Mechanics 10, 171–181. structure due to crack branching. Scripta Materialia 54, 1835–1840. doi:10.1080/19942060.2015.1128359 Krauss GB (2003) Solidification, segregation, and banding in carbon and alloy Bellucci JP and Bauer Jr KW (2017) A Taylor series approach to the robust steels. Metallurgical and Materials Transactions A 34, 781–792. parameter design of computer simulations using kriging and radial basis Lim D, Ong Y-S, Jin Y and Sendhoff B (2007) A study on metamodeling function neural networks. International Journal of Quality Engineering techniques, ensembles, and multi-surrogates in evolutionary computation. and Technology 6, 137–160. Paper Presented at the Proceedings of the 9th Annual Conference on Beynaghi A, Moztarzadeh F, Shahmardan A, Alizadeh R, Salimi J and Genetic and Evolutionary Computation, London, England. Mozafari M (2016) Makespan minimization for batching work and rework Liu K, Tovar A, Nutwell E and Detwiler D (2015) Thin-walled compliant process on a single facility with an aging effect: a hybrid meta-heuristic mechanism component design assisted by machine learning and multiple algorithm for sustainable production management. Journal of Intelligent surrogates. Paper Presented at the SAE 2015 World Congress & Manufacturing 30, 33–45. doi:10.1007/s10845-016-1223-0 Exhibition, Detroit, Michigan. Bhat S, Viana FAC, Lind R and Haftka R (2010) Control-oriented design Lv Z, Zhao J, Wang W and Liu Q (2018) A multiple surrogates based PSO using H-infinity synthesis and multiple surrogates. Paper Presented at the algorithm. Artificial Intelligence Review, 1–22. doi:10.1007/ 51st AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and s10462-017-9601-3 Materials Conference, Orlando, Florida. Mack Y, Goel T, Shyy W, Haftka R and Queipo N (2005) Multiple surrogates Bhattacharjee KS, Singh HK, Ray T and Branke J (2016) Multiple surrogate for the shape optimization of bluff body-facilitated mixing. Paper Presented assisted multiobjective optimization using improved pre-selection. Paper at the 43rd AIAA Aerospace Sciences Meeting and Exhibit, Reno, Nevada. Presented at the 2016 IEEE Congress on Evolutionary Computation (CEC). doi:10.2514/6.2005-333 Bhattacharjee KS, Singh HK and Tapabrata R (2018) Multiple surrogate Mirjalili S (2019) Evolutionary Radial Basis Function Networks. In Mirjalili S assisted many-objective optimization for computationally expensive engi- (ed.), Evolutionary Algorithms and Neural Networks: Theory and neering design. ASME Journal of Mechanical Design 140, 051403. Applications, Vol. 2. Cham: Springer International Publishing, pp. 105–139. Bishop CM (1995) Neural Networks for Pattern Recognition. Birmingham, Montgomery DC (2017) Design and Analysis of Experiments. Hoboken, NJ: UK: Department of Computer Science and Applied Mathematics Aston John Wiley & Sons. University, Oxford University Press. Nellippallil AB, Rangaraj V, Gautham BP, Singh AK, Allen JK and Bodnar R and Hansen S (1994) Effects of austenite grain size and cooling rate Mistree F (2017) A goal-oriented, inverse decision-based design method on Widmanstätten ferrite formation in low-alloy steels. Metallurgical and to achieve the vertical and horizontal integration of models in a hot rod roll- Materials Transactions A 25, 665–675. ing process chain. Paper Presented at the ASME 2017 International Design Chaudhuri A and Haftka RT (2014) Efficient global optimization with adap- Engineering Technical Conferences and Computers and Information in tive target setting. AIAA Journal 52, 1573–1578. doi:10.2514/1.J052930 Engineering Conference, Cleveland, Ohio, USA. Chaudhuri A, Haftka RT, Ifju P, Chang K, Tyler C and Schmitz T (2015) Nellippallil AB, Rangaraj V, Gautham BP, Singh AK, Allen JK and Experimental flapping wing optimization and uncertainty quantification Mistree F (2018) An inverse, decision‐based design method for integrated using limited samples. Structural and Multidisciplinary Optimization 51, design exploration of materials, products and manufacturing processes. 957–970. doi:10.1007/s00158-014-1184-x ASME Journal of Mechanical Design 140, 111403–111417. Ezhilsabareesh K, Rhee SH and Samad A (2018) Shape optimization of a Qiu Q, Li B and Feng P (2016) Optimal design of hydraulic excavator work- bidirectional impulse turbine via surrogate models. Engineering ing device based on multiple surrogate models. Advances in Mechanical Applications of Computational Fluid Mechanics 12, 1–12. doi:10.1080/ Engineering 8. doi:10.1177/1687814016647947 19942060.2017.1330709 Razavi S, Tolson BA and Burn DH (2012) Review of surrogate modeling Goel T, Haftka RT, Shyy W and Queipo NV (2006) Ensemble of surrogates. in water resources. Water Resources Research 48, 1–32. doi:10.1029/ Structural and Multidisciplinary Optimization 33, 199–216. doi:10.1007/ 2011WR011527 s00158-006-0051-9 Samad A, Kim K-Y, Goel T, Haftka RT and Shyy W (2006) Shape optimiza- Goel T, Haftka RT, Shyy W and Queipo NV (2007) Ensemble of surrogates. tion of turbomachinery blade using multiple surrogate models. Paper Structural and Multidisciplinary Optimization 33, 199–216. doi:10.1007/ Presented at the ASME 2006 2nd Joint U.S.-European Fluids Engineering s00158-006-0051-9 Summer Meeting Collocated With the 14th International Conference on Habib A, Kumar Singh H and Ray T (2017) A multiple surrogate assisted Nuclear Engineering, Miami, Florida, USA. evolutionary algorithm for optimization involving iterative solvers. Samad A, Lee K-D, Kim K-Y and Haftka R (2007) Application of multiple- Engineering Optimization 49, 1–20. doi:10.1080/0305215X.2017.1401068 surrogate model to optimization of a dimpled channel. Paper Presented at Artificial Intelligence for Engineering Design, Analysis and Manufacturing 499 the 7th World Congresses of Structural and Multidisciplinary Optimization, Viana F, Haftka R and Watson L (2010) Why not run the efficient global Seoul, Korea. optimization algorithm with multiple surrogates? Paper Presented at the Shankar Bhattacharjee K, Kumar Singh H and Ray T (2016) Multi-objective 51st AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and optimization with multiple spatially distributed surrogates. ASME Journal of Materials Conference, Orlando, Florida. Mechanical Design 138, 091401–091410. doi:10.1115/1.4034035 Viana FA, Haftka RT and Watson LT (2013) Efficient global optimization Shi R, Liu L, Long T and Liu J (2016) An efficient ensemble of radial basis algorithm assisted by multiple surrogate techniques. Journal of Global functions method based on quadratic programming. Engineering Optimization 56, 669–689. Optimization 48, 1202–1225. doi:10.1080/0305215X.2015.1100470 Villanueva D, Haftka RT, Le Riche R and Picard G (2013) Locating multiple Shyy W, Tucker PK and Vaidyanathan R (2001) Response surface and neural candidate designs with surrogate-based optimization. Paper Presented at the network techniques for rocket engine injector optimization. Journal of 10th World Congress on structural and multidisciplinary optimization, Propulsion and Power 17, 391–401. Orlando, Florida, USA. Song X, Lv L, Li J, Sun W and Zhang J (2018) An advanced and robust Wang H, Ye F, Li E and Li G (2016) A comparative study of expected ensemble surrogate model: extended adaptive hybrid functions. Journal of improvement-assisted global optimization with different surrogates. Mechanical Design 140, 041402–041409. doi:10.1115/1.4039128 Engineering Optimization 48, 1432–1458. doi:10.1080/0305215X.2015. Spitzig W (1983) Effect of sulfide inclusion morphology and pearlite banding 1115645 on anisotropy of mechanical properties in normalized C-Mn steels. Xu J and Zeger SL (2001) The evaluation of multiple surrogate endpoints. Metallurgical Transactions A 14, 271–283. Biometrics 57, 81–87. Tomita Y (1995) Effect of modified austemper on tensile properties of 0·52%C Yin H, Fang H, Wen G, Gutowski M and Xiao Y (2018) On the ensemble of steel. Materials Science and Technology 11, 994–997. metamodels with multiple regional optimized weight factors. Structural and Viana FAC and Haftka RT (2008) Using multiple surrogates for minimization Multidisciplinary Optimization 58, 245–263. doi:10.1007/s00158-017-1891-1 of the RMS Error in meta-modeling. Paper Presented at the ASME 2008 Zamani Sabzi H, Abudu S, Alizadeh R, Soltanisehat L, Dilekli N and International Design Engineering Technical Conferences and Computers King JP (2018) Integration of time series forecasting in a dynamic decision and Information in Engineering Conference, Brooklyn, New York, USA. support system for multiple reservoir management to conserve water Viana F, Haftka R, Steffen JV, Butkewitsch S and Leal MF (2008) Optimal sources. Energy Sources, Part A: Recovery, Utilization, and Environmental use of multiple surrogate for reduced RMS error in meta-model. Paper Effects 40, 1398–1416. doi:10.1080/15567036.2018.1476934. Presented at the NSF Engineering Research and Innovation Conference, Zerpa LE, Queipo NV, Pintos S and Salager J-L (2005) An optimization Knoxville, Tennessee. methodology of alkaline–surfactant–polymer flooding processes using Viana FA, Picheny V and Haftka RT (2009) Conservative prediction via field scale numerical simulation and multiple surrogates. Journal of safety margin: design through cross-validation and benefits of multiple sur- Petroleum Science and Engineering 47, 197–208. rogates. Paper Presented at the ASME 2009 International Design Engineering Zhou Q, Wang Y, Choi S, Jiang P, Shao X, Hu J and Shu L (2018) A robust Technical Conferences and Computers and Information in Engineering optimization approach based on multi-fidelity metamodel. Structural and Conference, San Diego, California, USA. Multidisciplinary Optimization 57, 775–797. 500 Reza Alizadeh et al. Appendix A Sample data for the hot rod rolling problem The sample points for the three objectives, allotriomorphic ferrite (Y1), Widmanstätten ferrite (Y2), and pearlite (Y3), are summarized in this section. To generate test data to validate the performance of these surrogate models, we repeat a ninefold cross-validation process ten times. In each run, all data sets are randomly partitioned into nine subsamples (groups). Of the nine subsam- ples, one subsample is used as the testing data set and the remaining eight sub- samples are used to train the model. Through nine repetitions, all observations are involved in training and testing. Three output variables, namely allotrio- morphic ferrite (Y1), Widmanstätten ferrite (Y2), and pearlite (Y3) of steel are predicted based on values of five input variables, namely carbon concen- tration rate, manganese concentration rate, grain size, cooling rate, and final temperature. The prediction results are shown in Table 12. Table 12. Values for input and output variables Input Output Concentrations (%) Grain size (μm) Cooling rate (K/min) Final temperature (K) Volume fractions (%) Carbon Manganese Allotriomorphic Widmanstatten Pearlite 0.18 0.7 30 11 639.53 0.6921 0.0195 0.2883 0.24 0.7 55 11 643.19 0.4612 0.1621 0.3761 0.3 0.7 100 11 643.67 0.2429 0.2705 0.4866 0.24 0.7 30 55 620.9 0.3453 0.2352 0.4195 0.3 0.7 55 55 612.31 0.2439 0.2442 0.5113 0.18 0.7 100 55 624.77 0.2094 0.4899 0.2991 0.3 0.7 30 100 601.36 0.3163 0.1155 0.5681 0.18 0.7 55 100 612.42 0.2955 0.3944 0.3101 0.24 0.7 100 100 598.16 0.1458 0.4292 0.4243 0.24 1.1 30 11 625.09 0.5684 0.019 0.4126 0.3 1.1 55 11 626.71 0.3384 0.1439 0.5176 0.18 1.1 100 11 633.67 0.3132 0.3937 0.2915 0.3 1.1 30 55 599.3 0.3316 0.0888 0.5976 0.18 1.1 55 55 608.67 0.3179 0.3714 0.3107 0.24 1.1 100 55 600.9 0.1016 0.4618 0.4366 0.18 1.1 30 100 596.93 0.4185 0.2573 0.3242 0.24 1.1 55 100 587.65 0.2197 0.3122 0.4667 0.3 1.1 100 100 572.42 0.1042 0.2828 0.613 0.3 1.5 30 11 615.95 0.4575 0 0.5425 0.3 1.5 55 11 611.88 0.3085 0.14 0.5515 0.24 1.5 100 11 614.11 0.2367 0.3321 0.4312 0.18 1.5 30 55 592.67 0.3148 0.3523 0.3329 0.24 1.5 55 55 584.69 0.2129 0.3236 0.4618 0.3 1.5 100 55 572.19 0.1106 0.2527 0.6367 0.24 1.5 30 100 567.66 0.3197 0.185 0.4953 0.3 1.5 55 100 556.64 0.1586 0.162 0.6794 0.18 1.5 100 100 568.25 0.1434 0.4933 0.3102 The bold numbers are the lowest root mean square errors. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 501 Reza Alizadeh received his BSc in Industrial Engineering from the Urmia 2015. He is currently an Associate Professor with the Beijing Institute of University of Technology in 2011. He also received his MSc in Technology Technology. His current research interests include reconfigurable manufac- Foresight from the Amirkabir University of Technology in 2013. He is turing systems, intelligent design, and knowledge engineering. currently a PhD candidate in Industrial and Systems Engineering at the University of Oklahoma, USA. His current research interests include simulation-based design, surrogate modeling, network analysis, data analy- Jia Hao received the MS and PhD degrees in Mechanical Engineering from the tics, decision-making under uncertainty, and cloud-based design. Beijing Institute of Technology, China, in 2010 and 2014, respectively. He is currently an Assistant Professor with the Beijing Institute of Technology. His current research interests include intelligent design, computational creativity, Liang Jia received the MS degree in Mechanical Engineering from the Beijing and related area. Institute of Technology, China, in 2018, where he is currently pursuing the PhD degree. His research interests include intelligent manufacturing systems and intelligent design. Janet K. Allen received the SB degree from MIT in 1967, and the PhD degree from the University of California, Berkeley, in 1973. She was a Senior Research Scientist at the George W. Woodruff School of Mechanical Anand Balu Nellippallil received his BTech, in Production Engineering from Engineering between 1992 and 2009. She is currently John and Mary the University of Calicut in 2012. He also received his MTech, in Materials Moore Chair and professor at the University of Oklahoma, USA. Her Science and Engineering from the Indian Institute of Technology current research interests include simulation-based design of engineering Bhubaneswar in 2014. He also received his PhD degree in Aerospace and systems, managing uncertainty in design, sustainability, and design Mechanical Engineering from the University of Oklahoma, USA, in 2018. pedagogy. He is currently a Research Engineer II at the Center for Advanced Vehicular Systems, Mississippi State University. His current research inter- ests include “integrated realization of engineered materials, products and Farrokh Mistree received the BS degree from IIT Kharagpur in 1967, the MS associated manufacturing processes” with emphasis on the following degree and the PhD degree from the University of California, Berkeley, in research thrusts: simulation-based design, robust design, design decision- 1970 and 1974, respectively. He was the Director of the School of making under uncertainty, and cloud-based design decision support. Aerospace and Mechanical Engineering between 2009 and 2013. He is cur- rently the L.A. Comp Chair and Professor at the University of Oklahoma, USA, from 2014 to 2015. Before joining the University of Oklahoma, he Guoxin Wang received the BS degree from Lanzhou Jiaotong University in was a professor at Georgia Institute of Technology and the Associate 2001, the MS degree from Lanzhou Jiaotong University in 2004, and the Chair of the Woodruff School of Mechanical Engineering. His current PhD degree from the Beijing Institute of Technology, China, in 2007. He research interests include knowledge-based realization and management of was a Visiting Scholar at the University of Oklahoma, USA, from 2014 to complex engineering systems.