Log in

Bayesian multi-objective optimization of process design parameters in constrained settings with noise: an engineering design application

  • Original Article
  • Published:
Engineering with Computers Aims and scope Submit manuscript

Abstract

The use of adhesive joints in various industrial applications has become increasingly popular due to their beneficial characteristics, including their high strength-to-weight ratio, design flexibility, limited stress concentrations, planar force transfer, good damage tolerance, and fatigue resistance. However, finding the best process parameters for adhesive bonding can be challenging. This optimization problem is inherently multi-objective, aiming to maximize break strength while minimizing cost and constrained to avoid any visual damage to the materials and ensure that stress tests do not result in adhesion-related failures. Additionally, testing the same process parameters several times may yield different break strengths, making the optimization process uncertain. Conducting physical experiments in a laboratory setting is costly, and traditional evolutionary approaches like genetic algorithms are not suitable due to the large number of experiments required for evaluation. Bayesian optimization is suitable in this context, but few methods simultaneously consider the optimization of multiple noisy objectives and constraints. This study successfully applies advanced learning techniques to emulate the objective and constraint functions based on limited experimental data. These are incorporated into a Bayesian optimization framework, which efficiently detects Pareto-optimal process configurations under strict budget constraints.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Germany)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

Data availability

The data that support the findings of this study are available from the authors but restrictions apply to the availability of these data, which were used under license from Flanders Make vzw for the current study, and so are not publicly available. Data are, however, available from the authors upon reasonable request and with permission from the aforementioned institution.

Notes

  1. https://www.flandersmake.be.

  2. Other noise factors could not be controlled in the simulator, so they are not further discussed.

References

  1. Brockmann W, Geiß PL, Klingen J, Schröder KB (2008) Adhesive bonding: materials, applications and technology. Wiley, New York

    Book  Google Scholar 

  2. Cavezza F, Boehm M, Terryn H, Hauffman T (2020) A review on adhesively bonded aluminium joints in the automotive industry. Metals. https://doi.org/10.3390/met10060730

    Article  Google Scholar 

  3. Licari JJ, Swanson DW (2011) Adhesives technology for electronic applications: materials, processing, reliability. William Andrew, Norwich

    Google Scholar 

  4. Correia S, Anes V, Reis L (2018) Effect of surface treatment on adhesively bonded aluminium–aluminium joints regarding aeronautical structures. Eng Fail Anal 84:34–45. https://doi.org/10.1016/j.engfailanal.2017.10.010

    Article  Google Scholar 

  5. da Silva LFM, Ochsner A, Adams RD, Spelt JK (2011) Handbook of adhesion technology. Springer, Berlin

    Book  Google Scholar 

  6. Pocius AV (2021) Adhesion and adhesives technology: an introduction. Carl Hanser Verlag GmbH Co KG, Munich

    Book  Google Scholar 

  7. Budhe S, Banea MD, De Barros S, Da Silva LFM (2017) An updated review of adhesively bonded joints in composite materials. Int J Adhes Adhes 72:30–42

    Article  Google Scholar 

  8. Habenicht G (2008) Applied adhesive bonding: a practical guide for flawless results. Wiley, West Sussex

    Book  Google Scholar 

  9. Chiarello F, Belingheri P, Fantoni G (2021) Data science for engineering design: state of the art and future directions. Comput Ind 129:103447. https://doi.org/10.1016/j.compind.2021.103447

    Article  Google Scholar 

  10. Zhou A, Qu B-Y, Li H, Zhao S-Z, Suganthan PN, Zhang Q (2011) Multiobjective evolutionary algorithms: a survey of the state of the art. Swarm Evolut Comput 1(1):32–49. https://doi.org/10.1016/j.swevo.2011.03.001

    Article  Google Scholar 

  11. Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic algorithm: Nsga-ii. IEEE Trans Evol Comput 6(2):182–197. https://doi.org/10.1109/4235.996017

    Article  Google Scholar 

  12. Corbett M, Sharos PA, Hardiman M, McCarthy CT (2017) Numerical design and multi-objective optimisation of novel adhesively bonded joints employing interlocking surface morphology. Int J Adhes Adhes 78:111–120. https://doi.org/10.1016/j.ijadhadh.2017.06.002

    Article  Google Scholar 

  13. Labbé S, Drouet J-M (2012) A multi-objective optimization procedure for bonded tubular-lap joints subjected to axial loading. Int J Adhes Adhes 33:26–35. https://doi.org/10.1016/j.ijadhadh.2011.09.005

    Article  Google Scholar 

  14. Brownlee AE, Wright JA (2015) Constrained, mixed-integer and multi-objective optimisation of building designs by NSGA-II with fitness approximation. Appl Soft Comput 33:114–126. https://doi.org/10.1016/j.asoc.2015.04.010

    Article  Google Scholar 

  15. Jain H, Deb K (2013) An evolutionary many-objective optimization algorithm using reference-point based nondominated sorting approach, part ii: Handling constraints and extending to an adaptive approach. IEEE Trans Evol Comput 18(4):602–622

    Article  Google Scholar 

  16. Li K, Chen R, Fu G, Yao X (2018) Two-archive evolutionary algorithm for constrained multiobjective optimization. IEEE Trans Evol Comput 23(2):303–315

    Article  Google Scholar 

  17. Gao Y-L, Qu M (2012) Constrained multi-objective particle swarm optimization algorithm. In: International Conference on Intelligent Computing. Springer, Berlin, pp 47–55

  18. Chugh T, Sindhya K, Hakanen J, Miettinen K (2017) A survey on handling computationally expensive multiobjective optimization problems with evolutionary algorithms. Soft Comput. https://doi.org/10.1007/s00500-017-2965-0

    Article  Google Scholar 

  19. Frazier PI (2018) Bayesian optimization. In: Recent advances in optimization and modeling of contemporary problems. Informs, pp 255–278

  20. Rojas Gonzalez S, van Nieuwenhuyse I (2020) A survey on kriging-based infill algorithms for multiobjective simulation optimization. Comput Oper Res 116:104869. https://doi.org/10.1016/j.cor.2019.104869

    Article  MathSciNet  Google Scholar 

  21. Hernández-Lobato JM, Gelbart M, Hoffman M, Adams R, Ghahramani Z (2015) Predictive entropy search for bayesian optimization with unknown constraints. In: International Conference on Machine Learning. PMLR, pp 1699–1707

  22. Gelbart MA, Adams RP, Hoffman MW, Ghahramani Z et al (2016) A general framework for constrained Bayesian optimization using information-based search. J Mach Learn Res 17(160):1–53

    MathSciNet  Google Scholar 

  23. Garrido-Merchán EC, Hernández-Lobato D (2019) Predictive entropy search for multi-objective Bayesian optimization with constraints. Neurocomputing 361:50–68

    Article  Google Scholar 

  24. Feliot P, Bect J, Vazquez E (2017) A Bayesian approach to constrained single-and multi-objective optimization. J Global Optim 67(1):97–133

    Article  MathSciNet  Google Scholar 

  25. Rojas Gonzalez S, Jalali H, Nieuwenhuyse IV (2020) A multiobjective stochastic simulation optimization algorithm. Eur J Oper Res 284(1):212–226. https://doi.org/10.1016/j.ejor.2019.12.014

    Article  MathSciNet  Google Scholar 

  26. Miettinen K (1999) Nonlinear multiobjective optimization, vol 12. Springer, Berlin

    Google Scholar 

  27. Forrester A, Sobester A, Keane A (2008) Engineering design via surrogate modelling (a practical guide), 1st edn. Wiley, West Sussex

    Book  Google Scholar 

  28. Horn D, Dagge M, Sun X, Bischl B (2017) First investigations on noisy model-based multi-objective optimization. In: International Conference on Evolutionary Multi-Criterion Optimization. Springer, Berlin, pp. 298–313. https://doi.org/10.1007/978-3-319-54157-0_21

  29. Loeppky JL, Sacks J, Welch W (2009) Choosing the sample size of a computer experiment: a practical guide. Technometrics 51(4):366–376

    Article  MathSciNet  Google Scholar 

  30. Zeng Y, Cheng Y, Liu J (2022) An efficient global optimization algorithm for expensive constrained black-box problems by reducing candidate infilling region. Inf Sci 609:1641–1669

    Article  Google Scholar 

  31. Jones DR, Schonlau M, Welch WJ (1998) Efficient global optimization of expensive black-box functions. J Global Optim 13(4):455–492. https://doi.org/10.1023/A:1008306431147

    Article  MathSciNet  Google Scholar 

  32. Williams CK, Rasmussen CE (2006) Gaussian processes for machine learning, vol 2. MIT press, Cambridge

    Google Scholar 

  33. Knowles J (2006) Parego: a hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems. IEEE Trans Evol Comput 10(1):50–66. https://doi.org/10.1109/TEVC.2005.851274

    Article  Google Scholar 

  34. Emmerich MT, Deutz AH, Klinkenberg JW (2011) Hypervolume-based expected improvement: monotonicity properties and exact computation. In: 2011 IEEE Congress of Evolutionary Computation (CEC). IEEE, pp 2147–2154

  35. Couckuyt I, Deschrijver D, Dhaene T (2014) Fast calculation of multiobjective probability of improvement and expected improvement criteria for pareto optimization. J Global Optim 60(3):575–594

    Article  MathSciNet  Google Scholar 

  36. Zhan D, **ng H (2020) Expected improvement for expensive optimization: a review. J Global Optim 78(3):507–544. https://doi.org/10.1007/s10898-020-00923-x

    Article  MathSciNet  Google Scholar 

  37. Daulton S, Balandat M, Bakshy E (2021) Parallel Bayesian optimization of multiple noisy objectives with expected hypervolume improvement. Adv Neural Inform Process Syst 34:2187–2200

    Google Scholar 

  38. Kleijnen JP (2018) Design and analysis of simulation experiments. Springer, Berlin

    Google Scholar 

  39. Ankenman B, Nelson BL, Staum J (2010) Stochastic kriging for simulation metamodeling. Oper Res 58(2):371–382. https://doi.org/10.1109/WSC.2008.4736089

    Article  MathSciNet  Google Scholar 

  40. Quan N, Yin J, Ng SH, Lee LH (2013) Simulation optimization via kriging: a sequential search using expected improvement with computing budget constraints. IIE Trans 45(7):763–780. https://doi.org/10.1080/0740817X.2012.706377

    Article  Google Scholar 

  41. Loka N, Couckuyt I, Garbuglia F, Spina D, Van Nieuwenhuyse I, Dhaene T (2022) Bi-objective Bayesian optimization of engineering problems with cheap and expensive cost functions. Eng Comput 39:1923–1933

    Article  Google Scholar 

  42. Zitzler E, Thiele L (1998) Multiobjective optimization using evolutionary algorithms—a comparative case study. In: Parallel Problem Solving from Nature—PPSN V: 5th International Conference Amsterdam, The Netherlands September 27–30, 1998 Proceedings 5. Springer, Berlin, pp 292–301

  43. Emmerich MT, Giannakoglou KC, Naujoks B (2006) Single-and multiobjective evolutionary optimization assisted by gaussian random field metamodels. IEEE Trans Evol Comput 10(4):421–439

    Article  Google Scholar 

  44. Daulton S, Balandat M, Bakshy E (2020) Differentiable expected hypervolume improvement for parallel multi-objective Bayesian optimization. Adv Neural Inform Process Syst 33:9851–9864

    Google Scholar 

  45. Koch P, Wagner T, Emmerich MT, Bäck T, Konen W (2015) Efficient multi-criteria optimization on noisy machine learning problems. Appl Soft Comput 29:357–370

    Article  Google Scholar 

  46. Qin S, Sun C, ** Y, Zhang G (2019) Bayesian approaches to surrogate-assisted evolutionary multi-objective optimization: a comparative study. In: 2019 IEEE Symposium Series on Computational Intelligence (SSCI), pp 2074–2080

  47. Chugh T, ** Y, Miettinen K, Hakanen J, Sindhya K (2016) A surrogate-assisted reference vector guided evolutionary algorithm for computationally expensive many-objective optimization. IEEE Trans Evol Comput 22(1):129–142

    Article  Google Scholar 

  48. Sobester A, Forrester A, Keane A (2008) Engineering design via surrogate modelling: a practical guide. Wiley, West Sussex

    Google Scholar 

  49. Wilson J, Hutter F, Deisenroth M (2018) Maximizing acquisition functions for Bayesian optimization. In: Advances in Neural Information Processing Systems, vol. 31. Curran Associates, Inc

  50. Yarat S, Senan S, Orman Z (2021) A comparative study on PSO with other metaheuristic methods. Springer, Cham, pp 49–72

    Google Scholar 

  51. Li M, Yao X (2019) Quality evaluation of solution sets in multiobjective optimisation: a survey. ACM Comput Surv 52(2):26

    Google Scholar 

  52. López-Ibánez M, Paquete L, Stützle T (2010) Exploratory analysis of stochastic local search algorithms in biobjective optimization. In: Experimental Methods for the Analysis of Optimization Algorithms. Springer, Berlin, pp. 209–222. https://doi.org/10.1007/978-3-642-02538-9_9

Download references

Acknowledgements

This work was supported by the Flanders Artificial Intelligence Research Program (FLAIR), the Research Foundation Flanders (FWO Grant 1216021N), and Flanders Make vzw.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Alejandro Morales-Hernández.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix A: Constrained expected improvement (CEI)

The Constrained Expected Improvement (CEI [36]) uses one OK metamodel to approximate the expensive objective and one metamodel to approximate each expensive constraint independently. Then, for a constrained optimization problem with c constraints

$$\begin{aligned} \begin{array}{ll} min &{} \ y(\textbf{x}), \textbf{x} \in {\mathbb {R}} \\ \\ \mathrm{s.t.} &{} \ g_i(\textbf{x}) \le 0, \, i=1, 2, \dots , c. \end{array} \end{aligned}$$
(A1)

the objective value of point \(\textbf{x}\) can be treated as a Gaussian random variable \({\mathcal {N}}\left( \widehat{y}(\textbf{x}), \widehat{s}\mathbf (x) \right)\) and the ith constraint value of \(\textbf{x}\) can be also treated as a Gaussian random variable \({\mathcal {N}}\left( \widehat{g}_i(\textbf{x}), \widehat{e}_i\mathbf (x) \right) , \; i=1, 2, \dots , c\). For this, \(\widehat{y}\) and \(\widehat{s}\) are the GP prediction and standard error of the objective function, respectively, and \(\widehat{g}_i\) and \(\widehat{e}_i\) are the GP prediction and standard error of the ith constraint function respectively.

Then, from Eq. 2 we can transform the constraint and derive the Probability of Feasibility as

$$\begin{aligned} g(\textbf{x})&=0.5-Pf(\textbf{x}) \le 0 = Pf(\textbf{x}) \ge 0.5 \nonumber \\&= \frac{Pf(\textbf{x}) - \widehat{g} (\textbf{x})}{\widehat{e}(\textbf{x})}\ge \frac{0.5 - \widehat{g} (\textbf{x})}{\widehat{e}(\textbf{x})} \end{aligned}$$
(A2)
$$\begin{aligned} PoF(\textbf{x})&= Prob\left( \frac{Pf(\textbf{x}) - \widehat{g}(\textbf{x})}{\widehat{e}(\textbf{x})} \ge \frac{0.5 - \widehat{g}(\textbf{x})}{\widehat{e} (\textbf{x})} \right) \nonumber \\&= 1-Prob\left( \frac{Pf(\textbf{x}) - \widehat{g} (\textbf{x})}{\widehat{e}(\textbf{x})} \le \frac{0.5 - \widehat{g}(\textbf{x})}{\widehat{e} (\textbf{x})} \right) \nonumber \\&= 1-\Phi \left( \frac{0.5 - \widehat{g}(\textbf{x})}{\widehat{e} (\textbf{x})} \right) . \end{aligned}$$
(A3)

In case the GP standing for the objective and constraint function are mutually independent, the CEI can be obtained by combining the EI (to be consistent with our work we used MEI instead and the predictors defined in Eq. 12 and Eq. 13) and PoF as

$$\begin{aligned} \text {CMEI-SK}(\textbf{x})&= \text {MEI}(\textbf{x}) \times \text {PoF}(\textbf{x}) \nonumber \\&=\left[ \left( \widehat{f}_{SK}(\textbf{x}_{min}) -\widehat{f}_{SK}(\textbf{x})\right) \Phi \left( {\mathcal {D}} \right) + \widehat{s}_{OK}(\textbf{x})\phi \left( {\mathcal {D}} \right) \right] \nonumber \\&\quad \times \left[ 1-\Phi \left( \frac{0.5 - \widehat{g} (\textbf{x})}{\widehat{e}(\textbf{x})}\right) \right] \end{aligned}$$
(A4)

and

$$\begin{aligned} {\mathcal {D}} = \frac{\widehat{f}_{SK}(\textbf{x}_{min}) -\widehat{f}_{SK}(\textbf{x})}{\widehat{s}_{OK}(\textbf{x})}. \end{aligned}$$
(A5)
Fig. 12
figure 12

Evolution of the mean hypervolume throughout the optimization (of 50 macro-replications). The reference point [production cost = 3, break strength = 4] is used to compute this metric

Note that this equation is different from our proposed Eq. 19 in the derivation of the PoF. Figure 12 shows that our proposed cMEI-SK got on average better Pareto fronts (higher hypervolume values) than that of CEI. This suggests that considering the uncertainty predicted by the GP may lead the optimization to points that do not cause an increase in the hypervolume. This was more evident when the improvement was measured with MEI (and fitting only one GP to the scalarized objectives).

Appendix B: Wilcoxon test results

See Fig. 13

Fig. 13
figure 13

Wilcoxon test results for significant differences between algorithms, using a hypervolume and b IGD+ metrics. Pink color indicate no significant differences (\(p\_value \ge 5 \%\))

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Morales-Hernández, A., Rojas Gonzalez, S., Van Nieuwenhuyse, I. et al. Bayesian multi-objective optimization of process design parameters in constrained settings with noise: an engineering design application. Engineering with Computers (2024). https://doi.org/10.1007/s00366-023-01922-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s00366-023-01922-8

Keywords

Navigation