This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Non-Myopic Multi-Objective Bayesian Optimization

Syrine Belakaria∗1             Alaleh Ahmadianshalchi∗2             Barbara Engelhardt1
Stefano Ermon1             Janardhan Rao Doppa2

1Stanford University           2Washington State University
Abstract

We consider the problem of finite-horizon sequential experimental design to solve multi-objective optimization (MOO) of expensive black-box objective functions. This problem arises in many real-world applications, including materials design, where we have a small resource budget to make and evaluate candidate materials in the lab. We solve this problem using the framework of Bayesian optimization (BO) and propose the first set of non-myopic methods for MOO problems. Prior work on non-myopic BO for single-objective problems relies on the Bellman optimality principle to handle the lookahead reasoning process. However, this principle does not hold for most MOO problems because the reward function needs to satisfy some conditions: scalar variable, monotonicity, and additivity. We address this challenge by using hypervolume improvement (HVI) as our scalarization approach, which allows us to use a lower-bound on the Bellman equation to approximate the finite-horizon using a batch expected hypervolume improvement (EHVI) acquisition function (AF) for MOO. Our formulation naturally allows us to use other improvement-based scalarizations and compare their efficacy to HVI. We derive three non-myopic AFs for MOBO: 1) the Nested AF, which is based on the exact computation of the lower bound, 2) the Joint AF, which is a lower bound on the nested AF, and 3) the BINOM AF, which is a fast and approximate variant based on batch multi-objective acquisition functions. Our experiments on multiple diverse real-world MO problems demonstrate that our non-myopic AFs substantially improve performance over the existing myopic AFs for MOBO.

**footnotetext: These authors contributed equally to this work

1 Introduction

Many engineering design and scientific applications involve performing a sequence of experiments under a small-resource budget to optimize multiple expensive-to-evaluate objective functions. Consider the example of searching nanoporous materials optimized for hydrogen-powered vehicles. In this real-world problem, we need to perform physical lab experiments to make a candidate material and evaluate its volumetric and gravimetric hydrogen uptake properties. Our goal is to approximate the optimal Pareto front of solutions under a given small resource budget for experiments.

We consider the problem of finite-horizon sequential experimental design (SED) to solve multi-objective optimization (MOO) problems over expensive-to-evaluate objectives to find high-quality Pareto solutions. We solve this problem within the framework of Bayesian optimization (BO) Shahriari et al. [2015], which has been shown to be sample-efficient in practice. The key idea behind BO is to train a surrogate model (e.g., a Gaussian process) from past experimental data, and intelligently select the future experiments guided by an acquisition function. Most of the prior acquisition functions for BO are myopic in nature (i.e., focused on the immediate advancement of the goal). There are a few non-myopic acquisition functions for BO in the single-objective setting Jiang et al. [2020a], Lam and Willcox [2016]. These non-myopic strategies rely on formulating the optimal policy for SED as a dynamic program using the recursive Bellman optimality principle, and then solving for the optimal policy using methods with varying approximations (e.g., rollout strategies).

This work studies the design of non-myopic acquisition functions for MOO problems for the first time. The Bellman optimality principle does not hold in most MOO problems and requires the reward function of the underlying Markov decision process (MDP) to satisfy some conditions Roijers et al. [2013], Van Moffaert and Nowé [2014] 1) only scalar rewards; 2) monotonically increasing rewards with respect to better actions; and 3) additivity condition—for the optimality to hold. These conditions are challenging to satisfy in MOO problems and require a careful design of the reward function. We address this challenge by using hypervolume improvement (HVI) as our scalarization approach to define the reward function. Unlike the hypervolume function, hypervolume improvement satisfies the additivity condition. A key factor for these conditions to hold is to define optimality as Pareto optimality with respect to HVI rather than individual optimality for each objective function. This reward function allows us to use a lower-bound on the Bellman equation to approximately solve the finite-horizon SED problem using a batch utility function for MOO (e.g., qEHVI Daulton et al. [2020]). Our approach generalizes to other scalarization approaches that satisfy the above conditions.

We derive three non-myopic multi-objective (NMMO) acquisition functions (AFs) for MOO with varying speed and accuracy trade-offs: 1) the NMMO-Nested AF, based on exact computation of the lower bound and computationally intractable even for small horizons, 2) the NMMO-Joint AF, a scalable lower bound on the nested AF, and 3) the BINOM AF, a fast and approximate version of EHVI. Our experiments on real-world MOO problems demonstrate that the proposed non-myopic acquisition functions substantially improve performance over the baseline BO methods for MOO.

Contributions. The key contribution of our work is the development and evaluation of principled non-myopic acquisition functions to solve MOO problems. In particular, we

  • validate the use of a lower-bound on the Bellman optimality principle for MOO problems by using hypervolume improvement (HVI) to define an appropriate reward function;

  • develop three non-myopic acquisition functions based on the derived lower bound—NMMO-Nested, NMMO-Joint, and BINOM—with varying speed and accuracy trade-offs;

  • instantiate our framework with several information gain-based scalarizations and compare with HVI.

  • experimentally evaluate the proposed non-myopic acquisition functions and baseline BO methods on multiple real-world MOO problems. Our code is provided at https://github.com/Alaleh/NMMO.

2 Problem Setup and Background

This section formally defines the problem setting and provides the background necessary for our solution approach.

Multi-objective optimization (MOO). Let 𝔛d\mathfrak{X}\subset\mathbb{R}^{d} be the input space of dd design variables, where each candidate input 𝐱𝔛\mathbf{x}\in\mathfrak{X} is a dd-dimensional vector. Let {f1,,fK}\{f_{1},\cdots,f_{K}\} with K2K\geq 2 represent the black-box objective functions defined over the input space 𝔛\mathfrak{X}, where f1(𝐱),,fK(𝐱):𝔛f_{1}(\mathbf{x}),\cdots,f_{K}(\mathbf{x}):\mathfrak{X}\rightarrow\mathbb{R}. We denote the function evaluation at 𝐱\mathbf{x} as 𝐲=[y1,,yK]\mathbf{y}=[y_{1},\cdots,y_{K}], where yi=fi(𝐱)y_{i}=f_{i}(\mathbf{x}) for all i{1,,K}i\in\{1,\cdots,K\}. For example, to optimize nanoporous materials for hydrogen-powered vehicles, 𝐱\mathbf{x} represents a candidate material, f1(𝐱)f_{1}(\mathbf{x}) and f2(𝐱)f_{2}(\mathbf{x}) corresponds to a physical lab experiment to evaluate the volumetric and gravimetric hydrogen uptake of the material. WLOG, we assume maximization for all KK objective functions.

An input 𝐱\mathbf{x} Pareto-dominates another input 𝐱\mathbf{x}^{\prime} if and only if i:fi(𝐱)fi(𝐱)\forall i:f_{i}(\mathbf{x})\geq f_{i}(\mathbf{x}^{\prime}) and j:fj(𝐱)>fj(𝐱)\exists j:f_{j}(\mathbf{x})>f_{j}(\mathbf{x}^{\prime}). The optimal solution of the MOO problem is a set of inputs 𝒳𝔛\mathcal{X}^{*}\subset\mathfrak{X} such that no input 𝐱𝔛𝒳\mathbf{x}\in\mathfrak{X}\setminus\mathcal{X}^{*} Pareto-dominates another input 𝐱𝒳\mathbf{x}^{\prime}\in\mathcal{X}^{*}. The set of input solutions 𝒳\mathcal{X}^{*} is called the optimal Pareto set, and the corresponding set of function values 𝒴\mathcal{Y}^{*} is called the optimal Pareto front. In SED, we select one input for evaluation in each iteration, and our goal is to uncover a high-quality Pareto front while minimizing the total number of expensive function evaluations. A typical measure to evaluate the quality of a Pareto front is the Pareto hypervolume (PHV) indicator Zitzler and Thiele [1999] of the Pareto front with respect to a reference point 𝐫\mathbf{r}.

Non-myopic MOO problem. Our goal is to approximate the optimal Pareto front for a MOO problem within a maximum experimental budget TT, which is typically small (i.e., finite-horizon SED). This problem can be formulated as:

maxXP(𝔛)max𝐱Xf1(𝐱)fK(𝐱), s.t. |X|=T,\max_{X\in P(\mathfrak{X})}\max_{\mathbf{x}\in X}f_{1}(\mathbf{x})\cdots f_{K}(\mathbf{x}),\ \ \ \ \text{~{}s.t.~{}}|X|=T, (1)

where P(𝔛)P(\mathfrak{X}) denotes the power set of 𝔛\mathfrak{X} and X={𝐱1𝐱T}X=\{\mathbf{x}_{1}\dots\mathbf{x}_{T}\} is the set of inputs selected for function evaluation. We solve this problem using a non-myopic decision policy, where, at each iteration, the algorithm selects one input for evaluation while reasoning about candidate experimental design plans within the budget TT.

Overview of Bayesian optimization (BO). BO is an effective framework for solving expensive black-box optimization problems. BO has three main steps: (i) conducting an expensive experiment with an input selected by BO, (ii) updating the surrogate model by including the new evaluations, and (iii) selecting a new candidate input for the next function evaluation. The two key components of BO are the surrogate model and an acquisition function. Surrogate models are probabilistic models (typically Gaussian processes Williams and Rasmussen [2006]) that are trained on data from past function evaluations. A surrogate model is able to predict the output of unknown inputs without requiring an expensive experiment and to quantify the uncertainty of its predictions. The acquisition function (e.g., expected improvement (EI) Mockus [1989]) is used to decide which input to select for the next function evaluation. It uses the surrogate model to score the utility of selecting inputs for the next experiment by trading off exploitation and exploration. The goal is to quickly direct the search toward high-quality solutions with the least number of iterations. BO in a multi-objective setting requires particular consideration for modeling and acquisition function design. For MOO problems, we model the objective functions f1,,fKf_{1},\cdots,f_{K} using KK independent GP models 𝒢𝒫1,,𝒢𝒫K\mathcal{GP}_{1},\cdots,\mathcal{GP}_{K} and discuss acquisition functions in Section 3.

Non-myopic decision policies for BO. We review important facts for optimal sequential decision-making Osborne and Osborne [2010]. Consider collecting tt observations DtD_{t}, and let uu denote a utility function defining a quality metric over the data DtD_{t}. The marginal gain in the utility of query 𝐱\mathbf{x} w.r.t. DtD_{t} is expressed as

u(𝐲|𝐱,Dt)=u(Dt(𝐱,𝐲))u(Dt).u(\mathbf{y}|\mathbf{x},D_{t})=u(D_{t}\cup(\mathbf{x},\mathbf{y}))-u(D_{t}). (2)

Consider the case where TtT-t steps are remaining. The expected marginal gain for TtT-t-steps is expressed through the Bellman recursion Jiang et al. [2020a] as

𝒰Tt(𝐱|Dt)=𝔼𝐲[u(𝐲|𝐱,Dt)]+𝔼𝐲[max𝐱𝒰Tt1(𝐱|Dt(𝐱,𝐲))].\mathcal{U}_{T-t}(\mathbf{x}|D_{t})=\mathbb{E}_{\mathbf{y}}[u(\mathbf{y}|\mathbf{x},D_{t})]+\mathbb{E}_{\mathbf{y}}\left[\max_{\mathbf{x}^{\prime}}\mathcal{U}_{T-t-1}(\mathbf{x}^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))\right]. (3)

The optimal expected policy to select the TtT-t lookahead steps can be obtained by sequentially maximizing Equation 3 at each step

𝐱=argmax𝐱𝔛𝒰Tt(𝐱|Dt).\mathbf{x}^{*}=\underset{\mathbf{x}\in\mathfrak{X}}{\mathrm{argmax}}~{}\mathcal{U}_{T-t}(\mathbf{x}|D_{t}). (4)

Being a sequence of TtT-t nested integrals, solving the optimization problem in equation 4 to optimality is intractable for even a small value of TtT-t Lam et al. [2016], Jiang et al. [2020b].

Recent work made a connection between non-myopic input selection and batch input selection Jiang et al. [2020a]. Assuming parallel evaluation, the expected marginal utility of a new batch of evaluations Y={𝐲t,,𝐲T}Y=\{\mathbf{y}_{t},\cdots,\mathbf{y}_{T}\} given their associated batch of inputs X={𝐱t,,𝐱T}X=\{\mathbf{x}_{t},\cdots,\mathbf{x}_{T}\} is defined as:

𝒰(X|Dt)=𝔼Y[u(Y|X,Dt)]\mathcal{U}(X|D_{t})=\mathbb{E}_{Y}[u(Y|X,D_{t})] (5)

Respectively, the maximum expected marginal utility of a batch of evaluations at input locations XX^{\prime} of size |X|=Tt1|X^{\prime}|=T-t-1 conditioned on the addition of an observation (𝐱,𝐲)(\mathbf{x},\mathbf{y}) to the data is defined as

maxX:|X|=Tt1𝔼𝐲[𝒰(X|Dt(𝐱,𝐲))].\max_{X^{\prime}:|X^{\prime}|=T-t-1}\mathbb{E}_{\mathbf{y}}[\mathcal{U}(X^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))]. (6)

Comparing equations 3 and 5, the second term in equation 3 has an adaptive utility with a nested maximization while equation 5 has a joint batch utility with an outer maximization. This leads to a lower-bound connection between the two:

maxX:|X|=Tt1𝔼𝐲[𝒰(X|Dt(𝐱,𝐲))]𝔼𝐲[max𝐱𝒰Tt1(𝐱|Dt(𝐱,𝐲))].\max_{X^{\prime}:|X^{\prime}|=T-t-1}\mathbb{E}_{\mathbf{y}}[\mathcal{U}(X^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))]\leq\mathbb{E}_{\mathbf{y}}[\max_{\mathbf{x}^{\prime}}\mathcal{U}_{T-t-1}(\mathbf{x}^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))]. (7)

Consequently, we obtain a lower bound on the expected marginal gain defined by the Bellman recursion in equation 3

α(𝐱|Dt)𝒰Tt(𝐱|Dt)\alpha(\mathbf{x}|D_{t})\leq\mathcal{U}_{T-t}(\mathbf{x}|D_{t}) (8)
withα(𝐱|Dt)=𝔼𝐲[u(𝐲|𝐱,Dt)]+maxX:|X|=Tt1𝔼𝐲[𝒰(X|Dt(𝐱,𝐲))].\text{with}~{}~{}~{}~{}\alpha(\mathbf{x}|D_{t})=\mathbb{E}_{\mathbf{y}}[u(\mathbf{y}|\mathbf{x},D_{t})]+\max_{X^{\prime}:|X^{\prime}|=T-t-1}\mathbb{E}_{\mathbf{y}}[\mathcal{U}(X^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))]. (9)

The lower bound on the Bellman equation (Equation 9) provides an efficient alternative for optimizing the sequential decision-making process in non-myopic single-objective BO Jiang et al. [2020a], where the utility function u(Dt)u(D_{t}) is defined as the incumbent, and the marginal utility u(y|𝐱,Dt)u(y|\mathbf{x},D_{t}) as the incumbent improvement. However, applying the same strategy in the multi-objective BO (MOBO) setting is non-trivial. The Bellman equation is inherently single-dimensional with respect to the output of the utility function and, therefore, presents challenges when faced with the multi-output nature of MOBO, where conflicting objectives require simultaneous consideration. This necessitates an adapted formulation capable of capturing the trade-offs across multiple objectives and guiding the exploration and exploitation balance. The methodological challenges to address the extension of the Bellman equation for MOBO and the proposed solutions are described in Section 4.

3 Related Work

Most of the prior work on acquisition functions in BO falls under the myopic category. There is less work on non-myopic acquisition functions with a focus on single-objective BO.

Myopic single-objective BO. Recent advances in single-objective BO have predominantly focused on refining and extending traditional acquisition functions. The most notable techniques Garnett [2023], Wang and Jegelka [2017], Hernández-Lobato et al. [2014], Ament et al. [2024] use Gaussian processes as surrogate models.

Non-myopic single-objective BO. Non-myopic approaches in single-objective BO Jiang et al. [2020a], Lam and Willcox [2016], González et al. [2016], Ginsbourger and Le Riche [2009], Osborne and Osborne [2010], Osborne et al. [2009] represent an emerging direction aimed at overcoming the limitations of myopic acquisition functions, which typically optimize for immediate gains. These lookahead methods consider the future impact of current decisions Wu and Frazier [2019], Lee et al. [2021], thereby enhancing long-term optimization outcomes when the experimental resource budget is small.

Multi-objective BO. When compared to single-objective BO, multi-objective BO (MOBO) has received relatively less attention. The predominant methods for designing MOBO strategies include information-theoretic methods Tu et al. [2022], Hernández-Lobato et al. [2016], Belakaria et al. [2019], Suzuki et al. [2020], Garrido-Merchán and Hernández-Lobato [2021], uncertainty based methods Belakaria et al. [2020b], hypervolume-based methods Konakovic Lukovic et al. [2020], Daulton et al. [2020], Ahmadianshalchi et al. [2024a], and scalarization-based methods Daulton et al. [2022], Knowles [2006]. The joint entropy search for MOBO (JESMO) algorithm Tu et al. [2022] aims to maximize information gain about the optimal Pareto front by evaluating inputs that maximally reduce its uncertainty. The expected hypervolume improvement (EHVI) method Daulton et al. [2020] extends the concept of EI from a single-objective to a MOO problem by assessing the potential improvement in hypervolume from new input evaluations. Several efforts to extend these approaches to handle constraints and preferences have resulted in more general frameworks that accommodate complex real-world decision-making scenarios Garrido-Merchán and Hernández-Lobato [2019], Belakaria et al. [2020a], Ahmadianshalchi et al. [2024b].

Non-myopic MOBO. Despite advances in myopic strategies for MOBO, a notable knowledge gap remains. There is no prior work on non-myopic acquisition functions for MOBO problems with a finite horizon. A recent approach proposed for settings with decoupled function evaluation Daulton et al. [2023] considers a one-step lookahead only and thus lacks strategic foresight. Our goal is to precisely fill this critical gap in the current state of knowledge.

4 Non-Myopic Multi-Objective BO

In this section, we provide details of our proposed approach for non-myopic MOBO. We first discuss the challenge of maintaining the validity of the Bellman optimality principle and, by consequence, the ability to use the Bellman equation and its lower-bound to solve the SED problem in the multi-objective setting. We then discuss design choices for the marginal utility function that enables the effective use of the Bellman equation in our setting. We finally provide the details of the non-myopic acquisition functions we propose, their computational challenges, and practical algorithms.

4.1 Challenges of Non-Myopic MOBO

For the Bellman optimality principle to hold and for the optimal policy to be computable by solving the Bellman equation, the reward function has to satisfy several criteria. In our problem setting, we refer to the reward and the marginal gain in utility defined in equation 2 interchangeably. We state below some of the reward function requirements that are not straightforward to satisfy when dealing with a multi-objective problem:

  • Scalarization: the reward has to be defined as a single scalar value and not a vector of values Boutilier et al. [1999], Roijers et al. [2013]. In single-objective optimization, the reward is defined as the improvement in the best-found function value Lam et al. [2016], Jiang et al. [2020a]. However, in MOBO problems, every input evaluation leads to a vector of values representing a trade-off where some of the objectives may improve while others degrade.

  • Monotonicity: the reward has to be a monotonically increasing function with respect to the quality of the actions (i.e., better actions lead to higher rewards). This monotonic property ensures that the optimization trajectory is aligned with improving performance Roijers et al. [2013].

  • Additivity: the reward must satisfy the additivity condition, i.e, the total reward is a sum of the rewards obtained at each step of the decision-making process Boutilier et al. [1999], Roijers et al. [2013], Van Moffaert and Nowé [2014]. Additivity is essential for the recursive decomposition of the sequential decision process.

It is important to note that the choice of the scalarization approach plays a key role in preserving the monotonicity and additivity conditions. In MOBO problems, the challenge lies in combining both the improvement and degradation incurred in several objectives into a single, quantifiable metric that accurately reflects the quality of an input in terms of Pareto dominance regardless of trade-offs in individual objectives. In broader applications of the Bellman equation (e.g., multi-objective reinforcement learning), linear scalarization is the most common technique known to preserve the additivity of the reward, where each objective is multiplied by a predetermined nonnegative weight reflecting its relative importance Barrett and Narayanan [2008]. Since linear scalarization computes a convex combination, it has the known fundamental limitation that it cannot recover solutions laying outside the convex regions of the Pareto front Van Moffaert and Nowé [2014].

4.2 Our Proposed Non-Myopic MOBO Method

In this section, we provide the details of the design choices of our utility function that enable the application of the Bellman equation in the MO setting. We then provide the details of the different acquisition functions we propose for non-myopic MOBO.

4.2.1 Utility for Multi-Objective Non-Myopic Setting

The hypervolume (HV) quality indicator provides a scalar measure of a particular Pareto front 𝒴\mathcal{Y}. HV is the volume between a reference point and the Pareto front. This measure has been of particular interest in MOBO problems because it is known to be strictly monotonically increasing with regard to Pareto dominance. However, the HV measure does not satisfy the additivity condition necessary for Bellman’s optimality principle Van Moffaert et al. [2013]. In this work, we propose to use the hypervolume improvement (HVI) to scalarize our multiple single-objective rewards defined as the improvement of each of the functions. This approach quantifies the quality of a new point 𝐲\mathbf{y} by the volume of the objective space that is newly encompassed by extending the Pareto front to include the new point.

HVI(𝐲|𝒴)=HV(𝒴𝐲)HV(𝒴)HVI(\mathbf{y}|\mathcal{Y})=HV(\mathcal{Y}\cup\mathbf{y})-HV(\mathcal{Y}) (10)

HVI provides a suitable scalarization because it naturally satisfies all of the reward requirements. First, it integrates the improvement of multiple objectives into a single scalar value. Second, it preserves the monotonicity by reflecting the Pareto quality of a new input (action) given the current Pareto front (state space representation).

Intuitively, HVI satisfies the additivity condition because the improvement in hypervolume by successive input evaluations (actions) can be summed to give a total improvement over the sequence of evaluations, thus adhering to the required recursive structure.

Lemma 1.

We denote HVItHVI_{t} the hypervolume improvement at tt with t[1,T]t\in[1,T] and HVItotalHVI_{total} be total hypervolume improvement collected at iteration TT. The additivity condition is satisfied because the total hypervolume collected at iteration TT can be decomposed as a summation of the intermediate hypervolume improvements

HVItotal=t=1t=THVItHVI_{total}=\sum_{t=1}^{t=T}HVI_{t} (11)

We provide a proof of lemma 1 in Appendix 8. It is important to note that the Bellman optimality principle holds with respect to HVI and Pareto optimality defined by the HV measure, rather than with respect to each objective function separately. This definition of optimality aligns well with the goals of MOBO, where the focus is on maximizing the objective space coverage efficiently, and where defining optimality with respect to HV and using HV as the primary evaluation metric is the most common approach Belakaria et al. [2019], Daulton et al. [2021]. We make this distinction to avoid confusion with the multi-objective RL setting, where theoretical optimality guarantees with respect to each function independently might not hold when using nonlinear scalarizations Roijers et al. [2013].

4.2.2 Proposed Acquisition Functions

Given our choice of reward as the HVI, we can define our SED problem using the Bellman equation. We formally define the marginal gain in the utility of our multi-objective setting as follows:

u(𝐲|𝐱,Dt)=HVI(𝐲|𝒴).u(\mathbf{y}|\mathbf{x},D_{t})=HVI(\mathbf{y}|\mathcal{Y}). (12)

Similar to the single-objective case, computing the optimal policy by fully solving the Bellman equation to optimality is intractable even for a moderately large horizon TtT-t Jiang et al. [2020b]. An effective approach to navigating this complexity is to approximate the optimal policy by optimizing the lower-bound on the Bellman equation (Equation 8). In what follows, we denote by XX the full remaining horizon at iteration tt with |X|=Tt|X|=T-t, 𝐱\mathbf{x} the next input to evaluate and the first input in the horizon XX, and XX^{\prime} the lookahead horizon with |X|=Tt1|X^{\prime}|=T-t-1 and X={𝐱,X}X=\{\mathbf{x},X^{\prime}\}. The selection of the next input to evaluate while accounting for the lookahead horizon (the remaining sequence of inputs in the horizon) can be achieved by maximizing the acquisition function α(𝐱|Dt)\alpha(\mathbf{x}|D_{t}) (Equation 9). By defining the marginal gain in utility as the HVI, we rewrite the acquisition function as:

α(𝐱|Dt)=EHVI(𝐱|Dt)+maxX:|X|=Tt1𝔼𝐲[BEHVI(X|Dt(𝐱,𝐲))]\alpha(\mathbf{x}|D_{t})=EHVI(\mathbf{x}|D_{t})+\max_{X^{\prime}:|X^{\prime}|=T-t-1}\mathbb{E}_{\mathbf{y}}[BEHVI(X^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))] (13)

The first term on the r.h.s is the expected hypervolume improvement (EHVI) for the input 𝐱\mathbf{x} given the observed data DtD_{t}. The second term is the maximum of the expectation over the batch expected hypervolume improvement (BEHVI) computed at the batch of inputs XX^{\prime} conditioned on the addition of the new observation (𝐱,𝐲)(\mathbf{x},\mathbf{y}) to the data. The expectation is computed with respect to the possible values of 𝐲\mathbf{y} sampled from the posterior of the surrogate models at the input 𝐱\mathbf{x}. We refer to this new acquisition function as the nested non-myopic multi-object acquisition function. Since the first part corresponds to the EHVI acquisition function Emmerich and Klinkenberg [2008], which corresponds to the one-step myopic policy, input selection by optimizing this lower bound is always at least as tight as optimizing the myopic policy. This approach, while approximate, leverages the strengths of the EHVI and BEHVI acquisition functions Emmerich and Klinkenberg [2008], Daulton et al. [2020] to efficiently compute the immediate impact of the selection of input 𝐱\mathbf{x} and to approximate the impact of the selection of 𝐱\mathbf{x} on the lookahead horizon XX^{\prime}.

Algorithm 1 NMMO Algorithm

Input: input space 𝔛\mathfrak{X}; KK blackbox objective functions f1(x),f2(x),,fK(x)f_{1}(x),f_{2}(x),\cdots,f_{K}(x); and maximum no. of iterations TT, selected non-myopic method {\in\{NMMO-Joint, NMMO-Nested, BINOM}\}.

1:Initialize GP models 𝒢𝒫1,𝒢𝒫2,,𝒢𝒫K\mathcal{GP}_{1},\mathcal{GP}_{2},\cdots,\mathcal{GP}_{K} by evaluating at N0N_{0} initial points
2:for each iteration tt = 11 to TT do
3:     if method==NMMO-Nested then
4:         Select 𝐱targmax𝐱𝔛α(𝐱|Dt)\mathbf{x}_{t}\leftarrow\arg max_{\mathbf{x}\in\mathfrak{X}}\hskip 5.69054pt\alpha(\mathbf{x}|D_{t}), where α\alpha is computed from Equation 13      
5:     if method==NMMO-Joint then
6:         Select 𝐱t,Xargmax𝐱𝔛α(𝐱,Xt|Dt)\mathbf{x}_{t},X^{{}^{\prime}}\leftarrow\arg max_{\mathbf{x}\in\mathfrak{X}}\hskip 5.69054pt\alpha(\mathbf{x},X_{t}^{{}^{\prime}}|D_{t}), where α\alpha is computed from Equation 4.2.2      
7:     if  method==BINOM then
8:         Select Xargmax𝐱𝔛α(X|Dt)X\leftarrow\arg max_{\mathbf{x}\in\mathfrak{X}}\hskip 5.69054pt\alpha(X|D_{t}), where α\alpha is computed from Equation 15
9:         Select 𝐱targmax𝐱XEHVI(𝐱|Dt)\mathbf{x}_{t}\leftarrow\arg max_{\mathbf{x}\in X}\hskip 5.69054ptEHVI(\mathbf{x}|D_{t})      
10:     Evaluate 𝐱t\mathbf{x}_{t}: 𝐲t(f1(𝐱t),,fK(𝐱t))\mathbf{y}_{t}\leftarrow(f_{1}(\mathbf{x}_{t}),\cdots,f_{K}(\mathbf{x}_{t}))
11:     Aggregate data: Dt+1Dt{(𝐱t,𝐲t)}D_{t+1}\leftarrow D_{t}\cup\{(\mathbf{x}_{t},\mathbf{y}_{t})\}
12:     Update models 𝒢𝒫1,𝒢𝒫2,,𝒢𝒫K\mathcal{GP}_{1},\mathcal{GP}_{2},\cdots,\mathcal{GP}_{K}
13:end for
14:return Pareto front of f1(x),,fK(x)f_{1}(x),\cdots,f_{K}(x) based on DTD_{T}

Computational challenges. The expression of the acquisition function (Equation 13) involves two computational challenges for which we propose solutions.

1. Nested Maximization: The second term in the nested non-myopic AF for MOO (Equation 13) is defined as a maximization problem. Thus, evaluating the acquisition function for each candidate input 𝐱\mathbf{x} requires solving an internal optimization problem to estimate the batch representing the lookahead horizon. This leads to a computationally prohibitive acquisition function optimization. To address this challenge, we propose two alternative acquisition functions:

Alternative lower-bound computation: One approach to circumvent the nested maximization problem is to reformulate the acquisition function as a new lower-bound that jointly optimizes over both the first input in the horizon 𝐱\mathbf{x} and the lookahead horizon XX^{\prime}. We refer to this approach as the joint non-myopic MO acquisition function.

α(𝐱,X|Dt)\displaystyle\alpha(\mathbf{x},X^{\prime}|D_{t}) =𝔼𝐲[u(𝐲|𝐱,Dt)]+𝔼𝐲[𝒰(X|Dt(𝐱,𝐲))]\displaystyle=\mathbb{E}_{\mathbf{y}}[u(\mathbf{y}|\mathbf{x},D_{t})]+\mathbb{E}_{\mathbf{y}}[\mathcal{U}(X^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))]
=EHVI(𝐱|Dt)+𝔼𝐲[BEHVI(X|Dt(𝐱,𝐲))]\displaystyle=EHVI(\mathbf{x}|D_{t})+\mathbb{E}_{\mathbf{y}}[BEHVI(X^{\prime}|D_{t}\cup(\mathbf{x},\mathbf{y}))] (14)

Approximation via batch selection: Approximating the selection of all the inputs in the horizon by a batch acquisition function has been previously used for non-myopic single-objective BO González et al. [2016], Jiang et al. [2020a]. Let XX^{*}, s.t |X|=Tt|X^{*}|=T-t be a set of inputs selected by maximizing a batch acquisition function 𝒰(X|Dt)\mathcal{U}(X|D_{t}). Jiang et al. [2020a] showed that choosing any input 𝐱X\mathbf{x}^{*}\in X^{*} can be approximated by selecting inputs that maximize α(𝐱|Dt)\alpha(\mathbf{x}|D_{t}) several times, with α\alpha as the acquisition function in equation 9. With this approximation, we propose to follow Jiang et al. [2020a], and approximate the non-myopic acquisition function by a fully joint batch acquisition function. We select the full horizon by maximizing the BEHVI. We call this selection approach Batch-Informed NOnmyopic Multi-objective optimization (BINOM):

α(X|Dt)=𝒰(X|Dt)=BEHVI(X|Dt).\alpha(X|D_{t})=\mathcal{U}(X|D_{t})=BEHVI(X|D_{t}). (15)

However, an important issue arises. The strategy does not differentiate the next input for evaluation 𝐱\mathbf{x}^{*} from the lookahead horizon XX^{\prime} in the way the acquisition functions in equations 9 and 4.2.2 do. By analogy to Jiang et al. [2020a], we select the next input to evaluate as the input with the highest immediate EHVI among the inputs in the selected batch.

2. Expectation computation via posterior sampling: The second term of the acquisition function (Equation 9) requires an expectation computed over the BEHVI. This expectation can be estimated by sampling several realizations of 𝐲\mathbf{y} from the posterior of the surrogate models. The accuracy of the expectation estimation depends on the number of samples used. In practice, more samples lead to a slower and more computationally expensive acquisition function, while fewer samples leads to inaccurate estimation and a higher variance. This results in unstable optimization behavior with unreliable outcomes. To mitigate the former issues, we propose to substitute the expectation by the use of a one-step lookahead GP model Lyu et al. [2021]. In this approach, the mean function of the GP remains unchanged, while the variance is updated based on the posterior conditioning on the newly added input.

In summary, these computational strategies are crucial for enhancing the efficiency and stability of the acquisition functions in multi-objective optimization scenarios. With these strategies, our decision-making process for finite-horizon sequential experimental design using MOBO becomes computationally tractable and more robust.

Alternative scalarization approaches The proof of additivity in Lemma 1 relies on the fact that HVI is an improvement-based scalarization. Therefore, this lemma theoretically holds for any improvement-based scalarization. Notably, information gain (IG) based scalarizations, defined as the decrease in the entropy of the Pareto front/set, can indeed satisfy the additivity condition. IG can be computed with respect to the input space (leading to an instantiation of our method with PESMO Hernández-Lobato et al. [2016]), the output space (leading to an instantiation of our method with MESMO Belakaria et al. [2019, 2021]), or both (leading to an instantiation of our method with JESMO Tu et al. [2022]). However, it is important to point out that the IG monotonicity with respect to better actions might not always hold. Typically, an input is considered a better choice/action, if it dominates existing points in the Pareto front and thus enhances the quality of the Pareto front. However, an input can have a high information gain even if it is not Pareto dominating other inputs, but it provides high information about a new region in the search space. This conflicting issue does not occur with HVI, since HVI is always monotonically increasing with respect to Pareto dominance.

Nevertheless, we provide an ablation study comparing varaints of our acquisition functions with EHVI, MESMO, and JESMO based scalarization (Appendix Section 7.3).

5 Experiments and Results

This section describes our experimental evaluation comparing the proposed non-myopic methods with baselines on several real-world and synthetic MOO problems.

Benchmark MOO problems: We provide a brief description of our real-world MOO benchmarks, including the number of input dimensions (dd) and the number of objectives (KK).

Metal-organic Framework Design (dd = 7, KK = 2) Kitagawa et al. [2014], Boyd et al. [2019]: Nanoporous materials (NPMs) Lu and Zhao [2004] represent a class of structures characterized by their microscopic pores at the nanoscale level. Metal-organic frameworks (MOFs) Ahmed et al. [2017], Ahmed and Siegel [2021], Boyd et al. [2019] are a class of NPMs, instrumental in several sustainable applications due to their selective gas absorption capabilities. The goal of this problem is to identify the Pareto front of MOFs from \approx 100K candidates that optimize gravimetric and volumetric hydrogen uptake for hydrogen-powered vehicles.

Reinforced Concrete Beam Design (dd = 3, KK = 2) Amir and Hasegawa [1989]: The goal is to optimize the cost of concrete and reinforcing steel used in the construction of a beam. Its input features are the area of reinforcement, the beam’s width, and its depth. This setup enabled us to explore a mixed input space to find cost-effective yet robust designs.

Four-Bar Truss Design (dd = 4, KK = 2) Cheng and Li [1999]: The goal is to minimize the structural volume and joint displacement of a four-bar truss to reduce the weight of the structure and the displacement of a specific node. The features are the areas of the four-member cross-sections.

Gear Train Design (dd = 4, KK = 3) Deb and Srinivasan [2006], Tanabe and Ishibuchi [2020]: The goal is to design a compound four-gear train to achieve a specific target gear ratio between the driver and driven shafts. The objectives are to minimize the error between the obtained gear ratio and the desired gear ratio and to minimize the maximum size of the four gears to optimize the compactness of the design.

Welded Beam Design (dd = 4, KK = 3) Coello and Montes [2002], Rao [2019]: The goal is to minimize the cost of fabrication and the end deflection of a welded beam by precisely adjusting four of its side lengths. To focus more directly on the relationship between the beam’s dimensions and the optimization objectives, we adopted a non-constrained version of this problem as in prior work Tanabe and Ishibuchi [2020], Konakovic Lukovic et al. [2020].

Disc Break Design (dd = 4, KK = 3) Ray and Liew [2002], Tanabe and Ishibuchi [2020]: The goals are to minimize the mass of the brake and the stopping time. The variables are the inner and outer radius of the discs, the engaging force, and the number of friction surfaces.

We also include the ZDT-3 (dd = 9, KK = 2) Zitzler et al. [2000] problem as a synthetic MOO benchmark. We provide additional results in the Appendix (Section 7.4) and an overview of all our MOO benchmarks(Appendix Table LABEL:tab:benchmark_details).

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 2: Hypervolume results on real-world problems: Non-myopic methods vs. EHVI and JESMO.

Baselines: We proposed three non-myopic acquisition functions (NMMO-Joint, NMMO-Nested, and BINOM) with varying trade-offs between computational complexity and approximation accuracy. In an ideal scenario with a perfect surrogate model, evaluating the entire lookahead horizon H=TtH=T-t at each iteration tt would yield the closest results to the optimal policy. However, previous research on non-myopic Bayesian optimization [Jiang et al., 2020a, b, Lee et al., 2021, 2020] has shown that models often struggle with long-term predictions due to compounding uncertainty at each lookahead step. This is because the model’s predictions at each step are based on the uncertain predictions from the previous steps. As a result, querying an extended horizon can hinder the optimization process by including suboptimal or misleading inputs in the non-myopic horizon, potentially guiding the search in less promising directions and incurring higher computational costs. To address this issue, we adopt the same strategy from previous work and introduce a maximum horizon length to limit the size of the horizon.

The NMMO-Nested method stands out with its potential for superior optimization performance due to its comprehensive approach to handling nested optimizations. However, this method has practical challenges. Notably, this method requires a pre-defined grid for optimization, as internal maximization prevents using gradient-based optimizers. Furthermore, Its computational complexity constrains the size of the grid that can be feasibly used, which in turn limits the method’s operational speed and accuracy. Hence, the nested approach does not perform as well as expected compared to the other proposed non-myopic approaches. We provide results including NMMO-Joint with lookahead horizon H{4,8}H\in\{4,8\} and BINOM with HH = 4. We report the results for NMMO-Nested with H=2H=2 in the Appendix Section 7.5.

We consider two state-of-the-art MOBO baselines to benchmark the performance of our non-myopic methods. First, we use the multi-objective variant of the joint entropy search for multi-objective Bayesian optimization (JESMO) Tu et al. [2022], Balandat et al. [2020]. Second, we use the expected hypervolume improvement (EHVI) method Emmerich et al. [2006], Balandat et al. [2020] as a hypervolume-based baseline. Another baseline is the hypervolume knowledge gradient (HVKG), a one-step lookahead method for cost-aware MOBO problems Daulton et al. [2023]. While the HVKG algorithm is designed for cost-aware settings with decoupled evaluations, we use a modified version of this method where the objective function costs are set to zero, and we calculate all objectives for a true evaluation of the hypervolume. We report the results in the Appendix Section 7.5.

The baseline methods were chosen due to their distinct but complementary approaches to handling multi-objective trade-offs, providing a comprehensive comparative perspective against our proposed NMMO method. The effectiveness of each method is measured using the hypervolume indicator.

Evaluation metric. The HVI is a widely-used metric in MOO that measures the volume of the space dominated by the Pareto front relative to a predefined reference point: a larger hypervolume indicates better Pareto solutions. We will compare the hypervolume metric after each BO iteration.

Experimental details: All experiments were averaged among 15 runs, initialized with five points randomly generated from Sobol sequences, and run for 65 iterations. We emphasize that non-myopic solutions are well-suited for the smaller experimental budgets. We model each of the multiple objectives using an independent GP with a Matérn 5/2 ARD kernel. We use implementations from the BoTorch Python package Balandat et al. [2020] for all baselines, including JESMO, EHVI, and HVKG. All methods were evaluated under identical initial conditions. This includes consistent initialization, identical computational budgets, and the same evaluation criteria across all experiments.

5.1 Results and Discussion

Figure 2 shows the hypervolume versus BO iterations for NMMO-Joint with lookahead horizon H{4,8}H\in\{4,8\} and BINOM with lookahead horizon HH = 4 against existing state-of-the-art baselines.

Non-myopic methods perform better than myopic ones. The non-myopic methods outperform the state-of-the-art myopic MOBO algorithms on all six real-world MOO problems (Figure 2). These results highlight the capability of non-myopic methods in maximizing the hypervolume when the experimental budget is small (i.e., 65 BO iterations).

Non-myopic methods with different horizons. From our results, we often observe a slight decrease in performance with increasing horizon values in non-myopic MOO. This trend can be primarily attributed to the increased complexity and model uncertainty as the decision horizon extends. A longer horizon necessitates the consideration of a larger set of future outcomes and their interactions, which not only complicates the optimization task but also introduces greater uncertainty and potential for errors in prediction. Consequently, these accumulated errors can degrade the optimization strategy, making it challenging for the algorithm to effectively balance short-term gains against long-term goals, and negatively impacting performance. This finding highlights the challenges in implementing long-term strategic planning in complex optimization environments. These challenges can be seen in the NMMO-Joint and BINOM methods run on each benchmark with increasing lookahead horizon H{2,4,6,8}H\in\{2,4,6,8\} (Appendix Figure 3 and Figure 4).

Computational runtime comparison. In our comparative analysis of myopic and non-myopic MOO methods, there are computational complexity and optimization performance trade-offs. Myopic MOO methods, due to their focus on immediate gains, have lower computational complexity, which translates into faster runtimes (Table LABEL:table:runtimes). However, this computational efficiency comes at the cost of performance. Our results indicate that, while myopic methods are faster, they underperform in comparison to their non-myopic counterparts in limited-budget settings. Hence, while non-myopic approaches require more computational resources, their better optimization results justify the additional complexity.

Ablation study of information-gain within BINOM. In addition to the EHVI acquisition function, we incorporated information-gain acquisition functions, specifically MESMO and JESMO, into our proposed BINOM approach. Even though information gain based scalarization does not always satisfy the monotonicity condition as discussed in section 7.3, These instantiations of our framework provide a complementary approach to the hypervolume-based strategies. The results, provided in the Appendix demonstrate that the non-myopic methods maintain strong performance across different instantiations of BINOM, highlighting the versatility and effectiveness of our framework. The results also show that the HVI instantiation performs better in most cases.

6 Summary

This paper introduced a novel set of non-myopic methods for multi-objective Bayesian optimization (MOBO), to specifically address the problem of finite-horizon sequential experimental design (SED) for MOO problems. By addressing the difficulties related to scalarization, monotonicity, and additivity of the reward function typically encountered in MOO scenarios, we successfully adapted the Bellman optimality principle for MOBO by adopting a new approach to scalarization using hypervolume improvement (HVI). Our proposed methods, including the Nested, Joint, and BINOM acquisition functions, represent the first non-myopic acquisition functions tailored for MOBO problems. Our experiments on multiple real-world problems, show that our methods show significant improvements by consistently outperforming the state-of-the-art myopic approaches in limited-budget settings.

References

  • Ahmadianshalchi et al. [2024a] A. Ahmadianshalchi, S. Belakaria, and J. R. Doppa. Pareto front-diverse batch multi-objective Bayesian optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 38:10784–10794, Mar. 2024a. doi: 10.1609/aaai.v38i10.28951. URL https://ojs.aaai.org/index.php/AAAI/article/view/28951.
  • Ahmadianshalchi et al. [2024b] A. Ahmadianshalchi, S. Belakaria, and J. R. Doppa. Preference-aware constrained multi-objective bayesian optimization. In Proceedings of the 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD), pages 182–191, 2024b.
  • Ahmed and Siegel [2021] A. Ahmed and D. J. Siegel. Predicting hydrogen storage in mofs via machine learning. Patterns, 2(7):100291, 2021. ISSN 2666-3899.
  • Ahmed et al. [2017] A. Ahmed, Y. Liu, J. Purewal, L. D. Tran, A. G. Wong-Foy, M. Veenstra, A. J. Matzger, and D. J. Siegel. Balancing gravimetric and volumetric hydrogen density in mofs. Energy Environ. Sci., 10:2459–2471, 2017.
  • Ament et al. [2024] S. Ament, S. Daulton, D. Eriksson, M. Balandat, and E. Bakshy. Unexpected improvements to expected improvement for Bayesian optimization. Advances in Neural Information Processing Systems, 36, 2024.
  • Amir and Hasegawa [1989] H. M. Amir and T. Hasegawa. Nonlinear mixed-discrete structural optimization. Journal of Structural Engineering, 115(3):626–646, 1989.
  • Balandat et al. [2020] M. Balandat, B. Karrer, D. R. Jiang, S. Daulton, B. Letham, A. G. Wilson, and E. Bakshy. BoTorch: A Framework for Efficient Monte-Carlo Bayesian Optimization. In Advances in Neural Information Processing Systems 33, 2020. URL http://arxiv.org/abs/1910.06403.
  • Barrett and Narayanan [2008] L. Barrett and S. Narayanan. Learning all optimal policies with multiple criteria. In Proceedings of the 25th international conference on Machine learning, pages 41–47, 2008.
  • Belakaria et al. [2019] S. Belakaria, A. Deshwal, and J. R. Doppa. Max-value entropy search for multi-objective Bayesian optimization. In Conference on Neural Information Processing Systems, 2019.
  • Belakaria et al. [2020a] S. Belakaria, A. Deshwal, and J. R. Doppa. Max-value entropy search for multi-objective Bayesian optimization with constraints. arXiv preprint arXiv:2009.01721, 2020a.
  • Belakaria et al. [2020b] S. Belakaria, A. Deshwal, N. K. Jayakodi, and J. R. Doppa. Uncertainty-aware search framework for multi-objective Bayesian optimization. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, 2020b.
  • Belakaria et al. [2021] S. Belakaria, A. Deshwal, and J. R. Doppa. Output space entropy search framework for multi-objective bayesian optimization. Journal of artificial intelligence research, 72:667–715, 2021.
  • Boutilier et al. [1999] C. Boutilier, T. Dean, and S. Hanks. Decision-theoretic planning: Structural assumptions and computational leverage. Journal of Artificial Intelligence Research, 11:1–94, 1999.
  • Boyd et al. [2019] P. G. Boyd, A. Chidambaram, E. García-Díez, C. P. Ireland, T. D. Daff, R. Bounds, A. Gładysiak, P. Schouwink, S. M. Moosavi, M. M. Maroto-Valer, J. A. Reimer, J. A. R. Navarro, T. K. Woo, S. Garcia, K. C. Stylianou, and B. Smit. Data-driven design of metal–organic frameworks for wet flue gas co2 capture. Nature, 576(7786):253–256, Dec. 2019. ISSN 1476-4687. doi: 10.1038/s41586-019-1798-7. URL http://dx.doi.org/10.1038/s41586-019-1798-7.
  • Cheng and Li [1999] F. Cheng and X. Li. Generalized center method for multiobjective engineering optimization. Engineering Optimization, 31(5):641–661, 1999.
  • Coello and Montes [2002] C. A. C. Coello and E. M. Montes. Constraint-handling in genetic algorithms through the use of dominance-based tournament selection. Advanced Engineering Informatics, 16(3):193–203, 2002.
  • Daulton et al. [2020] S. Daulton, M. Balandat, and E. Bakshy. Differentiable expected hypervolume improvement for parallel multi-objective Bayesian optimization. Advances in Neural Information Processing Systems, 33:9851–9864, 2020.
  • Daulton et al. [2021] S. Daulton, M. Balandat, and E. Bakshy. Parallel Bayesian optimization of multiple noisy objectives with expected hypervolume improvement. NeurIPS, 34, 2021.
  • Daulton et al. [2022] S. Daulton, S. Cakmak, M. Balandat, M. A. Osborne, E. Zhou, and E. Bakshy. Robust multi-objective Bayesian optimization under input noise. In K. Chaudhuri, S. Jegelka, L. Song, C. Szepesvari, G. Niu, and S. Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 4831–4866. PMLR, 17–23 Jul 2022. URL https://proceedings.mlr.press/v162/daulton22a.html.
  • Daulton et al. [2023] S. Daulton, M. Balandat, and E. Bakshy. Hypervolume knowledge gradient: a lookahead approach for multi-objective Bayesian optimization with partial information. In International Conference on Machine Learning, pages 7167–7204. PMLR, 2023.
  • Deb and Srinivasan [2006] K. Deb and A. Srinivasan. Innovization: Innovating design principles through optimization. In Proceedings of the 8th annual conference on Genetic and evolutionary computation, pages 1629–1636, 2006.
  • Emmerich and Klinkenberg [2008] M. Emmerich and J.-w. Klinkenberg. The computation of the expected improvement in dominated hypervolume of pareto front approximations. Technical Report, Leiden University, 34, 2008.
  • Emmerich et al. [2006] M. Emmerich, K. Giannakoglou, and B. Naujoks. Single- and multiobjective evolutionary optimization assisted by Gaussian random field metamodels. IEEE Transactions on Evolutionary Computation, 10(4):421–439, 2006. doi: 10.1109/TEVC.2005.859463.
  • Garnett [2023] R. Garnett. Bayesian optimization. Cambridge University Press, 2023.
  • Garrido-Merchán and Hernández-Lobato [2019] E. C. Garrido-Merchán and D. Hernández-Lobato. Predictive entropy search for multi-objective Bayesian optimization with constraints. Neurocomputing, 361:50–68, 2019.
  • Garrido-Merchán and Hernández-Lobato [2021] E. C. Garrido-Merchán and D. Hernández-Lobato. Parallel predictive entropy search for multi-objective Bayesian optimization with constraints, 2021.
  • Ginsbourger and Le Riche [2009] D. Ginsbourger and R. Le Riche. Towards gp-based optimization with finite time horizon. Advances in Model-Oriented Design and Analysis: Proceedings of the 9th International Workshop in Model-Oriented Design and Analysis, 10 2009.
  • González et al. [2016] J. González, M. Osborne, and N. Lawrence. Glasses: Relieving the myopia of Bayesian optimisation. In Artificial Intelligence and Statistics, pages 790–799. PMLR, 2016.
  • Hernández-Lobato et al. [2016] D. Hernández-Lobato, J. Hernandez-Lobato, A. Shah, and R. Adams. Predictive entropy search for multi-objective Bayesian optimization. In Proceedings of International Conference on Machine Learning (ICML), pages 1492–1501, 2016.
  • Hernández-Lobato et al. [2014] J. M. Hernández-Lobato, M. W. Hoffman, and Z. Ghahramani. Predictive entropy search for efficient global optimization of black-box functions. In NeurIPS, 2014.
  • Jiang et al. [2020a] S. Jiang, H. Chai, J. Gonzalez, and R. Garnett. BINOCULARS for efficient, nonmyopic sequential experimental design. In International Conference on Machine Learning, pages 4794–4803. PMLR, 2020a.
  • Jiang et al. [2020b] S. Jiang, D. Jiang, M. Balandat, B. Karrer, J. Gardner, and R. Garnett. Efficient nonmyopic Bayesian optimization via one-shot multi-step trees. In NeurIPS, 2020b.
  • Kitagawa et al. [2014] S. Kitagawa et al. Metal–organic frameworks (mofs). Chemical Society Reviews, 43(16):5415–5418, 2014.
  • Knowles [2006] J. Knowles. Parego: A hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems. IEEE Transactions on Evolutionary Computation, 10(1):50–66, 2006.
  • Konakovic Lukovic et al. [2020] M. Konakovic Lukovic, Y. Tian, and W. Matusik. Diversity-guided multi-objective Bayesian optimization with batch evaluations. Advances in Neural Information Processing Systems, 33:17708–17720, 2020.
  • Krause and Guestrin [2007] A. Krause and C. Guestrin. Nonmyopic active learning of Gaussian processes: an exploration-exploitation approach. In Proceedings of the 24th international conference on Machine learning, pages 449–456, 2007.
  • Lam and Willcox [2016] R. Lam and K. E. Willcox. Lookahead Bayesian optimization with inequality constraints. Proceedings of the 34th International Conference on Machine Learning, 56:1230–1239, 2016.
  • Lam et al. [2016] R. Lam, K. Willcox, and D. H. Wolpert. Bayesian optimization with a finite budget: An approximate dynamic programming approach. Advances in Neural Information Processing Systems, 29, 2016.
  • Lee et al. [2020] E. Lee, D. Eriksson, D. Bindel, B. Cheng, and M. Mccourt. Efficient rollout strategies for Bayesian optimization. In UAI. PMLR, 2020.
  • Lee et al. [2021] E. H. Lee, D. Eriksson, V. Perrone, and M. Seeger. A nonmyopic approach to cost-constrained Bayesian optimization, 2021.
  • Lu and Zhao [2004] G. M. Lu and X. S. Zhao. Nanoporous materials: science and engineering, volume 4. World Scientific, 2004.
  • Lyu et al. [2021] X. Lyu, M. Binois, and M. Ludkovski. Evaluating Gaussian process metamodels and sequential designs for noisy level set estimation. Statistics and Computing, 31(4):43, 2021.
  • Mockus [1989] J. Mockus. Bayesian Approach to Global Optimization: Theory and Applications. Kluwer Academic Publishers, 1989.
  • Osborne and Osborne [2010] M. Osborne and M. A. Osborne. Bayesian Gaussian processes for sequential prediction, optimisation and quadrature. PhD thesis, Oxford University, UK, 2010.
  • Osborne et al. [2009] M. A. Osborne, R. Garnett, and S. J. Roberts. Gaussian processes for global optimization. In 3rd international conference on learning and intelligent optimization (LION3), pages 1–15, 2009.
  • Rao [2019] S. S. Rao. Engineering optimization: theory and practice. John Wiley & Sons, 2019.
  • Ray and Liew [2002] T. Ray and K. Liew. A swarm metaphor for multiobjective design optimization. Engineering optimization, 34(2):141–153, 2002.
  • Roijers et al. [2013] D. M. Roijers, P. Vamplew, S. Whiteson, and R. Dazeley. A survey of multi-objective sequential decision-making. Journal of Artificial Intelligence Research, 48:67–113, 2013.
  • Shahriari et al. [2015] B. Shahriari, K. Swersky, Z. Wang, R. P. Adams, and N. De Freitas. Taking the human out of the loop: A review of Bayesian optimization. Proceedings of the IEEE, 2015.
  • Suzuki et al. [2020] S. Suzuki, S. Takeno, T. Tamura, K. Shitara, and M. Karasuyama. Multi-objective Bayesian optimization using pareto-frontier entropy. In International Conference on Machine Learning, pages 9279–9288. PMLR, 2020.
  • Tanabe and Ishibuchi [2020] R. Tanabe and H. Ishibuchi. An easy-to-use real-world multi-objective optimization problem suite. Applied Soft Computing, 89:106078, 2020.
  • Tu et al. [2022] B. Tu, A. Gandy, N. Kantas, and B. Shafei. Joint entropy search for multi-objective Bayesian optimization. Advances in Neural Information Processing Systems, 35:9922–9938, 2022.
  • Van Moffaert and Nowé [2014] K. Van Moffaert and A. Nowé. Multi-objective reinforcement learning using sets of pareto dominating policies. The Journal of Machine Learning Research, 15(1):3483–3512, 2014.
  • Van Moffaert et al. [2013] K. Van Moffaert, M. M. Drugan, and A. Nowé. Hypervolume-based multi-objective reinforcement learning. In Evolutionary Multi-Criterion Optimization: 7th International Conference, EMO 2013, Sheffield, UK, March 19-22, 2013. Proceedings 7, pages 352–366. Springer, 2013.
  • Wang and Jegelka [2017] Z. Wang and S. Jegelka. Max-value entropy search for efficient Bayesian optimization. In Proceedings of International Conference on Machine Learning (ICML), 2017.
  • Williams and Rasmussen [2006] C. K. Williams and C. E. Rasmussen. Gaussian processes for machine learning. MIT press Cambridge, MA, 2006.
  • Wu and Frazier [2019] J. Wu and P. Frazier. Practical two-step lookahead Bayesian optimization. Advances in neural information processing systems, 32, 2019.
  • Zitzler and Thiele [1999] E. Zitzler and L. Thiele. Multiobjective evolutionary algorithms: a comparative case study and the strength pareto approach. IEEE transactions on Evolutionary Computation, 3(4):257–271, 1999.
  • Zitzler et al. [2000] E. Zitzler, K. Deb, and L. Thiele. Comparison of multiobjective evolutionary algorithms: Empirical results. Evolutionary computation, 8(2):173–195, 2000.

7 Appendix

7.1 Benchmarks details

In Table LABEL:tab:benchmark_details, we include a concise overview of all of our MOO benchmarks consisting of the problem names, number of input dimensions (dd), number of output dimensions (KK), and the reference point (rr) used for calculating the hypervolume indicator for each benchmark.

7.2 Analysis of lookahead horizons

We include figures illustrating the performance of our non-myopic methods across different horizon values. The results reveal a general trend where larger horizons correlate with slight drops in performance. This observation can primarily be attributed to the increased uncertainty associated with extending the optimization horizon. As the horizon increases, the predictive model must account for a broader range of potential future outcomes. Their uncertainties can complicate the decision-making process, as the model faces increased difficulty in accurately estimating the long-term consequences of current decisions. Consequently, while longer horizons theoretically allow for more strategic and informed decision-making by considering future impacts, the added uncertainty can undermine performance by introducing greater variability in the optimization outcomes. Figure 3 and Figure 4 show the hypervolume results for NMMO-Joint and BINOM with horizon lengths 2, 4, 6, and 8.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 3: Hypervolume results for NMMO-Joint with lookahead horizon H{2,4,6,8}H\in\{2,4,6,8\}.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 4: Hypervolume results for BINOM with lookahead horizon H{2,4,6,8}H\in\{2,4,6,8\}.

7.3 Acquisition function ablation study

Here, we present the results of our ablation study comparing the performance of the BINOM acquisition function using Expected Hypervolume Improvement (EHVI), Max-value Entropy Search (MES), and Joint Entropy Search (JES) with horizon lengths of 2, 4, and 6, across six real-world benchmark problems. For clarity, we refer to the BINOM variant using EHVI as simply BINOM. The hypervolume plots for each benchmark illustrate that BINOM-EHVI generally outperforms the other acquisition functions, demonstrating its superiority in guiding the optimization process towards a high-quality Pareto front. Even in instances where BINOM-EHVI does not achieve the best performance, it consistently surpasses our myopic baseline methods, reaffirming the robustness and effectiveness of the BINOM approach in diverse optimization scenarios.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 5: Hypervolume results for BINOM with MESMO (BINOM-MES) , JESMO (BINOM-JES), and EHVI (BINOM) acquisition functions with lookahead horizon H{2,4,6}H\in\{2,4,6\}.

7.4 ZDT3 - Synthetic Benchmark

In addition to the benchmarks presented in the main paper, we evaluated our method on the ZDT3 benchmark in Figure 6. The ZDT3 problem is a widely used test in multi-objective optimization literature Konakovic Lukovic et al. [2020].

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 6: With respect to the ZDT3 problem, Figure 6(a) shows the comparison of all baselines, Figure 6(b) shows the hypervolume comparison for BINOM with varying horizon lengths, and Figure 6(c) depicts the hypervolume comparison for NMMO-joint with different horizon lengths.

7.5 Additional Baselines

HVKG Daulton et al. [2023]. HVKG is notable for its cost-aware decoupled strategy, which assigns specific costs to each objective function and avoids evaluating all objectives in every iteration. To ensure a fair comparison between HVKG and our NMMO approaches, we adjusted the objective function costs to zero for all problems in our experiments, which allows all methods to operate under similar cost conditions. Moreover, to accurately measure the hypervolume indicator, we implemented a procedure where all objective functions of the selected inputs under the HVKG method are evaluated at each iteration. These expensive evaluations are then used to calculate the hypervolume, ensuring that all methods’ performances are assessed similarly.

NMMO-Nested. We include the NMMO-Nested with H=2H=2 results in Figure 7. Due to computational cost, we were not able to run experiments of NMMO-Nested with additional horizon values. As explained in section 5, while the NMMO-Nested method should theoretically outperform the other non-myopic baselines, its optimization is very time-consuming and inefficient in practice, leading to slightly subpar results.

MESMO. We included MESMO (Max-value Entropy Search for Multi-objective Optimization Belakaria et al. [2019]) in our baseline comparisons to assess its performance on various benchmarks. MESMO employs an advanced output space entropy search strategy, aiming to maximize the information gained about the Pareto front with each iteration. Despite its innovative approach, MESMO has been somewhat overshadowed in recent evaluations by newer methods like JESMO, which combines the strengths of both output and input space entropy searches. We include a comparison of how MESMO’s performance stacks up against other state-of-the-art multi-objective Bayesian optimization techniques.

We use implementations from the BoTorch python package ***https://github.com/pytorch/botorch for all baselines (JESMO, EHVI, MESMO, HVKG).

7.6 Run times

The following are the computational complexities of our proposed methods.

BINOM: O(KN2H)O(KN^{2}H) Where K is the number of objectives, N is the number of observations, and H is the lookahead horizon.

NMMO-Joint: O(KN2H2)O(KN^{2}H^{2}) This method has an additional factor of H due to the joint optimization over the lookahead horizon.

NMMO-Nested: O(KN2H2M)O(KN^{2}H^{2}M) Where M is the size of the discretized input space used for the nested optimization.

For context, the baseline EHVI method typically has a complexity of O(KN2)O(KN^{2}).

While our non-myopic methods do introduce additional computational overhead compared to myopic approaches, the increase is not prohibitively high, especially considering the performance gains achieved and its intended use in budget-limited settings for scientific applications.

Table LABEL:table:runtimes, shows the average run-time per iteration of each baseline on multiple benchmarks. We utilized an NVIDIA Quadro RTX 6000 GPU with 24,576 MiB memory capacity. We provide the average and standard deviation of 65 iterations in one run of the algorithm (in seconds). It is noteworthy that while the non-myopic methods take longer, they also provide better performance with a limited budget.

Table 1: Runtime comparison (seconds per iteration) rounded to the closest integer.
Method H Reinforced concrete beam design Four-bar truss design Gear train design Metal-organic framework design
EHVI - 1±01\pm 0 0±00\pm 0 1±01\pm 0 5±05\pm 0
JESMO - 2±22\pm 2 4±34\pm 3 3±33\pm 3 6±06\pm 0
HVKG 1 47±2047\pm 20 32±532\pm 5 52±1452\pm 14 102±27102\pm 27
BINOM 2 4±24\pm 2 9±39\pm 3 13±313\pm 3 17±217\pm 2
4 7±47\pm 4 22±922\pm 9 24±324\pm 3 17±117\pm 1
NMMO-Joint 2 28±1328\pm 13 22±522\pm 5 59±1059\pm 10 47±747\pm 7
4 52±1952\pm 19 35±1535\pm 15 75±1075\pm 10 55±455\pm 4
8 104±33104\pm 33 125±41125\pm 41 181±52181\pm 52 73±1373\pm 13
NMMO-Nested 2 560±58560\pm 58 802±62802\pm 62 616±151616\pm 151 754±141754\pm 141
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 7: Hypervolume results on real-world benchmarks including additional baselines

7.7 Additional Experimental Details

Joint Entropy Search for Multi-Objective Optimization (JESMO): We use the JESMO Tu et al. [2022] algorithm as implemented in the BoTorch python package Balandat et al. [2020] as a baseline for comparison in our study. However, we encountered a significant challenge with this approach: the JESMO implementation in BoTorch failed almost immediately when applied to many of our real-world multi-objective problems. This failure was primarily due to the algorithm’s inability to identify a sufficient number of optimal Pareto sample points. Specifically, it struggled to find the required 10 sample points. To accommodate this limitation and ensure the feasibility of using JESMO in our experiments, we adjusted the requirement to a smaller set of 4 optimal Pareto sample points. We determined the optimal number of Pareto sample points by monitoring the instances where JESMO failed to identify a sufficient number of optimal points across various runs. This empirical approach allowed us to adjust the number of required sample points to a level where the algorithm could consistently perform without failure.

7.8 Limitations and Future Work.

Despite the proposed advancements in solving the finite-horizon multi-objective SED problem using hypervolume improvement strategies, several limitations remain in our approach, opening the way for future improvements. While the introduction of alternative lower bounds and the use of BEHVI are designed to provide a tractable alternative to solving the Bellman equation, these strategies raise questions about the tightness of the lower bounds and its ability to accurately represent the intractable optimal policy. This question is closely related to prior work on the adaptivity gap Krause and Guestrin [2007]. Given that we established the possibility of using the Bellman equation with HVI, a promising future direction would be to build a multi-step expected HVI approach analogous to the multi-step expected improvement Jiang et al. [2020b].

Table 2: Benchmark details.
Problem name d K r
Disc Brake Design 4 3 (6.1356, 6.3421, 12.9737)
Four Bar Truss Design 4 2 (2967.0243, 0.0383)
Gear Train Design 4 3 (6.6764, 59.0, 0.4633)
Reinforced Concrete Beam Design 3 2 (703.6860, 899.2291)
Welded Beam Design 4 3 (202.8569, 42.0653, 2111643.6209)
Metal-Organic Framework Design 7 2 (1.0, 1.0)
ZDT-3 9 2 (11.0, 11.0)

8 Hypervolume Improvement Additivity Proof

Let 𝒴0\mathcal{Y}_{0} be the initial Pareto front, and let (𝐱t,𝐲t)(\mathbf{x}_{t},\mathbf{y}_{t}) be the new data point added at time step tt, where t[1,T]t\in[1,T]. We denote the updated Pareto front at time step tt as 𝒴t\mathcal{Y}_{t}.

The hypervolume improvement at time step tt is given by:

HVIt=HVI(𝐲𝐭|𝒴t1)\displaystyle HVI_{t}=HVI(\mathbf{y_{t}}|\mathcal{Y}_{t-1}) =HV(𝒴t1𝐲t)HV(𝒴t1)\displaystyle=HV(\mathcal{Y}_{t-1}\cup\mathbf{y}_{t})-HV(\mathcal{Y}_{t-1}) =HV(𝒴t)HV(𝒴t1)\displaystyle=HV(\mathcal{Y}_{t})-HV(\mathcal{Y}_{t-1}) (16)

Now, let’s consider the final hypervolume improvement after TT time steps, we denote the finaly hypervolumen improvemennt as HVITotalHVI_{Total}

HVITotal=HV(𝒴T)HV(𝒴0)HVI_{Total}=HV(\mathcal{Y}_{T})-HV(\mathcal{Y}_{0})

We want to prove that:

HVITotal=t=1t=THVItHVI_{Total}=\sum_{t=1}^{t=T}HVI_{t}

By expanding the summation, we get a telescoping series:

t=1t=THVIt\displaystyle\sum_{t=1}^{t=T}HVI_{t} =t=1t=THV(𝒴t)HV(𝒴t1)\displaystyle=\sum_{t=1}^{t=T}HV(\mathcal{Y}_{t})-HV(\mathcal{Y}_{t-1}) (17)
=(HV(𝒴1)HV(𝒴0))+(HV(𝒴2)HV(𝒴1))++(HV(𝒴T)HV(𝒴T1))\displaystyle=(HV(\mathcal{Y}_{1})-HV(\mathcal{Y}_{0}))+(HV(\mathcal{Y}_{2})-HV(\mathcal{Y}_{1}))+\dots+(HV(\mathcal{Y}_{T})-HV(\mathcal{Y}_{T-1})) (18)

The intermediate terms of the series canceling out lead to:

t=1t=THVIt\displaystyle\sum_{t=1}^{t=T}HVI_{t} =HV(𝒴T)HV(𝒴0)\displaystyle=HV(\mathcal{Y}_{T})-HV(\mathcal{Y}_{0}) (19)

Therefore,

HVITotal=t=1t=THVIt\displaystyle HVI_{Total}=\sum_{t=1}^{t=T}HVI_{t} (20)

This proves that the final hypervolume improvement is indeed the summation of hypervolume improvements occurring at individual time steps.

The intuition behind this proof is that the hypervolume improvement at each time step represents the additional volume dominated by the updated Pareto front compared to the previous Pareto front. By summing up these individual improvements, we account for the total volume gained from the initial Pareto front to the final Pareto front.