Source code for botorch.acquisition.multi_objective.parego

# Copyright (c) Meta Platforms, Inc. and affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.

from typing import Callable, List, Optional, Union

import torch
from botorch.acquisition.logei import qLogNoisyExpectedImprovement, TAU_MAX, TAU_RELU
from botorch.acquisition.multi_objective.monte_carlo import (
    MultiObjectiveMCAcquisitionFunction,
)
from botorch.acquisition.multi_objective.objective import MCMultiOutputObjective
from botorch.acquisition.objective import GenericMCObjective
from botorch.models.model import Model
from botorch.posteriors.fully_bayesian import MCMC_DIM
from botorch.sampling.base import MCSampler
from botorch.utils.multi_objective.scalarization import get_chebyshev_scalarization
from botorch.utils.sampling import sample_simplex
from botorch.utils.transforms import is_ensemble
from torch import Tensor


[docs] class qLogNParEGO(qLogNoisyExpectedImprovement, MultiObjectiveMCAcquisitionFunction): def __init__( self, model: Model, X_baseline: Tensor, scalarization_weights: Optional[Tensor] = None, sampler: Optional[MCSampler] = None, objective: Optional[MCMultiOutputObjective] = None, constraints: Optional[List[Callable[[Tensor], Tensor]]] = None, X_pending: Optional[Tensor] = None, eta: Union[Tensor, float] = 1e-3, fat: bool = True, prune_baseline: bool = False, cache_root: bool = True, tau_relu: float = TAU_RELU, tau_max: float = TAU_MAX, ) -> None: r"""q-LogNParEGO supporting m >= 2 outcomes. This acquisition function utilizes qLogNEI to compute the expected improvement over Chebyshev scalarization of the objectives. This is adapted from qNParEGO proposed in [Daulton2020qehvi]_ to utilize log-improvement acquisition functions of [Ament2023logei]_. See [Knowles2005]_ for the original ParEGO algorithm. This implementation assumes maximization of all objectives. If any of the model outputs are to be minimized, either an `objective` should be used to negate the model outputs or the `scalarization_weights` should be provided with negative weights for the outputs to be minimized. Args: model: A fitted multi-output model, producing outputs for `m` objectives and any number of outcome constraints. NOTE: The model posterior must have a `mean` attribute. X_baseline: A `batch_shape x r x d`-dim Tensor of `r` design points that have already been observed. These points are considered as the potential best design point. scalarization_weights: A `m`-dim Tensor of weights to be used in the Chebyshev scalarization. If omitted, samples from the unit simplex. sampler: The sampler used to draw base samples. See `MCAcquisitionFunction` more details. objective: The MultiOutputMCAcquisitionObjective under which the samples are evaluated before applying Chebyshev scalarization. Defaults to `IdentityMultiOutputObjective()`. constraints: A list of constraint callables which map a Tensor of posterior samples of dimension `sample_shape x batch-shape x q x m'`-dim to a `sample_shape x batch-shape x q`-dim Tensor. The associated constraints are satisfied if `constraint(samples) < 0`. X_pending: A `batch_shape x q' x d`-dim Tensor of `q'` design points that have points that have been submitted for function evaluation but have not yet been evaluated. Concatenated into `X` upon forward call. Copied and set to have no gradient. eta: Temperature parameter(s) governing the smoothness of the sigmoid approximation to the constraint indicators. See the docs of `compute_(log_)smoothed_constraint_indicator` for details. fat: Toggles the logarithmic / linear asymptotic behavior of the smooth approximation to the ReLU. prune_baseline: If True, remove points in `X_baseline` that are highly unlikely to be the best point. This can significantly improve performance and is generally recommended. In order to customize pruning parameters, instead manually call `botorch.acquisition.utils.prune_inferior_points` on `X_baseline` before instantiating the acquisition function. cache_root: A boolean indicating whether to cache the root decomposition over `X_baseline` and use low-rank updates. tau_max: Temperature parameter controlling the sharpness of the smooth approximations to max. tau_relu: Temperature parameter controlling the sharpness of the smooth approximations to ReLU. """ MultiObjectiveMCAcquisitionFunction.__init__( self, model=model, sampler=sampler, objective=objective, constraints=constraints, eta=eta, ) org_objective = self.objective # Create the composite objective. with torch.no_grad(): Y_baseline = org_objective(model.posterior(X_baseline).mean) if is_ensemble(model): Y_baseline = torch.mean(Y_baseline, dim=MCMC_DIM) scalarization_weights = ( scalarization_weights if scalarization_weights is not None else sample_simplex( d=Y_baseline.shape[-1], device=X_baseline.device, dtype=X_baseline.dtype ).view(-1) ) chebyshev_scalarization = get_chebyshev_scalarization( weights=scalarization_weights, Y=Y_baseline, ) composite_objective = GenericMCObjective( objective=lambda samples, X=None: chebyshev_scalarization( org_objective(samples=samples, X=X), X=X ), ) qLogNoisyExpectedImprovement.__init__( self, model=model, X_baseline=X_baseline, sampler=sampler, # This overwrites self.objective with the composite objective. objective=composite_objective, X_pending=X_pending, constraints=constraints, eta=eta, fat=fat, prune_baseline=prune_baseline, cache_root=cache_root, tau_max=tau_max, tau_relu=tau_relu, ) # Set these after __init__ calls so that they're not overwritten / deleted. # These are intended mainly for easier debugging & transparency. self._org_objective: MCMultiOutputObjective = org_objective self.chebyshev_scalarization: Callable[[Tensor, Optional[Tensor]], Tensor] = ( chebyshev_scalarization ) self.scalarization_weights: Tensor = scalarization_weights self.Y_baseline: Tensor = Y_baseline