Source code for botorch.acquisition.acquisition
#!/usr/bin/env python3
# Copyright (c) Meta Platforms, Inc. and affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
r"""Abstract base module for all botorch acquisition functions."""
from __future__ import annotations
import warnings
from abc import ABC, abstractmethod
from typing import Optional
import torch
from botorch.exceptions import BotorchWarning
from botorch.models.model import Model, ModelDict
from botorch.posteriors.posterior import Posterior
from botorch.sampling.base import MCSampler
from botorch.sampling.get_sampler import get_sampler
from torch import Tensor
from torch.nn import Module
[docs]
class AcquisitionFunction(Module, ABC):
r"""Abstract base class for acquisition functions.
Please note that if your acquisition requires a backwards call,
you will need to wrap the backwards call inside of an enable_grad
context to be able to optimize the acquisition. See #1164.
"""
_log: bool = False # whether the acquisition utilities are in log-space
def __init__(self, model: Model) -> None:
r"""Constructor for the AcquisitionFunction base class.
Args:
model: A fitted model.
"""
super().__init__()
self.model: Model = model
[docs]
def set_X_pending(self, X_pending: Optional[Tensor] = None) -> None:
r"""Informs the acquisition function about pending design points.
Args:
X_pending: `n x d` Tensor with `n` `d`-dim design points that have
been submitted for evaluation but have not yet been evaluated.
"""
if X_pending is not None:
if X_pending.requires_grad:
warnings.warn(
"Pending points require a gradient but the acquisition function"
" will not provide a gradient to these points.",
BotorchWarning,
)
self.X_pending = X_pending.detach().clone()
else:
self.X_pending = X_pending
[docs]
@abstractmethod
def forward(self, X: Tensor) -> Tensor:
r"""Evaluate the acquisition function on the candidate set X.
Args:
X: A `(b) x q x d`-dim Tensor of `(b)` t-batches with `q` `d`-dim
design points each.
Returns:
A `(b)`-dim Tensor of acquisition function values at the given
design points `X`.
"""
pass # pragma: no cover
[docs]
class OneShotAcquisitionFunction(AcquisitionFunction, ABC):
r"""
Abstract base class for acquisition functions using one-shot optimization
"""
[docs]
@abstractmethod
def get_augmented_q_batch_size(self, q: int) -> int:
r"""Get augmented q batch size for one-shot optimization.
Args:
q: The number of candidates to consider jointly.
Returns:
The augmented size for one-shot optimization (including variables
parameterizing the fantasy solutions).
"""
pass # pragma: no cover
[docs]
@abstractmethod
def extract_candidates(self, X_full: Tensor) -> Tensor:
r"""Extract the candidates from a full "one-shot" parameterization.
Args:
X_full: A `b x q_aug x d`-dim Tensor with `b` t-batches of `q_aug`
design points each.
Returns:
A `b x q x d`-dim Tensor with `b` t-batches of `q` design points each.
"""
pass # pragma: no cover
[docs]
class MCSamplerMixin(ABC):
r"""A mix-in for adding sampler functionality into an acquisition function class.
Attributes:
_default_sample_shape: The `sample_shape` for the default sampler.
"""
_default_sample_shape = torch.Size([512])
def __init__(self, sampler: Optional[MCSampler] = None) -> None:
r"""Register the sampler on the acquisition function.
Args:
sampler: The sampler used to draw base samples for MC-based acquisition
functions. If `None`, a sampler is generated on the fly within
the `get_posterior_samples` method using `get_sampler`.
"""
self.sampler = sampler
[docs]
def get_posterior_samples(self, posterior: Posterior) -> Tensor:
r"""Sample from the posterior using the sampler.
Args:
posterior: The posterior to sample from.
"""
if self.sampler is None:
self.sampler = get_sampler(
posterior=posterior, sample_shape=self._default_sample_shape
)
return self.sampler(posterior=posterior)
@property
def sample_shape(self) -> torch.Size:
return (
self.sampler.sample_shape
if self.sampler is not None
else self._default_sample_shape
)
[docs]
class MultiModelAcquisitionFunction(AcquisitionFunction, ABC):
r"""Abstract base class for acquisition functions that require
multiple types of models.
The intended use case for these acquisition functions are those
where we have multiple models, each serving a distinct purpose.
As an example, we can have a "regression" model that predicts
one or more outcomes, and a "classification" model that predicts
the probabilty that a given parameterization is feasible. The
multi-model acquisition function can then weight the acquisition
value computed with the "regression" model with the feasibility
value predicted by the "classification" model to produce the
composite acquisition value.
This is currently only a placeholder to help with some development
in Ax. We plan to add some acquisition functions utilizing multiple
models in the future.
"""
def __init__(self, model_dict: ModelDict) -> None:
r"""Constructor for the MultiModelAcquisitionFunction base class.
Args:
model_dict: A ModelDict mapping labels to models.
"""
super(AcquisitionFunction, self).__init__()
self.model_dict: ModelDict = model_dict