#!/usr/bin/env python3
# Copyright (c) Meta Platforms, Inc. and affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
from __future__ import annotations
from collections import defaultdict
from copy import deepcopy
from logging import Logger
from typing import Any, Dict, List, Optional, Set, Tuple, Type
import pandas as pd
from ax.core.arm import Arm
from ax.core.base_trial import BaseTrial, TrialStatus
from ax.core.data import Data
from ax.core.experiment import Experiment
from ax.core.generator_run import GeneratorRun
from ax.core.observation import ObservationFeatures
from ax.exceptions.core import DataRequiredError, NoDataError, UserInputError
from ax.exceptions.generation_strategy import (
GenerationStrategyCompleted,
GenerationStrategyRepeatedPoints,
MaxParallelismReachedException,
)
from ax.modelbridge.base import ModelBridge
from ax.modelbridge.generation_node import GenerationStep
from ax.modelbridge.registry import _extract_model_state_after_gen, ModelRegistryBase
from ax.utils.common.base import Base
from ax.utils.common.logger import _round_floats_for_logging, get_logger
from ax.utils.common.typeutils import not_none
logger: Logger = get_logger(__name__)
MAX_CONDITIONS_GENERATED = 10000
MAX_GEN_DRAWS = 5
MAX_GEN_DRAWS_EXCEEDED_MESSAGE = (
f"GenerationStrategy exceeded `MAX_GEN_DRAWS` of {MAX_GEN_DRAWS} while trying to "
"generate a unique parameterization. This indicates that the search space has "
"likely been fully explored, or that the sweep has converged."
)
[docs]class GenerationStrategy(Base):
"""GenerationStrategy describes which model should be used to generate new
points for which trials, enabling and automating use of different models
throughout the optimization process. For instance, it allows to use one
model for the initialization trials, and another one for all subsequent
trials. In the general case, this allows to automate use of an arbitrary
number of models to generate an arbitrary numbers of trials
described in the `trials_per_model` argument.
Args:
steps: A list of `GenerationStep` describing steps of this strategy.
name: An optional name for this generaiton strategy. If not specified,
strategy's name will be names of its steps' models joined with '+'.
"""
_name: Optional[str]
_steps: List[GenerationStep]
_curr: GenerationStep # Current step in the strategy.
# Whether all models in this GS are in Models registry enum.
_uses_registered_models: bool
# All generator runs created through this generation strategy, in chronological
# order.
_generator_runs: List[GeneratorRun]
# Experiment, for which this generation strategy has generated trials, if
# it exists.
_experiment: Optional[Experiment] = None
# Trial indices as last seen by the model; updated in `_model` property setter.
# pyre-fixme[4]: Attribute must be annotated.
_seen_trial_indices_by_status = None
_model: Optional[ModelBridge] = None # Current model.
def __init__(self, steps: List[GenerationStep], name: Optional[str] = None) -> None:
assert isinstance(steps, list) and all(
isinstance(s, GenerationStep) for s in steps
), "Steps must be a GenerationStep list."
self._name = name
self._steps = steps
self._uses_registered_models = True
self._generator_runs = []
for idx, step in enumerate(self._steps):
if step.num_trials == -1:
if idx < len(self._steps) - 1:
raise UserInputError( # pragma: no cover
"Only last step in generation strategy can have `num_trials` "
"set to -1 to indicate that the model in the step should "
"be used to generate new trials indefinitely."
)
elif step.num_trials < 1: # pragma: no cover
raise UserInputError(
"`num_trials` must be positive or -1 (indicating unlimited) "
"for all generation steps."
)
if step.max_parallelism is not None and step.max_parallelism < 1:
raise UserInputError(
"Maximum parallelism should be None (if no limit) or a positive"
f" number. Got: {step.max_parallelism} for step {step.model_name}."
)
step.index = idx
if not isinstance(step.model, ModelRegistryBase):
self._uses_registered_models = False
if not self._uses_registered_models:
logger.info(
"Using model via callable function, "
"so optimization is not resumable if interrupted."
)
self._curr = steps[0]
self._seen_trial_indices_by_status = None
@property
def name(self) -> str:
"""Name of this generation strategy. Defaults to a combination of model
names provided in generation steps."""
if self._name is not None:
return not_none(self._name)
factory_names = (step.model_name for step in self._steps)
# Trim the "get_" beginning of the factory function if it's there.
factory_names = (n[4:] if n[:4] == "get_" else n for n in factory_names)
self._name = "+".join(factory_names)
return not_none(self._name)
@name.setter
def name(self, name: str) -> None:
"""Set generation strategy name."""
self._name = name
@property
def model_transitions(self) -> List[int]:
"""List of trial indices where a transition happened from one model to
another."""
gen_changes = [step.num_trials for step in self._steps]
return [sum(gen_changes[: i + 1]) for i in range(len(gen_changes))][:-1]
@property
def current_step(self) -> GenerationStep:
"""Current generation step."""
return self._curr # pragma: no cover
@property
def model(self) -> Optional[ModelBridge]:
"""Current model in this strategy. Returns None if no model has been set
yet (i.e., if no generator runs have been produced from this GS).
"""
return self._curr.model_spec._fitted_model
@property
def experiment(self) -> Experiment:
"""Experiment, currently set on this generation strategy."""
if self._experiment is None: # pragma: no cover
raise ValueError("No experiment set on generation strategy.")
return not_none(self._experiment)
@experiment.setter
def experiment(self, experiment: Experiment) -> None:
"""If there is an experiment set on this generation strategy as the
experiment it has been generating generator runs for, check if the
experiment passed in is the same as the one saved and log an information
statement if its not. Set the new experiment on this generation strategy.
"""
if self._experiment is None or experiment._name == self.experiment._name:
self._experiment = experiment
else: # pragma: no cover
raise ValueError(
"This generation strategy has been used for experiment "
f"{self.experiment._name} so far; cannot reset experiment"
f" to {experiment._name}. If this is a new optimization, "
"a new generation strategy should be created instead."
)
@property
def uses_non_registered_models(self) -> bool:
"""Whether this generation strategy involves models that are not
registered and therefore cannot be stored."""
return not self._uses_registered_models
@property
def last_generator_run(self) -> Optional[GeneratorRun]:
"""Latest generator run produced by this generation strategy.
Returns None if no generator runs have been produced yet.
"""
# Used to restore current model when decoding a serialized GS.
return self._generator_runs[-1] if self._generator_runs else None
@property
def trial_indices_by_step(self) -> Dict[int, Set[int]]:
"""Find trials in experiment that are not mapped to a generation step yet
and add them to the mapping of trials by generation step.
"""
trial_indices_by_step = defaultdict(set)
for trial_index, trial in self.experiment.trials.items():
if (
trial._generation_step_index is not None
and trial._generation_step_index <= self._curr.index
):
trial_indices_by_step[trial._generation_step_index].add(trial_index)
return trial_indices_by_step
@property
def trials_as_df(self) -> Optional[pd.DataFrame]:
"""Puts information on individual trials into a data frame for easy
viewing. For example:
Gen. Step | Model | Trial Index | Trial Status | Arm Parameterizations
0 | Sobol | 0 | RUNNING | {"0_0":{"x":9.17...}}
"""
logger.info(
"Note that parameter values in dataframe are rounded to 2 decimal "
"points; the values in the dataframe are thus not the exact ones "
"suggested by Ax in trials."
)
if self._experiment is None or all(
len(trials) == 0 for trials in self.trial_indices_by_step.values()
):
return None
records = [
{
"Generation Step": step_idx,
"Generation Model": self._steps[step_idx].model_name,
"Trial Index": trial_idx,
"Trial Status": self.experiment.trials[trial_idx].status.name,
"Arm Parameterizations": {
arm.name: _round_floats_for_logging(arm.parameters)
for arm in self.experiment.trials[trial_idx].arms
},
}
for step_idx, trials in self.trial_indices_by_step.items()
for trial_idx in trials
]
return pd.DataFrame.from_records(records).reindex(
columns=[
"Generation Step",
"Generation Model",
"Trial Index",
"Trial Status",
"Arm Parameterizations",
]
)
@property
def num_running_trials_this_step(self) -> int:
"""Number of trials in status `RUNNING` for the current generation step
of this strategy.
"""
num_running = 0
for trial in self.experiment.trials.values():
if (
trial._generation_step_index == self._curr.index
and trial.status.is_running
):
num_running += 1
return num_running
@property
def num_can_complete_this_step(self) -> int:
"""Number of trials for the current step in generation strategy that can
be completed (so are not in status `FAILED` or `ABANDONED`). Used to keep
track of how many generator runs (that become trials) can be produced
from the current generation step.
NOTE: This includes `COMPLETED` trials.
"""
step_trials = self.trial_indices_by_step[self._curr.index]
by_status = self.experiment.trial_indices_by_status
# Number of trials that will not be `COMPLETED`, used to avoid counting
# unsuccessfully terminated trials against the number of generated trials
# during determination of whether enough trials have been generated and
# completed to proceed to the next generation step.
num_will_not_complete = len(
step_trials.intersection(
by_status[TrialStatus.FAILED].union(by_status[TrialStatus.ABANDONED])
)
)
return len(step_trials) - num_will_not_complete
@property
def num_completed_this_step(self) -> int:
"""Number of trials in status `COMPLETED` or `EARLY_STOPPED` for
the current generation step of this strategy. We include early
stopped trials because their data will be used in the model,
so they are completed from the model's point of view and should
count towards that total.
"""
step_trials = self.trial_indices_by_step[self._curr.index]
by_status = self.experiment.trial_indices_by_status
return len(
step_trials.intersection(
by_status[TrialStatus.COMPLETED] | by_status[TrialStatus.EARLY_STOPPED]
)
)
[docs] def gen(
self,
experiment: Experiment,
data: Optional[Data] = None,
n: int = 1,
pending_observations: Optional[Dict[str, List[ObservationFeatures]]] = None,
**kwargs: Any,
) -> GeneratorRun:
"""Produce the next points in the experiment. Additional kwargs passed to
this method are propagated directly to the underlying model's `gen`, along
with the `model_gen_kwargs` set on the current generation step.
NOTE: Each generator run returned from this function must become a single
trial on the experiment to comply with assumptions made in generation
strategy. Do not split one generator run produced from generation strategy
into multiple trials (never making a generator run into a trial is allowed).
Args:
experiment: Experiment, for which the generation strategy is producing
a new generator run in the course of `gen`, and to which that
generator run will be added as trial(s). Information stored on the
experiment (e.g., trial statuses) is used to determine which model
will be used to produce the generator run returned from this method.
data: Optional data to be passed to the underlying model's `gen`, which
is called within this method and actually produces the resulting
generator run. By default, data is all data on the `experiment` if
`use_update` is False and only the new data since the last call to
this method if `use_update` is True.
n: Integer representing how many arms should be in the generator run
produced by this method. NOTE: Some underlying models may ignore
the `n` and produce a model-determined number of arms. In that
case this method will also output a generator run with number of
arms that can differ from `n`.
pending_observations: A map from metric name to pending
observations for that metric, used by some models to avoid
resuggesting points that are currently being evaluated.
"""
return self._gen_multiple(
experiment=experiment,
num_generator_runs=1,
data=data,
n=n,
pending_observations=pending_observations,
**kwargs,
)[0]
[docs] def current_generator_run_limit(
self,
) -> Tuple[int, bool]:
"""How many generator runs can this generation strategy generate right now,
assuming each one of them becomes its own trial, and whether optimization
is completed.
NOTE: This method might move the generation strategy to the next step, which
is safe, as the next call to ``gen`` will just pick up from there.
Returns: a two-item tuple of:
- the number of generator runs that can currently be produced, with -1
meaning unlimited generator runs,
- whether optimization is completed and the generation strategy cannot
generate any more generator runs at all.
"""
try:
self._maybe_move_to_next_step(raise_data_required_error=False)
except GenerationStrategyCompleted:
return 0, True
to_gen = self._num_trials_to_gen_and_complete_in_curr_step()[0]
if to_gen < -1:
# `_num_trials_to_gen_and_complete_in_curr_step()` should return value
# of -1 or greater always.
raise RuntimeError( # pragma: no cover
"Number of trials left to generate in current generation step is "
f"{to_gen}. This is an unexpected state of the generation strategy."
)
until_max_parallelism = self._num_remaining_trials_until_max_parallelism(
raise_max_parallelism_reached_exception=False
)
# If there is no limitation on the number of trials in the step and
# there is a parallelism limit, return number of trials until that limit.
if until_max_parallelism is not None and to_gen == -1:
return until_max_parallelism, False
# If there is a limitation on the number of trials in the step and also on
# parallelism, return the number of trials until either one of the limits.
if until_max_parallelism is not None: # NOTE: to_gen must be >= 0 here
return min(to_gen, until_max_parallelism), False
# If there is no limit on parallelism, return how many trials are left to
# gen in this step (might be -1 indicating unlimited).
return to_gen, False
[docs] def clone_reset(self) -> GenerationStrategy:
"""Copy this generation strategy without it's state."""
return GenerationStrategy(name=self.name, steps=self._steps)
def __repr__(self) -> str:
"""String representation of this generation strategy."""
repr = f"GenerationStrategy(name='{self.name}', steps=["
remaining_trials = "subsequent" if len(self._steps) > 1 else "all"
for step in self._steps:
num_trials = (
f"{step.num_trials}" if step.num_trials != -1 else remaining_trials
)
try:
model_name = step.model_name
except TypeError: # pragma: no cover
model_name = "model with unknown name"
repr += f"{model_name} for {num_trials} trials, "
repr = repr[:-2]
repr += "])"
return repr
# ------------------------- Candidate generation helpers. -------------------------
def _gen_multiple(
self,
experiment: Experiment,
num_generator_runs: int,
data: Optional[Data] = None,
n: int = 1,
pending_observations: Optional[Dict[str, List[ObservationFeatures]]] = None,
**kwargs: Any,
) -> List[GeneratorRun]:
"""Produce multiple generator runs at once, to be made into multiple
trials on the experiment.
NOTE: This is used to ensure that maximum paralellism and number
of trials per step are not violated when producing many generator
runs from this generation strategy in a row. Without this function,
if one generates multiple generator runs without first making any
of them into running trials, generation strategy cannot enforce that it only
produces as many generator runs as are allowed by the paralellism
limit and the limit on number of trials in current step.
Args:
experiment: Experiment, for which the generation strategy is producing
a new generator run in the course of `gen`, and to which that
generator run will be added as trial(s). Information stored on the
experiment (e.g., trial statuses) is used to determine which model
will be used to produce the generator run returned from this method.
data: Optional data to be passed to the underlying model's `gen`, which
is called within this method and actually produces the resulting
generator run. By default, data is all data on the `experiment` if
`use_update` is False and only the new data since the last call to
this method if `use_update` is True.
n: Integer representing how many arms should be in the generator run
produced by this method. NOTE: Some underlying models may ignore
the ``n`` and produce a model-determined number of arms. In that
case this method will also output a generator run with number of
arms that can differ from ``n``.
pending_observations: A map from metric name to pending
observations for that metric, used by some models to avoid
resuggesting points that are currently being evaluated.
"""
self.experiment = experiment
self._maybe_move_to_next_step()
self._fit_or_update_current_model(data=data)
# Make sure to not make too many generator runs and
# exceed maximum allowed paralellism for the step.
num_until_max_parallelism = self._num_remaining_trials_until_max_parallelism()
if num_until_max_parallelism is not None:
num_generator_runs = min(num_generator_runs, num_until_max_parallelism)
# Make sure not to extend number of trials expected in step.
if self._curr.enforce_num_trials and self._curr.num_trials > 0:
num_generator_runs = min(
num_generator_runs,
self._curr.num_trials - self.num_can_complete_this_step,
)
generator_runs = []
for _ in range(num_generator_runs):
try:
generator_run = _gen_from_generation_step(
generation_step=self._curr,
input_max_gen_draws=MAX_GEN_DRAWS,
n=n,
pending_observations=pending_observations,
model_gen_kwargs=kwargs,
should_deduplicate=self._curr.should_deduplicate,
arms_by_signature=self.experiment.arms_by_signature,
)
generator_run._generation_step_index = self._curr.index
self._generator_runs.append(generator_run)
generator_runs.append(generator_run)
except DataRequiredError as err:
# Model needs more data, so we log the error and return
# as many generator runs as we were able to produce, unless
# no trials were produced at all (in which case its safe to raise).
if len(generator_runs) == 0: # pragma: no cover
raise # pragma: no cover
logger.debug(f"Model required more data: {err}.") # pragma: no cover
break # pragma: no cover
return generator_runs
# ------------------------- Model selection logic helpers. -------------------------
def _fit_or_update_current_model(self, data: Optional[Data]) -> None:
"""Fits or update the model on the current generation step (does not move
between generation steps).
Args:
data: Optional ``Data`` to fit or update with; if not specified, generation
strategy will obtain the data via ``experiment.lookup_data``.
"""
if self._model is not None and self._curr.use_update:
new_data = self._get_data_for_update(passed_in_data=data)
if new_data is not None:
self._update_current_model(new_data=new_data)
else:
self._fit_current_model(data=self._get_data_for_fit(passed_in_data=data))
self._save_seen_trial_indices()
def _num_trials_to_gen_and_complete_in_curr_step(self) -> Tuple[int, int]:
"""Returns how many generator runs (to be made into a trial each) are left to
generate in current step and how many are left to be completed in it before
this generation strategy can move to the next step.
NOTE: returns (-1, -1) if the number of trials to be generated from the given
step is unlimited (and therefore it must be the last generation step).
"""
if self._curr.num_trials == -1:
return -1, -1
# More than `num_trials` can be generated (if not `enforce_num_trials=False`)
# and more than `min_trials_observed` can be completed (if `min_trials_observed
# < `num_trials`), so `left_to_gen` and `left_to_complete` should be clamped
# to lower bound of 0.
left_to_gen = max(self._curr.num_trials - self.num_can_complete_this_step, 0)
left_to_complete = max(
self._curr.min_trials_observed - self.num_completed_this_step, 0
)
return left_to_gen, left_to_complete
def _num_remaining_trials_until_max_parallelism(
self, raise_max_parallelism_reached_exception: bool = True
) -> Optional[int]:
"""Returns how many generator runs (to be made into a trial each) are left to
generate before the `max_parallelism` limit is reached for the current
generation step.
Args:
raise_max_parallelism_reached_exception: Whether to raise
``MaxParallelismReachedException`` if number of trials running in
this generation step exceeds maximum parallelism for it.
"""
max_parallelism = self._curr.max_parallelism
num_running = self.num_running_trials_this_step
if max_parallelism is None:
return None # There was no `max_parallelism` limit.
if raise_max_parallelism_reached_exception and num_running >= max_parallelism:
raise MaxParallelismReachedException(
step_index=self._curr.index,
model_name=self._curr.model_name,
num_running=num_running,
)
return max_parallelism - num_running
def _maybe_move_to_next_step(self, raise_data_required_error: bool = True) -> bool:
"""Moves this generation strategy to next step if conditions for moving are met.
This method is safe to use both when generating candidates or simply checking
how many generator runs (to be made into trials) can currently be produced.
Conditions for moving to next step:
1. ``num_trials`` in current generation step have been generated (generation
strategy produced that many generator runs, which were then attached to
trials),
2. ``min_trials_observed`` in current generation step have been completed,
3. current step is not the last in this generation strategy.
NOTE: this method raises ``GenerationStrategyCompleted`` error if conditions 1
and 2 above are met, but the current step is the last in generation strategy.
It also raises ``DataRequiredError`` if all conditions below are true:
1. ``raise_data_required_error`` argument is ``True``,
2. ``num_trials`` in current generation step have been generated,
3. ``min_trials_observed`` in current generation step have not been completed,
4. ``enforce_num_trials`` in current generation step is ``True``.
Args:
raise_data_required_error: Whether to raise ``DataRequiredError`` in the
case detailed above. Not raising the error is useful if just looking to
check how many generator runs (to be made into trials) can be produced,
but not actually producing them yet.
Returns:
Whether generation strategy moved to the next step.
"""
to_gen, to_complete = self._num_trials_to_gen_and_complete_in_curr_step()
if to_gen == to_complete == -1: # Unlimited trials, never moving to next step.
return False
enforcing_num_trials = self._curr.enforce_num_trials
trials_left_to_gen = to_gen > 0
trials_left_to_complete = to_complete > 0
# If there is something left to gen or complete, we don't move to next step.
if trials_left_to_gen or trials_left_to_complete:
# Check that minimum observed_trials is satisfied if it's enforced.
raise_error = raise_data_required_error
if raise_error and enforcing_num_trials and not trials_left_to_gen:
raise DataRequiredError(
"All trials for current model have been generated, but not enough "
"data has been observed to fit next model. Try again when more data"
" are available."
)
return False
# If nothing left to gen or complete, move to next step if one is available.
if len(self._steps) == self._curr.index + 1:
raise GenerationStrategyCompleted(
f"Generation strategy {self} generated all the trials as "
"specified in its steps."
)
self._curr = self._steps[self._curr.index + 1]
# Moving to the next step also entails unsetting this GS's model (since
# new step's model will be initialized for the first time, so we don't
# try to `update` it but rather initialize with all the data even if
# `use_update` is true for the new generation step; this is done in
# `self._fit_or_update_current_model).
self._model = None
return True
def _fit_current_model(self, data: Data) -> None:
"""Instantiate the current model with all available data."""
# If last generator run's index matches the current step, extract
# model state from last generator run and pass it to the model
# being instantiated in this function.
lgr = self.last_generator_run
# NOTE: This will not be easily compatible with `GenerationNode`;
# will likely need to find last generator run per model. Not a problem
# for now though as GS only allows `GenerationStep`-s for now.
# Potential solution: store generator runs on `GenerationStep`-s and
# split them per-model there.
model_state_on_lgr = {}
if (
lgr is not None
and lgr._generation_step_index == self._curr.index
and lgr._model_state_after_gen
and self.model
):
# TODO[drfreund]: Consider moving this to `GenerationStep` or
# `GenerationNode`.
model_state_on_lgr = _extract_model_state_after_gen(
generator_run=lgr,
model_class=not_none(self.model).model.__class__,
)
if not data.df.empty:
trial_indices_in_data = sorted(data.df["trial_index"].unique())
logger.debug(f"Fitting model with data for trials: {trial_indices_in_data}")
self._curr.fit(experiment=self.experiment, data=data, **model_state_on_lgr)
self._model = self._curr.model_spec.fitted_model
def _get_data_for_fit(self, passed_in_data: Optional[Data]) -> Data:
if passed_in_data is None:
if self._curr.use_update:
# If the new step is using `update`, it's important to instantiate
# the model with data for completed trials only, so later we can
# update it with data for new trials as they become completed.
# `experiment.lookup_data` can lookup all available data, including
# for non-completed trials (depending on how the experiment's metrics
# implement `fetch_experiment_data`). We avoid fetching data for
# trials with statuses other than `COMPLETED`, by fetching specifically
# for `COMPLETED` trials.
avail_while_running_metrics = {
m.name
for m in self.experiment.metrics.values()
if m.is_available_while_running()
}
if avail_while_running_metrics:
raise NotImplementedError(
f"Metrics {avail_while_running_metrics} are available while "
"trial is running, but use of `update` functionality in "
"generation strategy relies on new data being available upon "
"trial completion."
)
data = self.experiment.lookup_data(
trial_indices=self.experiment.trial_indices_by_status[
TrialStatus.COMPLETED
]
)
else:
data = self.experiment.lookup_data()
else:
data = passed_in_data
# By the time we get here, we will have already transitioned
# to a new step, but if previous step required observed data,
# we should raise an error even if enough trials were completed.
# Such an empty data case does indicate an invalid state; this
# check is to improve the experience of detecting and debugging
# the invalid state that led to this.
previous_step_required_observations = (
self._curr.index > 0
and self._steps[self._curr.index - 1].min_trials_observed > 0
)
if data.df.empty and previous_step_required_observations:
raise NoDataError( # pragma: no cover
f"Observed data is required for generation step #{self._curr.index} "
f"(model {self._curr.model_name}), but fetched data was empty. "
"Something is wrong with experiment setup -- likely metrics do not "
"implement fetching logic (check your metrics) or no data was "
"attached to experiment for completed trials."
)
return data
def _update_current_model(self, new_data: Data) -> None:
"""Update the current model with new data (data for trials that have been
completed since the last call to `GenerationStrategy.gen`).
"""
if self._model is None: # Should not be reachable.
raise ValueError( # pragma: no cover
"Cannot update if no model instantiated."
)
trial_indices_in_new_data = sorted(new_data.df["trial_index"].unique())
logger.info(f"Updating model with data for trials: {trial_indices_in_new_data}")
# TODO[drfreund]: Switch to `self._curr.update` once `GenerationNode` supports
not_none(self._model).update(experiment=self.experiment, new_data=new_data)
def _get_data_for_update(self, passed_in_data: Optional[Data]) -> Optional[Data]:
# Should only pass data that is new since last call to `gen`, to the
# underlying model's `update`.
newly_completed_trials = self._find_trials_completed_since_last_gen()
if len(newly_completed_trials) == 0:
logger.debug(
"There were no newly completed trials since last model update."
)
return None
if passed_in_data is None:
new_data = self.experiment.lookup_data(trial_indices=newly_completed_trials)
if new_data.df.empty:
logger.info(
"No new data is attached to experiment; no need for model update."
)
return None
return new_data # pragma: no cover
elif passed_in_data.df.empty:
logger.info( # pragma: no cover
"Manually supplied data is empty; no need for model update."
)
return None # pragma: no cover
return Data(
# pyre-ignore[6]: Expected `Optional[pd.core.frame.DataFrame]`
# for 1st param. `df` to `Data.__init__` but got `pd.core.series.Series`
df=passed_in_data.df[
passed_in_data.df.trial_index.isin(newly_completed_trials)
]
)
def _restore_model_from_generator_run(
self, models_enum: Optional[Type[ModelRegistryBase]] = None
) -> None:
"""Reinstantiates the most recent model on this generation strategy
from the last generator run it produced.
NOTE: Uses model and model bridge kwargs stored on the generator run, as well
as the model state attributes stored on the generator run.
"""
self._fit_or_update_current_model(
data=self._get_data_for_fit(passed_in_data=None)
)
# ------------------------- State-tracking helpers. -------------------------
def _save_seen_trial_indices(self) -> None:
"""Saves Experiment's `trial_indices_by_status` at the time of the model's
last `gen` (so these `trial_indices_by_status` reflect which trials model
has seen the data for). Useful when `use_update=True` for a given
generation step.
"""
self._seen_trial_indices_by_status = deepcopy(
self.experiment.trial_indices_by_status
)
def _find_trials_completed_since_last_gen(self) -> Set[int]:
"""Retrieves indices of trials that have been completed or updated with data
since the last call to `GenerationStrategy.gen`.
"""
completed_now = self.experiment.trial_indices_by_status[TrialStatus.COMPLETED]
if self._seen_trial_indices_by_status is None:
return completed_now
completed_before = not_none(self._seen_trial_indices_by_status)[
TrialStatus.COMPLETED
]
return completed_now.difference(completed_before)
def _register_trial_data_update(self, trial: BaseTrial) -> None:
"""Registers that a given trial has new data even though it's a trial that has
been completed before. Useful only for generation steps that have `use_update=
True`, as the information registered by this function is used for identifying
new data since last call to `GenerationStrategy.gen`.
"""
# TODO[T65857344]: store information about trial update to pass with `new_data`
# to `model_update`. This information does not need to be stored, since when
# restoring generation strategy from serialized form, all data will is
# refetched and the underlying model is re-fit.
if any(s.use_update for s in self._steps): # pragma: no cover
raise NotImplementedError(
"Updating completed trials with new data is not yet supported for "
"generation strategies that leverage `model.update` functionality."
)
def _gen_from_generation_step(
input_max_gen_draws: int,
generation_step: GenerationStep,
n: int,
pending_observations: Optional[Dict[str, List[ObservationFeatures]]],
# pyre-fixme[2]: Parameter annotation cannot be `Any`.
model_gen_kwargs: Any,
should_deduplicate: bool,
arms_by_signature: Dict[str, Arm],
) -> GeneratorRun:
"""Produces a ``GeneratorRun`` with ``n`` arms using the provided ``model``. if
``should_deduplicate is True``, these arms are deduplicated against previous arms
using rejection sampling before returning. If more than ``input_max_gen_draws``
samples are generated during deduplication, this function produces a
``GenerationStrategyRepeatedPoints`` exception.
"""
# TODO[drfreund]: Consider moving dedulication to generation step itself.
# NOTE: Might need to revisit the behavior of deduplication when
# generating multi-arm generator runs (to be made into batch trials).
should_generate_run = True
generator_run = None
n_gen_draws = 0
# Keep generating until each of `generator_run.arms` is not a duplicate
# of a previous arm, if `should_deduplicate is True`
while should_generate_run:
if n_gen_draws > input_max_gen_draws:
raise GenerationStrategyRepeatedPoints(MAX_GEN_DRAWS_EXCEEDED_MESSAGE)
generator_run = generation_step.gen(
n=n,
pending_observations=pending_observations,
**model_gen_kwargs,
)
should_generate_run = should_deduplicate and any(
arm.signature in arms_by_signature for arm in generator_run.arms
)
n_gen_draws += 1
return not_none(generator_run)