Source code for gpytorch.distributions.multivariate_normal

#!/usr/bin/env python3

import math
import warnings

import torch
from torch.distributions import MultivariateNormal as TMultivariateNormal
from torch.distributions.kl import register_kl
from torch.distributions.utils import _standard_normal, lazy_property

from .. import settings
from ..lazy import DiagLazyTensor, LazyTensor, delazify, lazify
from ..utils.broadcasting import _mul_broadcast_shape
from ..utils.warnings import NumericalWarning
from .distribution import Distribution


[docs]class MultivariateNormal(TMultivariateNormal, Distribution): """ Constructs a multivariate normal random variable, based on mean and covariance. Can be multivariate, or a batch of multivariate normals Passing a vector mean corresponds to a multivariate normal. Passing a matrix mean corresponds to a batch of multivariate normals. :param torch.tensor mean: Vector n or matrix b x n mean of mvn distribution. :param ~gpytorch.lazy.LazyTensor covar: Matrix n x n or batch matrix b x n x n covariance of mvn distribution. """ def __init__(self, mean, covariance_matrix, validate_args=False): self._islazy = isinstance(mean, LazyTensor) or isinstance(covariance_matrix, LazyTensor) if self._islazy: if validate_args: ms = mean.size(-1) cs1 = covariance_matrix.size(-1) cs2 = covariance_matrix.size(-2) if not (ms == cs1 and ms == cs2): raise ValueError(f"Wrong shapes in {self._repr_sizes(mean, covariance_matrix)}") self.loc = mean self._covar = covariance_matrix self.__unbroadcasted_scale_tril = None self._validate_args = validate_args batch_shape = _mul_broadcast_shape(self.loc.shape[:-1], covariance_matrix.shape[:-2]) event_shape = self.loc.shape[-1:] # TODO: Integrate argument validation for LazyTensors into torch.distribution validation logic super(TMultivariateNormal, self).__init__(batch_shape, event_shape, validate_args=False) else: super().__init__(loc=mean, covariance_matrix=covariance_matrix, validate_args=validate_args) @property def _unbroadcasted_scale_tril(self): if self.islazy and self.__unbroadcasted_scale_tril is None: # cache root decoposition ust = delazify(self.lazy_covariance_matrix.cholesky()) self.__unbroadcasted_scale_tril = ust return self.__unbroadcasted_scale_tril @_unbroadcasted_scale_tril.setter def _unbroadcasted_scale_tril(self, ust): if self.islazy: raise NotImplementedError("Cannot set _unbroadcasted_scale_tril for lazy MVN distributions") else: self.__unbroadcasted_scale_tril = ust def add_jitter(self, noise=1e-4): return self.__class__(self.mean, self.lazy_covariance_matrix.add_jitter(noise)) def expand(self, batch_size): new_loc = self.loc.expand(torch.Size(batch_size) + self.loc.shape[-1:]) new_covar = self._covar.expand(torch.Size(batch_size) + self._covar.shape[-2:]) res = self.__class__(new_loc, new_covar) return res
[docs] def confidence_region(self): """ Returns 2 standard deviations above and below the mean. :rtype: (torch.Tensor, torch.Tensor) :return: pair of tensors of size (b x d) or (d), where b is the batch size and d is the dimensionality of the random variable. The first (second) Tensor is the lower (upper) end of the confidence region. """ std2 = self.stddev.mul_(2) mean = self.mean return mean.sub(std2), mean.add(std2)
@staticmethod def _repr_sizes(mean, covariance_matrix): return f"MultivariateNormal(loc: {mean.size()}, scale: {covariance_matrix.size()})" @lazy_property def covariance_matrix(self): if self.islazy: return self._covar.evaluate() else: return super().covariance_matrix
[docs] def get_base_samples(self, sample_shape=torch.Size()): """Get i.i.d. standard Normal samples (to be used with rsample(base_samples=base_samples))""" with torch.no_grad(): shape = self._extended_shape(sample_shape) base_samples = _standard_normal(shape, dtype=self.loc.dtype, device=self.loc.device) return base_samples
@lazy_property def lazy_covariance_matrix(self): """ The covariance_matrix, represented as a LazyTensor """ if self.islazy: return self._covar else: return lazify(super().covariance_matrix) def log_prob(self, value): if settings.fast_computations.log_prob.off(): return super().log_prob(value) if self._validate_args: self._validate_sample(value) mean, covar = self.loc, self.lazy_covariance_matrix diff = value - mean # Repeat the covar to match the batch shape of diff if diff.shape[:-1] != covar.batch_shape: if len(diff.shape[:-1]) < len(covar.batch_shape): diff = diff.expand(covar.shape[:-1]) else: padded_batch_shape = (*(1 for _ in range(diff.dim() + 1 - covar.dim())), *covar.batch_shape) covar = covar.repeat( *(diff_size // covar_size for diff_size, covar_size in zip(diff.shape[:-1], padded_batch_shape)), 1, 1, ) # Get log determininat and first part of quadratic form inv_quad, logdet = covar.inv_quad_logdet(inv_quad_rhs=diff.unsqueeze(-1), logdet=True) res = -0.5 * sum([inv_quad, logdet, diff.size(-1) * math.log(2 * math.pi)]) return res def rsample(self, sample_shape=torch.Size(), base_samples=None): covar = self.lazy_covariance_matrix if base_samples is None: # Create some samples num_samples = sample_shape.numel() or 1 # Get samples res = covar.zero_mean_mvn_samples(num_samples) + self.loc.unsqueeze(0) res = res.view(sample_shape + self.loc.shape) else: # Make sure that the base samples agree with the distribution if self.loc.shape != base_samples.shape[-self.loc.dim() :]: raise RuntimeError( "The size of base_samples (minus sample shape dimensions) should agree with the size " "of self.loc. Expected ...{} but got {}".format(self.loc.shape, base_samples.shape) ) # Determine what the appropriate sample_shape parameter is sample_shape = base_samples.shape[: base_samples.dim() - self.loc.dim()] # Reshape samples to be batch_size x num_dim x num_samples # or num_bim x num_samples base_samples = base_samples.view(-1, *self.loc.shape) base_samples = base_samples.permute(*range(1, self.loc.dim() + 1), 0) # Now reparameterize those base samples covar_root = covar.root_decomposition().root # If necessary, adjust base_samples for rank of root decomposition if covar_root.shape[-1] < base_samples.shape[-2]: base_samples = base_samples[..., : covar_root.shape[-1], :] elif covar_root.shape[-1] > base_samples.shape[-2]: raise RuntimeError("Incompatible dimension of `base_samples`") res = covar_root.matmul(base_samples) + self.loc.unsqueeze(-1) # Permute and reshape new samples to be original size res = res.permute(-1, *range(self.loc.dim())).contiguous() res = res.view(sample_shape + self.loc.shape) return res def sample(self, sample_shape=torch.Size(), base_samples=None): with torch.no_grad(): return self.rsample(sample_shape=sample_shape, base_samples=base_samples)
[docs] def to_data_independent_dist(self): """ Convert a MVN into a batched Normal distribution :returns: the bached data-independent Normal :rtype: gpytorch.distributions.Normal """ # Create batch distribution where all data are independent, but the tasks are dependent try: # If pyro is installed, use that set of base distributions import pyro.distributions as base_distributions except ImportError: # Otherwise, use PyTorch import torch.distributions as base_distributions return base_distributions.Normal(self.mean, self.stddev)
@property def stddev(self): # self.variance is guaranteed to be positive, because we do clamping. return self.variance.sqrt() @property def variance(self): if self.islazy: # overwrite this since torch MVN uses unbroadcasted_scale_tril for this diag = self.lazy_covariance_matrix.diag() diag = diag.view(diag.shape[:-1] + self._event_shape) variance = diag.expand(self._batch_shape + self._event_shape) else: variance = super().variance # Check to make sure that variance isn't lower than minimum allowed value (default 1e-6). # This ensures that all variances are positive min_variance = settings.min_variance.value(variance.dtype) if variance.lt(min_variance).any(): warnings.warn( f"Negative variance values detected. " "This is likely due to numerical instabilities. " f"Rounding negative variances up to {min_variance}.", NumericalWarning, ) variance = variance.clamp_min(min_variance) return variance def __add__(self, other): if isinstance(other, MultivariateNormal): return self.__class__( mean=self.mean + other.mean, covariance_matrix=(self.lazy_covariance_matrix + other.lazy_covariance_matrix), ) elif isinstance(other, int) or isinstance(other, float): return self.__class__(self.mean + other, self.lazy_covariance_matrix) else: raise RuntimeError("Unsupported type {} for addition w/ MultivariateNormal".format(type(other))) def __radd__(self, other): if other == 0: return self return self.__add__(other) def __mul__(self, other): if not (isinstance(other, int) or isinstance(other, float)): raise RuntimeError("Can only multiply by scalars") if other == 1: return self return self.__class__(mean=self.mean * other, covariance_matrix=self.lazy_covariance_matrix * (other ** 2)) def __truediv__(self, other): return self.__mul__(1.0 / other) def __getitem__(self, idx): if not isinstance(idx, tuple): idx = (idx,) rest_idx = idx[:-1] last_idx = idx[-1] new_mean = self.mean[idx] if len(idx) <= self.mean.dim() - 1 and (Ellipsis not in rest_idx): new_cov = self.lazy_covariance_matrix[idx] elif len(idx) > self.mean.dim(): raise IndexError(f"Index {idx} has too many dimensions") else: # In this case we know last_idx corresponds to the last dimension # of mean and the last two dimensions of lazy_covariance_matrix if isinstance(last_idx, int): new_cov = DiagLazyTensor(self.lazy_covariance_matrix.diag()[(*rest_idx, last_idx)]) elif isinstance(last_idx, slice): new_cov = self.lazy_covariance_matrix[(*rest_idx, last_idx, last_idx)] elif last_idx is (...): new_cov = self.lazy_covariance_matrix[rest_idx] else: new_cov = self.lazy_covariance_matrix[(*rest_idx, last_idx, slice(None, None, None))][..., last_idx] return self.__class__(mean=new_mean, covariance_matrix=new_cov)
@register_kl(MultivariateNormal, MultivariateNormal) def kl_mvn_mvn(p_dist, q_dist): output_shape = _mul_broadcast_shape(p_dist.batch_shape, q_dist.batch_shape) if output_shape != p_dist.batch_shape: p_dist = p_dist.expand(output_shape) if output_shape != q_dist.batch_shape: q_dist = q_dist.expand(output_shape) q_mean = q_dist.loc q_covar = q_dist.lazy_covariance_matrix p_mean = p_dist.loc p_covar = p_dist.lazy_covariance_matrix root_p_covar = p_covar.root_decomposition().root.evaluate() mean_diffs = p_mean - q_mean if isinstance(root_p_covar, LazyTensor): # right now this just catches if root_p_covar is a DiagLazyTensor, # but we may want to be smarter about this in the future root_p_covar = root_p_covar.evaluate() inv_quad_rhs = torch.cat([mean_diffs.unsqueeze(-1), root_p_covar], -1) logdet_p_covar = p_covar.logdet() trace_plus_inv_quad_form, logdet_q_covar = q_covar.inv_quad_logdet(inv_quad_rhs=inv_quad_rhs, logdet=True) # Compute the KL Divergence. res = 0.5 * sum([logdet_q_covar, logdet_p_covar.mul(-1), trace_plus_inv_quad_form, -float(mean_diffs.size(-1))]) return res