GPyTorch
latest

Tutorials:

  • GPyTorch Regression Tutorial
    • Introduction
      • Set up training data
    • Setting up the model
      • The GP Model
      • The likelihood
      • Model modes
    • Training the model
    • Make predictions with the model
    • Plot the model fit

Examples:

  • Basic Usage
    • Hyperparameters in GPyTorch
      • Defining an example model
      • Viewing model hyperparameters
        • Raw vs Actual Parameters
        • How do constraints work?
        • Convenience Getters/Setters for Transformed Values
      • Changing Parameter Constraints
      • Priors
      • Putting it Together
      • Initializing hyperparameters in One Call
    • Saving and Loading Models
      • Saving a Simple Model
        • Change Model State
        • Getting Model State
        • Saving Model State
        • Loading Model State
      • A More Complex Example
        • Getting Model State
    • Implementing a custom kernel in GPyTorch
      • A first kernel
      • Adding hyperparameters
    • Metrics in GPyTorch
      • Negative Log Predictive Density (NLPD)
      • Mean Standardized Log Loss (MSLL)
      • Mean Squared Error (MSE)
      • Mean Absolute Error (MAE)
  • Exact GPs (Regression)
    • GPyTorch Regression Tutorial
      • Introduction
        • Set up training data
      • Setting up the model
        • The GP Model
        • The likelihood
        • Model modes
      • Training the model
      • Make predictions with the model
      • Plot the model fit
    • Spectral GP Learning with Deltas
      • Load Data
      • Define a Model
      • Train
      • Plot Results
    • GP Regression with a Spectral Mixture Kernel
      • Introduction
      • Set up the model
    • Fully Bayesian GPs - Sampling Hyperparamters with NUTS
      • Running Sampling
      • Loading Samples
      • Plot Mean Functions
      • Simulate Loading Model from Disk
    • GP Regression with Uncertain Inputs
      • Introduction
        • Using a distributional kernel to deal with uncertain inputs
    • Exact GP Regression on Classification Labels
      • Generate Data
      • Setting Up the Model
      • Model Predictions
  • Exact GPs with Scalable (GPU) Inference
    • BlackBox Matrix-Matrix Inference (BBMM)
    • LancZos Variance Estimates (LOVE)
      • GP Regression with LOVE for Fast Predictive Variances and Sampling
        • Overview
        • The KISS-GP + LOVE GP Model
        • Computing predictive variances (KISS-GP or Exact GPs)
        • Computing posterior samples (KISS-GP only)
    • Exact GPs with GPU Acceleration
      • GPyTorch Regression Tutorial (GPU)
        • Introduction
        • Using the GPU
      • Exact GP Regression with Multiple GPUs and Kernel Partitioning
        • Introduction
        • How many GPUs do you want to use?
        • Automatically determining GPU Settings
        • Computing test time caches
      • GPyTorch Regression With KeOps
        • Introduction
        • Using KeOps with a GPyTorch Model
    • Scalable Posterior Sampling with CIQ
      • Scalable Exact GP Posterior Sampling using Contour Integral Quadrature
        • Are we running with KeOps?
        • Define an Exact GP Model and train
        • Define test set
        • Draw a sample with CIQ
    • Scalable Kernel Approximations
      • Sparse Gaussian Process Regression (SGPR)
        • Overview
        • Defining the SGPR Model
      • Structured Kernel Interpollation (SKI/KISS-GP)
        • Overview
        • KISS-GP for 1D Data
        • KISS-GP for 2D-4D Data
        • KISS-GP for higher dimensional data w/ Additive Structure
        • Scaling to more dimensions (without additive structure)
      • Scalable Kernel Interpolation for Product Kernels (SKIP)
        • Overview
        • Defining the SKIP GP Model
    • Structure-Exploiting Kernels
      • GP Regression with Grid Structured Training Data
        • Make the grid and training data
        • Creating the Grid GP Model
  • Multitask/Multioutput GPs with Exact Inference
    • Multi-output (vector valued functions)
      • Multitask GP Regression
        • Introduction
        • Define a multitask model
      • Batch Independent Multioutput GP
        • Introduction
        • Define a batch GP model
      • ModelList (Multi-Output) GP Regression
        • Introduction
        • Set up the sub-models
    • Scalar function with multiple tasks
      • Hadamard Multitask GP Regression
        • Introduction
        • Set up a Hadamard multitask model
  • Gaussian Process Latent Variable Models (GPLVM) with SVI
    • Introduction
      • Set up training data
    • Defining the GPLVM model
    • Training the model
    • Visualising the 2d latent subspace
  • Variational and Approximate GPs
    • Stochastic Variational GP Regression
      • Overview
      • Creating a DataLoader
      • Creating a SVGP Model
        • The GP Model
    • Modifying the Variational Strategy/Variational Distribution
      • Experimental setup
      • Some quick training/testing code
        • The Standard Approach
        • Reducing parameters
      • MeanFieldVariationalDistribution: a diagonal \(\mathbf S\) matrix
      • DeltaVariationalDistribution: no \(\mathbf S\) matrix
        • Reducing computation (through decoupled inducing points)
    • Using Natural Gradient Descent with Variational Models
      • Overview
      • What is natural gradient descent (NGD)?
      • Jointly optimizing variational parameters/hyperparameters
      • SVGP models for NGD
        • Difference #1: NaturalVariationalDistribution
        • Difference #2: Two optimizers - one for the variational parameters; one for the hyperparameters
        • Difference #3: The updated training loop
        • Evaluation
    • Stochastic Variational GP Regression with Contour Integral Quadrature
      • Overview
      • DataLoaders with CIQ-SVGP
      • Number of inducing points
      • CIQ - SVGP models
    • Different objective functions for Approximate GPs
      • Overview
        • Experimental setup
      • Objective Funtion 1) The Variational ELBO
        • How to use the variational ELBO in GPyTorch
      • Objective Funtion 2) The Predictive Log Likelihood
        • How to use the predictive log likelihood in GPyTorch
    • Non-Gaussian Likelihoods
      • Introduction
        • Set up training data
      • Setting up the classification model
        • Model modes
      • Learn the variational parameters (and other hyperparameters)
      • Make predictions with the model
    • Using Pòlya-Gamma Auxiliary Variables for Binary Classification
      • Overview
      • Pòlya-Gamma Augmentation
      • Setup
      • Variational Inference with PG Auxiliaries
        • Setup optimizers
        • Define training loop
        • Visualization and Evaluation
    • Variational GPs w/ Multiple Outputs
      • Introduction
        • Set up training data
      • Define a multitask model
        • Types of Variational Multitask Models
        • Output modes
        • Train the model
        • Make predictions with the model
    • GP Regression with Uncertain Inputs
      • Introduction
      • Using stochastic variational inference to deal with uncertain inputs
        • Set up training data
      • Setting up the model
      • Training the model with uncertain features
  • Deep GP and Deep Sigma Point Processes
    • Deep GP
    • Deep Sigma Point Processes (DSPP)
      • Deep Gaussian Processes
        • Introduction
        • Defining GP layers
        • Building the deep GP
        • Objective function (approximate marginal log likelihood/ELBO)
        • Training/Testing
      • Deep Sigma Point Processes
        • Basic settings
        • Create PyTorch DataLoader objects
        • Initialize Hidden Layer Inducing Points
        • Create The DSPPHiddenLayer Class
        • Create the DSPP Class
        • Train the Model
        • Make Predictions, compute RMSE and Test NLL
      • Deep GPs and DSPPs w/ Multiple Outputs
        • Introduction
        • Structure of a multitask deep GP
  • PyTorch NN Integration (Deep Kernel Learning)
    • Exact DKL (Deep Kernel Learning) Regression w/ KISS-GP
      • Overview
      • Loading Data
      • Defining the DKL Feature Extractor
      • Defining the DKL-GP Model
        • The forward method
        • Training the model
        • Making Predictions
    • SVDKL (Stochastic Variational Deep Kernel Learning) on CIFAR10/100
      • Set up data augmentation
      • Create DataLoaders
      • Creating the DenseNet Model
      • Creating the GP Layer
      • Creating the full SVDKL Model
      • Defining Training and Testing Code
  • Pyro Integration
    • High-level Pyro Interface (for predictive models)
      • Predictions with Pyro + GPyTorch (High-Level Interface)
        • Overview
        • The PyroGP model
        • Performing inference with Pyro
        • Making predictions
        • Next steps
      • Clustered Multitask GP (w/ Pyro/GPyTorch High-Level Interface)
        • Introduction
        • Adding additional latent variables to the likelihood
        • Constructing the PyroGP model
    • Low-level Pyro Interface (for latent function inference)
      • Latent Function Inference with Pyro + GPyTorch (Low-Level Interface)
        • Overview
        • Using the low-level Pyro/GPyTorch interface
        • Performing inference with Pyro
        • Making predictions
        • Next steps
      • Cox Processes (w/ Pyro/GPyTorch Low-Level Interface)
        • Introduction
        • Parameterizing the intensity function using a GP
        • The Cox process likelihood
        • Using the low-level Pyro/GPyTorch interface
        • Performing inference
        • Evaluate the inferred intensity function
  • Advanced Usage
    • Batch GPs
      • Batch GP Regression
        • Introduction
        • Set up training data
        • Setting up the model
        • Training the model
        • Make predictions with the model
    • GPs with Derivatives
      • GPyTorch regression with derivative information
        • Introduction
        • Setting up the training data
        • Setting up the model
      • GPyTorch regression with derivative information in 2d
        • Introduction
        • Franke function
        • Setting up the training data
        • Setting up the model
    • Variational Fantasization
    • Converting Models to TorchScript
      • Converting Exact GP Models to TorchScript
        • Define and train an exact GP
        • Trace the Model
        • Compare Predictions from TorchScript model and Torch model
      • Converting Variational Models to TorchScript
        • Download Data and Define Model
        • Load a Trained Model
        • Define a Wrapper
        • Trace the Model
      • SVGP Model Updating
        • Training Data
        • Model definition
        • Model Training
        • Model Evaluation
        • Model Updating

Package Reference

  • gpytorch.models
    • Models for Exact GP Inference
      • ExactGP
    • Models for Approximate GP Inference
      • ApproximateGP
    • Models for Deep GPs
      • deep_gps.DeepGP
      • deep_gps.DeepGPLayer
    • Gaussian Process Latent Variable Models (GPLVM)
      • gplvm.BayesianGPLVM
      • gplvm.PointLatentVariable
      • gplvm.MAPLatentVariable
      • gplvm.VariationalLatentVariable
    • Models for integrating with Pyro
      • PyroGP
  • gpytorch.likelihoods
    • Likelihood
    • One-Dimensional Likelihoods
      • GaussianLikelihood
      • GaussianLikelihoodWithMissingObs
      • FixedNoiseGaussianLikelihood
      • DirichletClassificationLikelihood
      • BernoulliLikelihood
      • BetaLikelihood
      • LaplaceLikelihood
      • StudentTLikelihood
    • Multi-Dimensional Likelihoods
      • MultitaskGaussianLikelihood
      • SoftmaxLikelihood
  • gpytorch.kernels
    • Kernel
    • Standard Kernels
      • CosineKernel
      • CylindricalKernel
      • LinearKernel
      • MaternKernel
      • PeriodicKernel
      • PiecewisePolynomialKernel
      • PolynomialKernel
      • PolynomialKernelGrad
      • RBFKernel
      • RQKernel
      • SpectralDeltaKernel
      • SpectralMixtureKernel
    • Composition/Decoration Kernels
      • AdditiveKernel
      • MultiDeviceKernel
      • AdditiveStructureKernel
      • ProductKernel
      • ProductStructureKernel
      • ScaleKernel
    • Specialty Kernels
      • ArcKernel
      • IndexKernel
      • LCMKernel
      • MultitaskKernel
      • RBFKernelGrad
    • Kernels for Scalable GP Regression Methods
      • GridKernel
      • GridInterpolationKernel
      • InducingPointKernel
      • RFFKernel
  • gpytorch.means
    • Mean
    • Standard Means
      • ZeroMean
      • ConstantMean
      • LinearMean
    • Specialty Means
      • MultitaskMean
      • ConstantMeanGrad
  • gpytorch.mlls
    • Exact GP Inference
      • ExactMarginalLogLikelihood
      • LeaveOneOutPseudoLikelihood
    • Approximate GP Inference
      • VariationalELBO
      • PredictiveLogLikelihood
      • GammaRobustVariationalELBO
      • DeepApproximateMLL
  • gpytorch.constraints
    • Parameter Constraints
      • Interval
      • GreaterThan
      • Positive
      • LessThan
  • gpytorch.distributions
    • Distribution
    • MultivariateNormal
    • MultitaskMultivariateNormal
    • Delta
  • gpytorch.priors
    • Prior
    • Standard Priors
      • GammaPrior
      • HalfCauchyPrior
      • LKJCovariancePrior
      • MultivariateNormalPrior
      • NormalPrior
      • SmoothedBoxPrior
  • gpytorch.variational
    • Variational Strategies
      • _VariationalStrategy
      • VariationalStrategy
      • BatchDecoupledVariationalStrategy
      • CiqVariationalStrategy
      • OrthogonallyDecoupledVariationalStrategy
      • UnwhitenedVariationalStrategy
      • GridInterpolationVariationalStrategy
    • Variational Strategies for Multi-Output Functions
      • LMCVariationalStrategy
      • IndependentMultitaskVariationalStrategy
    • Variational Distributions
      • _VariationalDistribution
      • CholeskyVariationalDistribution
      • DeltaVariationalDistribution
      • MeanFieldVariationalDistribution
    • Variational Distributions for Natural Gradient Descent
      • NaturalVariationalDistribution
      • TrilNaturalVariationalDistribution
  • gpytorch.optim
    • NGD

Settings and Beta Features

  • gpytorch.settings
  • gpytorch.beta_features

Advanced Package Reference

  • gpytorch.Module
  • gpytorch.lazy
    • LazyTensor
    • Kernel LazyTensors
    • Structured LazyTensors
      • BlockDiagLazyTensor
      • CatLazyTensor
      • CholLazyTensor
      • DiagLazyTensor
      • MatmulLazyTensor
      • RootLazyTensor
      • NonLazyTensor
      • ToeplitzLazyTensor
      • ZeroLazyTensor
    • Composition/Decoration LazyTensors
      • AddedDiagLazyTensor
      • ConstantMulLazyTensor
      • InterpolatedLazyTensor
      • KroneckerProductLazyTensor
      • MulLazyTensor
      • PsdSumLazyTensor
      • SumBatchLazyTensor
  • gpytorch.functions
    • Functions
  • gpytorch.utils
    • Utilities
      • Lanczos Utilities
      • Permutation Utilities
      • Quadrature Utilities
      • Sparse Utilities
      • Grid Utilities
GPyTorch
  • Docs »
  • <no title>
  • Edit on GitHub

Next Previous

© Copyright 2020, Cornellius GP Revision 538648b9.

Built with Sphinx using a theme provided by Read the Docs.
Read the Docs v: latest
Versions
latest
stable
v1.6.0
v1.5.1
v1.4.2
v1.3.1
Downloads
On Read the Docs
Project Home
Builds

Free document hosting provided by Read the Docs.