MITB Banner

Guide To GPyTorch: A Python Library For Gaussian Process Models

Share
GPyTorch

GPyTorch is a PyTorch-based library designed for implementing Gaussian processes. It was introduced by Jacob R. Gardner, Geoff Pleiss, David Bindel, Kilian Q. Weinberger and Andrew Gordon Wilson – researchers at Cornel University (research paper).

Before going into the details of GPyTorch, let us first understand what a Gaussian process means, in short.

Gaussian Process

In probability theory and statistics, the Gaussian process refers to a stochastic process i.e. a collection of random variables indexed by time or space in such a way that each finite collection of the random variables has a multivariate normal distribution (every finite linear combination of the variables is normally distributed). 

You might have heard about statistical inference techniques such as Bayesian inference using which one can represent uncertainty over numeric values like the outcome of a dice roll or the height of a person. Gaussian process instead is a probability distribution over possible functions. Find a detailed description of the Gaussian process here.

Overview of GPyTorch

GPyTorch enables easy creation of flexible, scalable and modular Gaussian process models. It is implemented using PyTorch. It performs GP inference via Blackbox Matrix-Matrix multiplication (BBMM). 

Pros of GPyTorch

  1. Scalability: It enables training of GPs with millions of data points
  2. Modular design: It has the capability of easily integrating GPs with deep neural networks
  3. Speed: It can utilize state-of-the-art inference algorithms (such as SKI/KISS-GP, stochastic Lanczos expansions, LOVE, SKIP, stochastic variational deep kernel learning)  and hardware acceleration using GPUs

Practical implementation

Here’s a demonstration of training an RBF kernel Gaussian process on the following function:

y = sin(2x) + E             …(i)

E ~ (0, 0.04)

 (where 0 is mean of the normal distribution and 0.04 is the variance)

The code has been implemented in Google colab with Python 3.7.10 and GPyTorch 1.4.0 versions. Step-wise explanation of the code is as follows:

  1. Install the GPyTorch library

!pip install gpytorch

  1. Import required libraries
 import math
 import torch
 import gpytorch
 from matplotlib import pyplot as plt
 %matplotlib inline #for visualization plots to appear at the frontend 
  1. Prepare training data
 # Choose regularly spaced 100 points from the interval [0,1] 
 x_train = torch.linspace(0, 1, 100)
 # Compute label as sin(2*pi*x) with Gaussian noise as described by eq.(i) above
 y_train = torch.sin(x_train * (2 * math.pi)) + torch.randn(x_train.size()) * math.sqrt(0.04) 
  1. Define the GP model

We have used exact inference – the simplest form of GP model

 class ExactGPModel(gpytorch.models.ExactGP):
     def __init__(self, x_train, y_train, likelihood):
         super(ExactGPModel, self).__init__(x_train, y_train, likelihood)
         self.mean_module = gpytorch.means.ConstantMean() #prior mean
         #covariance 
         self.covar_module = gpytorch.kernels.ScaleKernel
         (gpytorch.kernels.RBFKernel())
     def forward(self, x):
         mean_x = self.mean_module(x)
         covar_x = self.covar_module(x)
         return gpytorch.distributions.MultivariateNormal(mean_x, covar_x) 

For most of the GP regression models, following objects should be constructed:

  • A ‘GP Model’ which handles most of the inference
  • A ‘Likelihood’
  • A ‘Mean’ defining prior mean of GP
  • A ‘Kernel’ defining covariance of GP
  • A Multivariate Normal Distribution

The two methods defined above are components of the Exact (non-variational) GP model.

The _init_ method takes a likelihood and the training data. It then constructs objects like mean module and kernel module required for the ‘forward’ method of the model. The ‘forward’ method takes in some data x. It returns a multivariate normal distribution with prior mean and covariance computed at x.

  1. Initialize likelihood

lkh = gpytorch.likelihoods.GaussianLikelihood()

Initialize the GP model

model = ExactGPModel(x_train, y_train, lkh)

  1. Find optimal hyperparameters of the model
 model.train()
 lkh.train() 

Output: 

 GaussianLikelihood(
    (noise_covar): HomoskedasticNoise(
                 (raw_noise_constraint): GreaterThan(1.000E-04)
               )
            ) 
  1. Use Adam optimization algorithm

opt = torch.optim.Adam(model.parameters(), lr=0.1)

  1. Define loss for GP

l = gpytorch.mlls.ExactMarginalLogLikelihood(lkh, model)

  1. Compute loss, length scale (i.e. length of twists and turns in the function) and noise for each iteration of the GP
 for i in range(20):
     # Zero gradients from previous iteration
     opt.zero_grad()
     # Store output from the model
     op = model(x_train)
     # Compute loss and backprop gradients
     loss = -l(op, y_train)
     loss.backward()  #back propagation
 #torch.autograd.backward() calculates sum of gradients of given tensors
 #Print the loss, length scale and noise for 20 iterations
     print('Iter %d/%d - Loss: %.3f   lengthscale: %.3f   noise:  
     %.3f' % (
         i + 1, 20, loss.item(),  #iteration number
         model.covar_module.base_kernel.lengthscale.item(), #length scale
         model.likelihood.noise.item() #noise
     )) 
  1. Make prediction with the model
 #Evaluation (predictive posterior) mode
 model.eval()
 lkh.eval() 
  1.  Make predictions by feeding model through likelihood
 with torch.no_grad(), gpytorch.settings.fast_pred_var():
     x_test = torch.linspace(0, 1, 51) 
#equally spaced 51 test points in [0,1]
     observed_pred = likelihood(model(x_test)) 
  1. Plot the fitted model
 with torch.no_grad():  #disable gradient computation
     # Initialize plot
     fig, axis = plt.subplots(1, 1, figsize=(4, 3))
     # Upper and lower confidence bounds
     lower, upper = observed_pred.confidence_region()
     # Plot training data as black stars
     axis.plot(x_train.numpy(), y_train.numpy(), 'k*')
     # Plot predictive means as blue line
     axis.plot(x_test.numpy(), observed_pred.mean.numpy(), 'b')
     # Shade between the lower and upper confidence bounds
     #fill the area showing confidence
     axis.fill_between(x_test.numpy(), lower.numpy(), upper.numpy(),  
     alpha=0.5)
     axis.set_ylim([-3, 3])  #Y-direction limits
     axis.legend(['Observed Data', 'Mean', 'Confidence']) 

Output plot:

References

To get a detailed understanding of the GPyTorch library, refer to the following web links:

PS: The story was written using a keyboard.
Share
Picture of Nikita Shiledarbaxi

Nikita Shiledarbaxi

A zealous learner aspiring to advance in the domain of AI/ML. Eager to grasp emerging techniques to get insights from data and hence explore realistic Data Science applications as well.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India