Gaussian Process Regression: use an isotropic covariance kernel

In typical machine learning applications, Gaussian process regression surrogate models take several inputs, and those inputs are usually heterogeneous (e.g. in the cantilever beam use case, inputs are various physical quantities).

In geostatistical applications however, inputs are typically spatial coordinates, which means one can assume the output varies at the same rate in all directions. This calls for a specific kind of covariance kernel, represented in the library by the IsotropicCovarianceModel class.

# TODO : change reference to plot_gpr_cantilever_beam

Modeling temperature across a surface

In this example, we collect temperature data over a floorplan using sensors.

import numpy as np
import openturns as ot
import openturns.experimental as otexp
import matplotlib.pyplot as plt


coordinates = ot.Sample(
    [
        [100.0, 100.0],
        [500.0, 100.0],
        [900.0, 100.0],
        [100.0, 350.0],
        [500.0, 350.0],
        [900.0, 350.0],
        [100.0, 600.0],
        [500.0, 600.0],
        [900.0, 600.0],
    ]
)
observations = ot.Sample(
    [[25.0], [25.0], [10.0], [20.0], [25.0], [20.0], [15.0], [25.0], [25.0]]
)

Let us plot the data.

# Extract coordinates.
x = np.array(coordinates[:, 0])
y = np.array(coordinates[:, 1])

# Plot the data with a scatter plot and a color map.
fig = plt.figure()
plt.scatter(x, y, c=observations, cmap="viridis")
plt.colorbar()
plt.show()
plot gpr isotropic

Because we are going to view several Gaussian Process Regression models in this example, we use a function to automate the process of optimizing the scale parameter and producing the metamodel.

lower = 50.0
upper = 1000.0


def fitGPR(coordinates, observations, covarianceModel, basis):
    """
    Fit the parameters of a Gaussian Process Regression surrogate model.
    """
    # Prepare to fit Gaussian process hyperparameters.
    fitter = otexp.GaussianProcessFitter(
        coordinates, observations, covarianceModel, basis
    )

    # Set the optimization bounds for the scale parameter to sensible values
    # given the data set.
    scale_dimension = covarianceModel.getScale().getDimension()
    fitter.setOptimizationBounds(
        ot.Interval([lower] * scale_dimension, [upper] * scale_dimension)
    )

    # Fit the GP hyperparameters.
    fitter.run()
    fitter_result = fitter.getResult()

    # Based on the GP hyperparameters perform the regression.
    regression = otexp.GaussianProcessRegression(fitter_result)
    regression.run()
    result = regression.getResult()
    surrogate = result.getMetaModel()
    return result, surrogate

Let us define a helper function to plot Gaussian Process Regression predictions.

def plotGPRPredictions(gprModel):
    """
    Plot the predictions of a Gaussian Process Regression surrogate model.
    """
    # Create the mesh of the box [0., 1000.] * [0., 700.]
    myInterval = ot.Interval([0.0, 0.0], [1000.0, 700.0])

    # Define the number of intervals in each direction of the box
    nx = 20
    ny = 20
    myIndices = [nx - 1, ny - 1]
    myMesher = ot.IntervalMesher(myIndices)
    myMeshBox = myMesher.build(myInterval)

    # Predict
    vertices = myMeshBox.getVertices()
    predictions = gprModel(vertices)

    # Format for plot
    X = np.array(vertices[:, 0]).reshape((ny, nx))
    Y = np.array(vertices[:, 1]).reshape((ny, nx))
    predictions_array = np.array(predictions).reshape((ny, nx))

    # Plot
    plt.figure()
    plt.pcolormesh(X, Y, predictions_array, shading="auto")
    plt.colorbar()
    plt.show()
    return

Predict with an anisotropic geometric covariance kernel

In order to illustrate the usefulness of isotropic covariance kernels, we first perform prediction with an anisotropic geometric kernel.

Keep in mind that, when there are more than one input dimension, the CovarianceModel classes use a multidimensional scale parameter \vect{\theta}. They are anisotropic geometric by default.

Our example has two input dimensions, so \vect{\theta} = (\theta_1, \theta_2).

inputDimension = 2
basis = ot.ConstantBasisFactory(inputDimension).build()
covarianceModel = ot.SquaredExponential(inputDimension)
gpr_result, surrogate_model = fitGPR(coordinates, observations, covarianceModel, basis)
plotGPRPredictions(surrogate_model)
plot gpr isotropic

We see weird vertical columns on the plot. How did this happen? Let us have a look at the optimized scale parameter \hat{\vect{\theta}} = (\hat{\theta}_1, \hat{\theta}_2).

print(gpr_result.getCovarianceModel().getScale())
[50,80.4]

The value of \hat{\theta}_1 is actually equal to the lower bound:

print(lower)
50.0

This means that temperatures are likely to vary a lot along the X axis and much slower across the Y axis based on the observation data.

Predict with an isotropic covariance kernel

If we know that variations of the temperature are isotropic (i.e. with no priviledged direction), we can embed this information within the covariance kernel.

isotropic = ot.IsotropicCovarianceModel(ot.SquaredExponential(), inputDimension)

The IsotropicCovarianceModel class creates an isotropic version with a given input dimension of a CovarianceModel. Because is isotropic, it only needs one scale parameter \theta_{iso} and it will make sure \theta_1 = \theta_2 = \theta_{iso} at all times during the optimization.

gpr_result, surrogate_model = fitGPR(coordinates, observations, isotropic, basis)
print(gpr_result.getCovarianceModel().getScale())
[100.9]

Prediction with the isotropic covariance kernel is much more satisfactory.

# sphinx_gallery_thumbnail_number = 3
plotGPRPredictions(surrogate_model)
plot gpr isotropic