Note
Go to the end to download the full example code.
Gaussian Process Regression: use an isotropic covariance kernel¶
In typical machine learning applications, Gaussian process regression surrogate models take several inputs, and those inputs are usually heterogeneous (e.g. in the cantilever beam use case, inputs are various physical quantities).
In geostatistical applications however, inputs are typically spatial
coordinates, which means one can assume the output varies at the same rate
in all directions.
This calls for a specific kind of covariance kernel, represented
in the library by the IsotropicCovarianceModel
class.
# TODO : change reference to plot_gpr_cantilever_beam
Modeling temperature across a surface¶
In this example, we collect temperature data over a floorplan using sensors.
import numpy as np
import openturns as ot
import openturns.experimental as otexp
import matplotlib.pyplot as plt
coordinates = ot.Sample(
[
[100.0, 100.0],
[500.0, 100.0],
[900.0, 100.0],
[100.0, 350.0],
[500.0, 350.0],
[900.0, 350.0],
[100.0, 600.0],
[500.0, 600.0],
[900.0, 600.0],
]
)
observations = ot.Sample(
[[25.0], [25.0], [10.0], [20.0], [25.0], [20.0], [15.0], [25.0], [25.0]]
)
Let us plot the data.
# Extract coordinates.
x = np.array(coordinates[:, 0])
y = np.array(coordinates[:, 1])
# Plot the data with a scatter plot and a color map.
fig = plt.figure()
plt.scatter(x, y, c=observations, cmap="viridis")
plt.colorbar()
plt.show()
Because we are going to view several Gaussian Process Regression models in this example, we use a function to automate the process of optimizing the scale parameter and producing the metamodel.
lower = 50.0
upper = 1000.0
def fitGPR(coordinates, observations, covarianceModel, basis):
"""
Fit the parameters of a Gaussian Process Regression surrogate model.
"""
# Prepare to fit Gaussian process hyperparameters.
fitter = otexp.GaussianProcessFitter(
coordinates, observations, covarianceModel, basis
)
# Set the optimization bounds for the scale parameter to sensible values
# given the data set.
scale_dimension = covarianceModel.getScale().getDimension()
fitter.setOptimizationBounds(
ot.Interval([lower] * scale_dimension, [upper] * scale_dimension)
)
# Fit the GP hyperparameters.
fitter.run()
fitter_result = fitter.getResult()
# Based on the GP hyperparameters perform the regression.
regression = otexp.GaussianProcessRegression(fitter_result)
regression.run()
result = regression.getResult()
surrogate = result.getMetaModel()
return result, surrogate
Let us define a helper function to plot Gaussian Process Regression predictions.
def plotGPRPredictions(gprModel):
"""
Plot the predictions of a Gaussian Process Regression surrogate model.
"""
# Create the mesh of the box [0., 1000.] * [0., 700.]
myInterval = ot.Interval([0.0, 0.0], [1000.0, 700.0])
# Define the number of intervals in each direction of the box
nx = 20
ny = 20
myIndices = [nx - 1, ny - 1]
myMesher = ot.IntervalMesher(myIndices)
myMeshBox = myMesher.build(myInterval)
# Predict
vertices = myMeshBox.getVertices()
predictions = gprModel(vertices)
# Format for plot
X = np.array(vertices[:, 0]).reshape((ny, nx))
Y = np.array(vertices[:, 1]).reshape((ny, nx))
predictions_array = np.array(predictions).reshape((ny, nx))
# Plot
plt.figure()
plt.pcolormesh(X, Y, predictions_array, shading="auto")
plt.colorbar()
plt.show()
return
Predict with an anisotropic geometric covariance kernel¶
In order to illustrate the usefulness of isotropic covariance kernels, we first perform prediction with an anisotropic geometric kernel.
Keep in mind that, when there are more than one input dimension,
the CovarianceModel
classes use a multidimensional
scale parameter .
They are anisotropic geometric by default.
Our example has two input dimensions,
so .
inputDimension = 2
basis = ot.ConstantBasisFactory(inputDimension).build()
covarianceModel = ot.SquaredExponential(inputDimension)
gpr_result, surrogate_model = fitGPR(coordinates, observations, covarianceModel, basis)
plotGPRPredictions(surrogate_model)
We see weird vertical columns on the plot.
How did this happen? Let us have a look at the optimized scale parameter
.
print(gpr_result.getCovarianceModel().getScale())
[50,80.4]
The value of is actually equal to the lower bound:
print(lower)
50.0
This means that temperatures are likely to vary a lot along the X axis and much slower across the Y axis based on the observation data.
Predict with an isotropic covariance kernel¶
If we know that variations of the temperature are isotropic (i.e. with no priviledged direction), we can embed this information within the covariance kernel.
isotropic = ot.IsotropicCovarianceModel(ot.SquaredExponential(), inputDimension)
The IsotropicCovarianceModel
class creates an isotropic
version with a given input dimension of a CovarianceModel
.
Because is isotropic, it only needs one scale parameter
and it will make sure
at all times
during the optimization.
gpr_result, surrogate_model = fitGPR(coordinates, observations, isotropic, basis)
print(gpr_result.getCovarianceModel().getScale())
[100.9]
Prediction with the isotropic covariance kernel is much more satisfactory.
# sphinx_gallery_thumbnail_number = 3
plotGPRPredictions(surrogate_model)