Note
Go to the end to download the full example code.
Create a polynomial chaos metamodel from a data set¶
In this example, we create a polynomial chaos expansion (PCE) using a data set. More precisely, given a pair of input and output samples, we create a PCE without the knowledge of the input distribution. In this example, we use a relatively small sample size equal to 80.
In this example we create a global approximation of a model response using polynomial chaos expansion.
Let be the function defined by:
for any .
We assume that
and that and are independent.
An interesting point in this example is that the output is multivariate. This is why we are going to use the getMarginal method in the script in order to select the output marginal that we want to manage.
Simulate input and output samples¶
import openturns as ot
import openturns.viewer as viewer
from matplotlib import pylab as plt
ot.Log.Show(ot.Log.NONE)
We first create the function model.
ot.RandomGenerator.SetSeed(2)
dimension = 2
input_names = ["x1", "x2"]
formulas = ["cos(x1 + x2)", "(x2 + 1) * exp(x1)"]
model = ot.SymbolicFunction(input_names, formulas)
Then we create a sample inputSample and compute the corresponding output sample outputSample.
distribution = ot.Normal(dimension)
samplesize = 80
inputSample = distribution.getSample(samplesize)
outputSample = model(inputSample)
Create the PCE¶
Create a functional chaos model.
The algorithm needs to fit a distribution on the input sample.
To do this, the algorithm in FunctionalChaosAlgorithm
uses the BuildDistribution
static method to fit the distribution to the input sample.
Please read Fit a distribution from an input sample
for an example of this method.
The algorithm does this automatically using the Lilliefors test.
In order to make the algorithm a little faster, we reduce the
value of the sample size used in the Lilliefors test.
ot.ResourceMap.SetAsUnsignedInteger("FittingTest-LillieforsMaximumSamplingSize", 50)
The main topic of this example is to introduce the next constructor of
FunctionalChaosAlgorithm
.
Notice that the only input arguments are the input and output sample.
algo = ot.FunctionalChaosAlgorithm(inputSample, outputSample)
algo.run()
result = algo.getResult()
result
Not all coefficients are selected in this PCE.
Indeed, the default constructor of FunctionalChaosAlgorithm
creates a sparse PCE.
Please read Create a full or sparse polynomial chaos expansion
for more details on this topic.
Get the metamodel.
metamodel = result.getMetaModel()
Plot the second output of our model depending on with . In order to do this, we create a ParametricFunction and set the value of . Then we use the getMarginal method to extract the second output (which index is equal to 1).
x1index = 0
x1value = 0.5
x2min = -3.0
x2max = 3.0
outputIndex = 1
metamodelParametric = ot.ParametricFunction(metamodel, [x1index], [x1value])
graph = metamodelParametric.getMarginal(outputIndex).draw(x2min, x2max)
graph.setLegends(["Metamodel"])
modelParametric = ot.ParametricFunction(model, [x1index], [x1value])
curve = modelParametric.getMarginal(outputIndex).draw(x2min, x2max).getDrawable(0)
curve.setColor("red")
curve.setLegend("Model")
graph.add(curve)
graph.setLegendPosition("lower right")
graph.setXTitle("X2")
graph.setTitle("Metamodel Validation, output #%d" % (outputIndex))
view = viewer.View(graph)
We see that the metamodel fits approximately to the model, except
perhaps for extreme values of .
However, there is a better way of globally validating the metamodel,
using the MetaModelValidation
on a validation design of experiment.
n_valid = 100
inputTest = distribution.getSample(n_valid)
outputTest = model(inputTest)
Plot the corresponding validation graphics.
val = ot.MetaModelValidation(outputTest, metamodel(inputTest))
R2 = val.computeR2Score()
graph = val.drawValidation()
graph.setTitle("Metamodel validation R2=" + str(R2))
view = viewer.View(graph)
The coefficient of predictivity is not extremely satisfactory for the first output, but is would be sufficient for a central dispersion study. The second output has a much more satisfactory R2: only one single extreme point is far from the diagonal of the graphics.
Compute and print Sobol’ indices¶
chaosSI = ot.FunctionalChaosSobolIndices(result)
chaosSI
Let us analyse the results of this global sensitivity analysis.
We see that the first output involves significant multi-indices with higher marginal degree.
For the second output, the first variable is especially significant, with a significant contribution of the interactions. The contribution of the interactions are very significant in this model.
Draw Sobol’ indices.
sensitivityAnalysis = ot.FunctionalChaosSobolIndices(result)
first_order = [sensitivityAnalysis.getSobolIndex(i) for i in range(dimension)]
total_order = [sensitivityAnalysis.getSobolTotalIndex(i) for i in range(dimension)]
input_names = model.getInputDescription()
graph = ot.SobolIndicesAlgorithm.DrawSobolIndices(input_names, first_order, total_order)
graph.setLegendPosition("center")
view = viewer.View(graph)
Testing the sensitivity to the degree¶
With the specific constructor of FunctionalChaosAlgorithm that we use, the FunctionalChaosAlgorithm-MaximumTotalDegree in the ResourceMap configure the maximum degree explored by the algorithm. This degree is a trade-off.
If the maximum degree is too low, the polynomial may miss some coefficients so that the quality is lower than possible.
If the maximum degree is too large, the number of coefficients to explore is too large, so that the coefficients might be poorly estimated.
This is why the following for loop explores various degrees to see the sensitivity of the metamodel predictivity depending on the degree.
The default value of this parameter is 10.
ot.ResourceMap.GetAsUnsignedInteger("FunctionalChaosAlgorithm-MaximumTotalDegree")
10
This is why we explore the values from 1 to 15.
degrees = range(1, 12)
r2 = ot.Sample(len(degrees), 2)
for maximumDegree in degrees:
ot.ResourceMap.SetAsUnsignedInteger(
"FunctionalChaosAlgorithm-MaximumTotalDegree", maximumDegree
)
print("Maximum total degree =", maximumDegree)
algo = ot.FunctionalChaosAlgorithm(inputSample, outputSample)
algo.run()
result = algo.getResult()
metamodel = result.getMetaModel()
for outputIndex in range(2):
val = ot.MetaModelValidation(
outputTest[:, outputIndex], metamodel.getMarginal(outputIndex)(inputTest)
)
r2Value = min(1.0, max(0.0, val.computeR2Score()[0])) # Get lucky.
r2[maximumDegree - degrees[0], outputIndex] = r2Value
Maximum total degree = 1
Maximum total degree = 2
Maximum total degree = 3
Maximum total degree = 4
Maximum total degree = 5
Maximum total degree = 6
Maximum total degree = 7
Maximum total degree = 8
Maximum total degree = 9
Maximum total degree = 10
Maximum total degree = 11
graph = ot.Graph("Predictivity", "Total degree", "R2", True)
cloud = ot.Cloud([[d] for d in degrees], r2[:, 0])
cloud.setLegend("Output #0")
cloud.setPointStyle("bullet")
graph.add(cloud)
cloud = ot.Cloud([[d] for d in degrees], r2[:, 1])
cloud.setLegend("Output #1")
cloud.setColor("red")
cloud.setPointStyle("bullet")
graph.add(cloud)
graph.setLegendPosition("upper right")
view = viewer.View(graph, legend_kw={"bbox_to_anchor": (1.0, 1.0), "loc": "upper left"})
plt.subplots_adjust(right=0.7)
plt.show()
We see that the R2 score increases then gets constant or decreases. A low total polynomial degree is not sufficient to describe the first output with good predictivity. However, the coefficient of predictivity can decrease when the total degree gets greater. The predictivity of the second output seems to be much less satisfactory: a little more work would be required to improve the metamodel.
In this situation, the following methods may be used.
Since the distribution of the input is known, we may want to give this information to the FunctionalChaosAlgorithm. This prevents the algorithm from trying to fit the input distribution which best fit to the data.
We may want to customize the adaptiveStrategy by selecting an enumerate function which best fit to this particular example. In this specific example, however, the interactions plays a great role so that the linear enumerate function may provide better results than the hyperbolic rule.
We may want to customize the projectionStrategy by selecting a method to compute the coefficient which improves the estimation. For example, it might be interesting to try an integration rule instead of the least squares method. Notice that a specific design of experiment is required in this case.
Reset default settings
ot.ResourceMap.Reload()
Total running time of the script: (0 minutes 13.119 seconds)