The fourth UQSay seminar, organized by L2S and MSSMAT, will take place on Thursday afternoon, October 3, 2019, at CentraleSupelec Paris-Saclay (Eiffel building, amphi V).
We will have two talks:
Bayesian calibration and validation of a numerical model: an overview
Computer experiments are widely used in industrial studies to
complement or replace costly physical experiments, in many
applications: design, reliability, risk assessment, etc. One main
concern with such a widespread use is the confidence one can have
in the outcome of a numerical simulation, that aims to mimick an
actual physical phenomenon. Indeed, the result of a simulation is
tainted by different sources of uncertainty: numerical,
parametric, due to modelisation and/or extrapolation, to cite only
a few. Quantifying all sources of uncertainty, and their influence
on the result of the study, is the primary goal of the
verification, validation and uncertainty quantification (VVUQ)
framework. An important step of VVUQ is calibration, wherein
uncertain parameters within the computer model are tuned to reduce
the gap between computations and available field measures.
EDF R&D has devoted considerable efforts in the last few years to
develop generic, mathematically well-grounded and computationally
efficient calibration and validation methods, adapted to
industrial applications. Two PhD programs and a post-doc have been
devoted to this subject, whose main outcomes are reviewed in this
talk. Hence, we will present the main methods available today to
quantify and reduce the uncertainty on the result of a numerical
experiment through calibration (from ordinary least squares (OLS)
to sequential strategies adapted to costly black-box models) and
validation, seen as the task of detecting and accounting for a
possible systematic model bias (or model discrepancy) term, based
on Bayesian model averaging. All proposed methods are illustrated
using several industrial case studies, and we discuss available
implementations.
Joint work with Pierre Barbillon, Mathieu Carmassi, Matthieu Chiodetti, Guillaume Damblin, Cédric Gœury, Kaniav Kamary, Éric Parent.
References: arXiv:1711.10016, arXiv:1903.03387, arXiv:1801.01810, arXiv:1808.01932.
Blending Physics-Based numerical simulations and seismic databases using Generative Adversarial Network (GAN)
On the one hand, High Performance Computing (HPC) allows the
numerical simulation of highly complicated physics-based scenarios
accounting, to a certain extent, for Uncertainty Quantification
and Propagation (UQ). On the second hand, Machine Learning (ML)
techniques and Artificial Neural Networks (ANN) have reached
outstanding but yet-not-fully-understood prediction capabilities
for both supervised and unsupervised learning, at least in fields
such as image or speech recognition. Yet, ANN are both prone to
overfitting and highly sensitive to outliers questioning their
usefulness in risk assessment studies. However, development of
generative networks has allowed to better constrain the ANN
responses and quantify the related Uncertainty. Adversarial
training techniques have also appeared to provide a generic and
efficient way to train these generative networks on huge
un-labelled datasets.
In this talk, we will first show how Generative Adversarial
Networks (GAN) can be cast and used in the framework of
Uncertainty quantification. Then we will propose an adversarial
Generative Auto-Encoder aiming at transforming medium resolution
signals obtained by physics-based methods into broadband seismic
signals similar to those recorded in seismic databases.
Joint work with Filippo Gatti.
References: DOI:10.1785/0120170293 and hal-01860115.
Organizers: Julien Bect (L2S) and Fernando Lopez Caballero (MSSMAT).
No registration is needed, but an email would be appreciated if you intend to come.