Events


Probabilistic sampling for physics: Finding needles in a field of high-dimensional haystacks

Institut Pascal - Université Paris-Saclay , September 4th-22th 2023

Event Website: https://indico.ijclab.in2p3.fr/event/9042/

Organizer: Thomas Swinburne (CINaM, CNRS, Aix-Marseille Université) and Manon Michel (LMBP, CNRS, Université Clermont Auvergne).

Scientific committee: Alain Durmus (École Polytechnique, Massy-Palaiseau), Virginie Ehrlacher (École des Ponts, Marne la Vallée), Guilhem Lavaux (IAP, Paris), Mihai-Cosmin Marinica (CEA, Saclay), Martin Weigel (TU Chemnitz)

Description

This 3-week meeting aims at gathering physicists and mathematicians working on stochastic sampling. The overall goal is to foster new interdisciplinary collaborations to answer the challenges encountered while sampling probability distributions presenting high-energy barriers and/or a large number of metastable states. Such challenges are faced by researchers working in a diverse range of fields in science, from fundamental problems in statistical mechanics to large-scale simulations of materials; in recent years significant progress has been made on several aspects, including:

  • design of Markov processes which break free from a random-walk behavior,
  • numerical methods for dimensionality reduction of input data,
  • mode-hopping non-local moves by generative models,
  • density region detection,
  • and analytical derivations of robustness guarantees.

By bringing together established experts and early-career researchers across a wide range of disciplines we aim to find new synergies to answer important and general questions, including:

  • How to associate fastest local mixing strategies with mode-hopping moves?
  • What physics can we preserve and extract along the dynamics?
  • How to build reduced representations that can detect previously unseen events or regions?

Workshop on Scaling in Piecewise Deterministic Markov Processes

LMBP - ROOM 218, Université Clermont Auvergne , March 6th-9th 2023

Organizer: Please contact Manon Michel (LMBP, Université Clermont Auvergne) for any inquiry.

Participants: Andreas Eberle (Bonn University), Francis Lörle (Bonn university), Katharina Schuh (TU Wien), Pierre Monmarché (Sorbonne Université), Arnaud Guillin (UCA), Léo Hahn (UCA), Manon Michel (UCA)

Probabilistic numerical approaches in the context of large-scale physics

LMBP - ROOM 218, Université Clermont Auvergne , June 28th 2022

Organizer: Please contact Manon Michel (LMBP, Université Clermont Auvergne) for any inquiry.

Speakers: David Hill (ISIMA/LIMOS, UCA), Jens Jasche (Stockholm University), Guilhem Lavaux (IAP, France), Athina Monemvassitis (LMBP, UCA) and Lars Röhrig (TU Dortmund and LPC, UCA)

Schedule (click on title for abstract)

9.00 AM - 9.50 AM Guilhem Lavaux
The Aquila program: cosmological inference with complicated datasets

Cosmology deals with the study of the universe as a singular physical object with specific global properties. Over the last century, scientists have tried to build a consistent picture of our universe from astronomical data sets. Unfortunately, those data sets are both incomplete and complicated to interpret.

Over the last ten years, new data assimilation techniques got developed and made possible through advances in statistics, computer science, and a significant increase in the quantity and quality of data. The Aquila consortium intends to push the scientific analysis of those data sets to the next level. I wish to provide a panorama of the activities of the Aquila consortium. I will present the specific challenges of cosmological datasets and some of our statistical modeling techniques: from “likelihood full” with the BORG algorithm (Bayesian Origin Reconstruction from Galaxies) to “implicit likelihood” methods that rely on the use of neural networks. Beyond pure cosmological inference, members of Aquila also validate results by correlating with other data sets. This phase of validation also sometimes provides additional constraints.

9.50 AM - 10.40 AM Jens Jasche
Large-scale Bayesian inference of cosmic structures in galaxy surveys

The standard model of cosmology predicts a rich phenomenology to test the fundamental physics of the origin of cosmic structure, the accelerating cosmic expansion, and dark matter with next-generation galaxy surveys.

However, progress in the field critically depends on our ability to connect theory with observation and to infer relevant cosmological information from next-generation galaxy data. State-of-the-art data analysis methods focus on extracting only information from a limited number of statistical summaries but ignore significant information in the complex filamentary distribution of the three-dimensional cosmic matter field.

To go beyond classical approaches, I will present our Bayesian physical forward modeling approach aiming at extracting the full physical plausible information from cosmological large-scale structure data in this talk. Using a physics model of structure formation, the approach infers 3D initial conditions from which observed structures originate, maps non-linear density and velocity fields, and provides dynamic structure formation histories including a detailed treatment of uncertainties. A hierarchical Bayes approach paired with an efficient implementation of a Hamiltonian Monte Carlo sampler permits to account for various observational systematic effects while exploring a multi-million-dimensional parameter space. The method will be illustrated through various data applications providing an unprecedented view of the dynamic evolution of structures surrounding us. Inferred mass density fields are in agreement with and provide complementary information to gold-standard gravitational weak lensing and X-ray observations. I will discuss how using Bayesian forward modeling of the three-dimensional cosmic structure permits us to use the cosmic structure as a laboratory for fundamental physics and to gain insights into the cosmic origin, the dark matter, and dark energy phenomenology as well as the nature of gravity. Finally, I will outline a new program to use inferred posterior distributions and information-theoretic concepts to devise new algorithms for optimal acquisition of data and automated scientific discovery.

10.40 AM - 11.10 AM Coffee break
11.10 AM - 11.35 AM Athina Monemvassitis
On the implementation of discrete PDMP-based algorithms

Traditional Markov-chain Monte Carlo methods produce samples from a target probability density by exploring it through a reversible Markov chain. In most cases, this chain satisfies the detail balance condition thanks to the introduction of rejections. Lately, non-reversible algorithms based on continuous Piecewise deterministic Markov processes (PDMP) have been shown to improve the sampling efficiency. Those algorithms produce irreversible ballistic motion while ensuring the global balance by direction changes instead of rejections. Much more efficient than the historical Metropolis-Hastings algorithm, they however require either some upper bound of the gradient of the log of the probability density or knowledge on both its inverse and the zeros of its gradient. In case one does not have direct access to those quantities, a compromise is to discretize the continuous PDMP process, at the cost of introducing direction flips and diminishing the persistent nature of the dynamics. In this talk, after introducing some properties of the continuous PDMPs, I will present the discretization before showing some numerical preliminary results on its efficiency with respect to its continuous counterpart and the Metropolis-Hastings algorithm.

11.35 AM - 12.00 AM Lars Röhrig
The Bayesian Analysis Toolkit and applications Slides here .

This talk faces the main aspects and key features of the Bayesian Analysis Toolkit in julia (BAT.jl), a software package to perform Bayesian inference in the julia programming language. BAT.jl offers a variety of efficient and modular algorithms for sampling, optimization and integration to explore posterior distributions in high-dimensional parameter spaces. After giving an introduction to the main aspects and algorithms of BAT.jl, a use-case is presented to perform indirect searches for physics beyond the Standard Model.

12.00 AM - 12.50 AM David Hill
Stochastic Parallel Simulations, repeatability and reproducibility, what is possible?

Parallel stochastic simulations are too often presented as non-reproducible. However, anyone wishing to produce a computer science work of scientific quality must pay attention to the numerical reproducibility of his simulation results. The main purpose is to obtain at least the same scientific conclusions and when it’s possible the same numerical results. However significant differences can be observed in the results of parallel stochastic simulations if the practitioner fails to apply the best practices. Remembering that pseudo-random number generators are deterministic, it is often possible to reproduce the same numerical results for parallel stochastic simulations by implementing a rigorous method tested up to a billion threads. An interesting point for some practitioners is that method enables to check the parallel results with their sequential counterpart before scaling, thus gaining confidence in the proposed models. This talk will present the concepts of reproducibility and repeatability from an epistemological perspective, then the state of the art for parallel random numbers and the method previously mentioned will be exposed in the current context of high performance computing including silent errors impacting all top 500 machines, including the new Exascale supercomputer.

Workshop funded by the 80 prime project CosmoBayes, associated to SuSa.

On future synergies for stochastic and learning algorithms

CIRM , September 27th – October 1st 2021

Event Website: https://conferences.cirm-math.fr/2389.html

Organizers: Alain Durmus (CMLA, ENS Paris Saclay), Manon Michel (LMBP, Université Clermont Auvergne), Gareth Roberts (Department of statistics, Warwick University) and Lenka Zdeborova (Institut de Physique Théorique, CEA).

Scientific committee: Arnaud Guillin (LMBP, Université Clermont Auvergne), Rémi Monasson (LPENS, ENS Paris), Éric Moulines (CMAP, École Polytechnique), Judith Rousseau (Statistics department, Oxford University), Benjamin Wandelt (IAP; ILP; Flatiron Institute).

Description

This workshop intends to gather computational statisticians, applied mathematicians and theoretical and computational physicists, working on the development and analysis of stochastic and learning algorithms. All these different fields revolve indeed around a probabilistic modelling, either of physical systems or data points, which are then completely characterised by a probability distribution. In physics, given the number of interacting objects is high enough, the system properties can indeed be casted into a Boltzmann distribution, whose corresponding potential function describes the interactions between the objects. In statistics, Bayesian inference replaces standard point-estimates of the parameters by a full probability distribution, which allows to take into consideration a priori information on the problem, but at the cost of having to deal with high-dimensional integrals. The goal is to strengthen the already known connections but more importantly to create new ones around Monte Carlo methods, Bayesian machine learning, probabilistic modelling, inverse problems, uncertainty quantification and applications to large-scale datasets in physics and statistics.

Our goal is to bring people from the different communities to discuss the practical and theoretical problems they face and their own take on them. For that purpose, the conference will be organized around 4 tutorial lectures, invited talks and contributed talks.