Measure the ellipticity $\epsilon = \epsilon_i + \gamma$ of all galaxies
$\Longrightarrow$ Noisy tracer of the weak lensing shear $\gamma$
Compute summary statistics based on 2pt functions, e.g. the power spectrum
Run an MCMC to recover a posterior on model parameters, using an analytic likelihood
$$ p(\theta | x ) \propto \underbrace{p(x | \theta)}_{\mathrm{likelihood}} \ \underbrace{p(\theta)}_{\mathrm{prior}}$$
Main limitation: the need for an explicit likelihood
We can only compute from theory the likelihood for simple summary statistics and on large scales
$\Longrightarrow$ We are dismissing a significant fraction of the information!
The search for optimal summary statistics
Pires, Starck, Amara, Réfrégier, and R. Teyssier (2009)
Wavelet peak counting statistics
There still an entire industry today of handcrafting summary statistics (see Ajani et al. 2023):
one-point PDF, peak counts, Minkowski functionals, Betti numbers, persistent homology Betti numbers and heatmap, and scattering transform coefficients...
Traditional work on handcrafted summary statistics had plateaued because of two major limitations
Handcrafted summaries can be more powerful than the power spectrum, but never guaranteed to be optimal (sufficient).
It is impossible to analytically model the likelihood of complex, non-linear summary statistics
$\Longrightarrow$ Until Deep Learning Changed The Game!
First Realisation: Summaries can be learned instead of handcrafted
Second Realisation: Proper Inference with Simulators is possible
Cranmer, Pavez, Louppe (2015)
$$ r(\bf{x}; \theta_0, \theta_1) = \frac{p(\bf{x} | \theta_0)}{p(\bf{x} | \theta_1)} $$
can be estimated by training a classifier to distinguish samples from $p(\bf{x} | \theta_0)$ and $p(\bf{x} | \theta_1)$.
Implicit Cosmological Inference
The land of Simulations and Neural Density Estimation
The alternative to theoretical likelihoods: Simulation-Based Inference
Instead of trying to analytically evaluate the likelihood of
sub-optimal summary statistics, let us build a forward model of the full observables.
$\Longrightarrow$ The simulator becomes the physical model.
Benefits of a forward modeling approach
Model the full information content of the data
(aka "full field inference").
Easy to incorporate systematic effects.
Easy to combine multiple cosmological probes by joint simulations.
Why is it classicaly hard to do simulation-based inference?
The Challenge of Simulation-Based Inference
$$ p(x|\theta) = \int p(x, z | \theta) dz = \int p(x | z, \theta) p(z | \theta) dz $$
Where $z$ are stochastic latent variables of the simulator.
$\Longrightarrow$ This marginal likelihood is intractable!
$\Longrightarrow$ Minimizing the Negative Log Likelihood (NLL) over the joint distribution $p(x, \theta)$
leads to minimizing the KL divergence between the model posterior $q_{\phi}(\theta|x)$ and true posterior $p(\theta | x)$.
Finally, SBI has reached the mainstream: Official DES year 3 SBI wCDM results
Jeffrey et al. (2024)
Alright, so, we know how to do SBI...
Has it delivered everything we hoped for?
Example of unforeseen impact of shortcuts in simulations
Gatti, Jeffrey, Whiteway et al. (2023)
Is it ok to distribute lensing source galaxies randomly in simulations, or should they be clustered?
$\Longrightarrow$ An SBI analysis could be biased by this effect and you would never know it!
How much usable information is there beyond the power spectrum?
Chisari et al. (2018)
Ratio of power spectrum in hydrodynamical simulations vs. N-body simulations
Secco et al. (2021)
DES Y3 Cosmic Shear data vector
$\Longrightarrow$ Can we find non-Gaussian information that is not affected by baryons?
takeways and motivations for a Challenge
Will we be able to exploit all of the information content of
LSST, Euclid, DESI?
$\Longrightarrow$ Not rightaway, but it is not the fault of the inference
methodology!
Deep Learning has redefined the limits of our statistical
tools, creating
additional demand on the accuracy of simulations far
beyond the power spectrum.
Neural compression methods have the downside of being opaque.
It is much harder to detect unknown systematics.
We will need a significant number of
large volume, high resolution simulations.
What the NeurIPS 2025 Fair Universe Challenge will teach us:
Phase I: Strategies for training neural summary statistics with very few simulations
Phase II: Strategies for robustness to unknown model mispecification
And finally, on a personal note, the stakes have never been higher!