Previous generation survey: SDSS
Current generation survey: DES
LSST precursor survey: HSC
Deep Learning does not create new information, but
allows you to manipulate existing knowledge to answer new questions.
Simplest architecture: Multilayer Perceptron (MLP)
Performance can be improved for particular types of data by making use of inductive biases
Observed $y$
$f_\theta$
Unknown $x$
$$ y $$
$$f_\theta$$
$$ \mbox{True } x $$
Let's try to understand the neural network output by looking at the loss function
$$ \mathcal{L} = \sum_{(x_i, y_i) \in \mathcal{D}} \parallel x_i - f_\theta(y_i)\parallel^2 \quad \simeq \quad \int \parallel x - f_\theta(y) \parallel^2 \ p(x,y) \ dx dy $$Your goal:
Building a regression model with a Mean Squared Error loss in JAX/FlaxStill a latent variable model, except the mapping $f_\theta$ is made to be bijective.
We want to solve for the Maximum A Posterior solution:
$$\arg \max - \frac{1}{2} \parallel {\color{Orchid} y} - {\color{SkyBlue} x} \parallel_2^2 + \log p_\theta({\color{SkyBlue} x})$$ This can be done by gradient descent as long as one has access to the score function $\frac{\color{orange} d \color{orange}\log \color{orange}p\color{orange}(\color{orange}x\color{orange})}{\color{orange} d \color{orange}x}$.