At the bottom of the graphical model shown above we have the dark matter halo properties and at the top of the model we have the astrophysical parameters.
Application to a complex hydrodynamic simulation
We applied this model to a realistic simulation of the universe where a list of dark matter properties were passed as input to the model and a list of galaxy properties were passed as output to the model. The modern learns the probability distribution of galaxy properties conditioned on dark matter halo properties. Here is an example:
In the contour plot shown here, we see the joint probability distribution of the stellar mass of galaxies and the maximum circular velocity of dark matter halos. The empty(filled) contours correspond to the network prediction(simulation).
We see that the mixture denisty networks provide a fast way of evaluating the hard-to-evaluate conditional probability inside our cosmological inferences.
Semantic segmentation is the task of assigning labels to different pixels in an image. So you can think of it as a binary or multi-label classification. That means the targets have the same size as their corresponding images. How do we organize the directory for such a task and how can we make use of the ImageDataGenerator? Let’s start with a simple example.
I recently read this excellent book on financial machine learning which has a whole chapter dedicated to feature importance and its importance! It offers nice guidelines and some of the best practices for investigating feature importance in problems where we would like to know the extent to which different features contribute to the outcome of a machine learning model. Let’s imagine you are trying to predict whether you need to sell or buy stocks of a certain commidity in the market. Given the historical time evolution of the price, you can hand-engineer a large number of features. This includes:
Named entity recognition (NER) is the practice of recognizing names in a given body of text and identifying the entity category associated with them: persons, locations, etc.
Optimization of deep Neural Networks is often done using Gradient-based methods such as mini-batch gradient descent and its extensions such as Momentm, RMSprop, and Adam. Second order optimization methods such as Newton, BFGS, etc are widely used in different areas of statsitics and Machine Learning. Why are these methods are not popular in deep learning?