Flow matching (FM) is a method that draws -dimensional () samples from any (data) distribution . FM “transforms” noise to data by means of an ordinary differential equation (ODE). The general FM procedure goes like this: We will show which choice of the marginal vector field ensures that item iv) is satisfied, i.e., FM manages to transform…
In this post, we shed some light on the adjoint state method as used in the famous “Neural ODE” paper [1]. In Section 1, we start by introducing the adjoint state method in its raw form (ODE, loss minimization, adjoint equations), in continuous time (denoted by [C]). If this is already clear to you, then… no…
Consider the problem of measuring the discrepancy between the distributions of two sets of samples and . Amongst various options (KL divergence, Wasserstein distance, etc.), the Maximum Mean Discrepancy (MMD) is a beautifully elegant one, gaining popularity in recent years in the machine learning community. In this post, instead of defining upfront the MMD in…
Proximal Policy Optimization (PPO) algorithm is arguably the default choice in modern reinforcement learning (RL) libraries. In this post we understand how to derive PPO from first principles. First, we brush up our memory on the underlying Markov Decision Process (MDP) model. 1. Preliminaries on Markov Decision Process (MDP) In an MDP, an agent (say,…
This post explores the Gauss’s divergence theorem through intuitive and visual reasoning. To engage the reader’s imagination, we use water flux as our running example, although the reasoning applies to any vector field, e.g., electric, magnetic, heat or gravity field. Moreover, to keep things simple we work on the two dimensions, although the same principles…