By Jannik Wagner, Katrina Liang, Niclas Popp
Neural ordinary differential equations are a class of neural networks with infinitely many layers that generalize residual networks. In this report we present the fundamental theory behind neural ODEs together with two applications. We explain how neural ODEs can be used for descriptive time series modelling and show an example using COVID-19 infections in Sweden. Application two concerns density estimation using a family of generative models called continuous normalizing flows.
We apply Neural ODEs to epidemic time series data and compare the results to the classical modelling approach using the SEIR equations. The data comprises of the daily new COVID-19 infections in Sweden from 1st November to 20th of April with irregular timestamps. The following figure shows the result of the time series analysis model:
First, we applied Continuous Normalizing Flows (CNFs) to artificial data generated from a two dimensional density based on an image of the flag of Sweden. The GIF shows the contiuous transformation over time of a normal distribution to the learned distribution, using a hypernetwork with depth 3, hpernet width 16, and width 64:
We then experimented with high dimensional real world data and used the MNIST data set as samples from a 784 dimensional distribution. Generated images of our own implementation and more complex CNF trained on MNIST are shown below:
Parts of the code are inspired by code from torchdiffeq and FFJORD.