Rough paths: machine learning for sequential data

How can rough path theory help us understand complex streams of data?

Status

Ongoing

Introduction

Understanding complex, multimodal, high dimensional streams of data is a key challenge in data science. Rough path theory (RPT) provides us with powerful mathematical tools that can be used to design new models for learning with time series data. Applications include weather modelling, cybersecurity, financial modelling and healthcare decision making. The focus of this interest group will be to discuss various topics at the intersection between RPT, differential equations, deep learning and kernel methods.

Explaining the science

RPT was developed in the 1990s by Terry Lyons with the initial goal of making precise the mechanism of how irregular signals interact with non-linear systems. It provides a deterministic toolbox to recover many classical results in stochastic analysis without using specific probabilistic arguments. In particular, it extends Itô's theory of SDEs far beyond the semimartingale setting. At the heart of it is the challenge of developing a robust solution theory for controlled differential equations driven by very irregular signals.

One of its core tools, the Signature, is a homomorphism from the monoid of paths with concatenation into the grouplike elements embedded into the tensor algebra; it provides a graded and faithful description of a path (up to appropriate reparameterizations) by locally removing the need to look at its fine structure and summarising it over short intervals. Linear functionals on the Signature form a unital algebra (with the shuffle product) that separates points, therefore they form a basis for continuous functions on compact sets of paths. In practice, this translates into the following - almost magical - fact: the problem of learning a complex, highly non-linear function on a dataset of irregular time series can be replaced by a simple, two-steps procedure: 1) extract features from the stream by computing its Signature and 2) perform LINEAR regression on the Signature features.

RPT had many recent, exciting research spin-offs. On the theoretical side, Martin Hairer generalized RPT to construct a robust solution theory for certain classes of ill-posed stochastic PDEs - known as the theory of regularity structures - for which he was awarded a Fields medal in 2014. On the applied side, RPT has enabled the development of various learning tools - such as the signature kernel and the neural controlled/rough differential equation models - which are emerging are leading machine learning tools for time series data.

Aims

We are primarily interested in bringing together mathematicians, statisticians, data scientists and practitioners with an interest in machine learning for sequential data. We will host a series of talks on various topics at the intersection between RPT, differential equations, deep learning and kernel methods.

Talking points

- Neural differential equations.

- Kernel methods for time series.

- Applications to medicine, finance, physics, biology.

How to get involved

Please click here for an up-to-date list of past and future talks.

Please click here to find all past events that have been recorded.

Click here to join us and request sign-up

 

Organisers

Dr Cris Salvi

Lecturer (Assistant Professor) in Mathematics, Imperial College London

Contact info

Cristopher Salvi, [email protected]