Information theory for data-driven model reduction in physics and biology

Type: Preprint

Publication Date: 2024-04-25

Citations: 0

DOI: https://doi.org/10.1101/2024.04.19.590281

Abstract

Model reduction is the construction of simple yet predictive descriptions of the dynamics of many-body systems in terms of a few relevant variables. A prerequisite to model reduction is the identification of these relevant variables, a task for which no general method exists. Here, we develop a systematic approach based on the information bottleneck to identify the relevant variables, defined as those most predictive of the future. We elucidate analytically the relation between these relevant variables and the eigenfunctions of the transfer operator describing the dynamics. Further, we show that in the limit of high compression, the relevant variables are directly determined by the slowest-decaying eigenfunctions. Our information-based approach indicates when to optimally stop increasing the complexity of the reduced model. Furthermore, it provides a firm foundation to construct interpretable deep learning tools that perform model reduction. We illustrate how these tools work in practice by considering uncurated videos of atmospheric flows from which our algorithms automatically extract the dominant slow collective variables, as well as experimental videos of cyanobacteria colonies in which we discover an emergent synchronization order parameter. Significance Statement The first step to understand natural phenomena is to intuit which variables best describe them. An ambitious goal of artificial intelligence is to automate this process. Here, we develop a framework to identify these relevant variables directly from complex datasets. Very much like MP3 compression is about retaining information that matters most to the human ear, our approach is about keeping information that matters most to predict the future. We formalize this insight mathematically and systematically answer the question of when to stop increasing the complexity of minimal models. We illustrate how interpretable deep learning tools built on these ideas reveal emergent collective variables in settings ranging from satellite recordings of atmospheric fluid flows to experimental videos of cyanobacteria colonies.

Locations

  • arXiv (Cornell University) - View - PDF
  • bioRxiv (Cold Spring Harbor Laboratory) - View - PDF

Similar Works

Action Title Year Authors
+ Information theory for model reduction in stochastic dynamical systems 2023 Matthew S. Schmitt
Maciej Koch-Janusz
Michel Fruchart
Daniel S. Seara
Vincenzo Vitelli
+ Objective discovery of dominant dynamical processes with intelligible machine learning 2021 Bryan Kaiser
Juan A. Sáenz
Maike Sonnewald
Daniel Livescu
+ PDF Chat Simultaneous Dimensionality Reduction for Extracting Useful Representations of Large Empirical Multimodal Datasets 2024 Eslam Abdelaleem
+ PDF Chat Which bits went where? Past and future transfer entropy decomposition with the information bottleneck 2024 Kieran A. Murphy
Zifei Yin
Dani S. Bassett
+ Understanding and Designing Complex Systems: Response to "A framework for optimal high-level descriptions in science and engineering---preliminary report" 2014 James P. Crutchfield
Ryan G. James
Sarah Marzen
D. P. Varn
+ PDF Chat Theoretical investigations of an information geometric approach to complexity 2017 Sean Alan Ali
Carlo Cafaro
+ PDF Chat Opportunities for machine learning in scientific discovery 2024 Ricardo Vinuesa
Jean Rabault
Hossein Azizpour
Sebastian Bauer
Bingni W. Brunton
Arne Elofsson
Elias Jarlebring
Hedvig Kjellström
Stefano Markidis
David Marlevi
+ Optimal high-level descriptions of dynamical systems 2014 David H. Wolpert
Joshua A. Grochow
Eric Libby
Simon DeDeo
+ Optimal Latent Representations: Distilling Mutual Information into Principal Pairs 2019 Max Tegmark
+ PDF Chat Learning Interpretable Hierarchical Dynamical Systems Models from Time Series Data 2024 Manuel Brenner
Edward C. Weber
Georgia Koppe
Daniel Durstewitz
+ PDF Chat Information-driven transitions in projections of underdamped dynamics 2022 Giorgio Nicoletti
Amos Maritan
Daniel Maria Busiello
+ The Distributed Information Bottleneck reveals the explanatory structure of complex systems 2022 Kieran Murphy
Dani S. Bassett
+ Latent Representations of Dynamical Systems: When Two is Better Than One 2019 Max Tegmark
+ An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems 2017 Héctor Zenil
Narsis A. Kiani
Francesco Marabita
Yue Deng
Szabolcs Éliás
Angelika Schmidt
G. C. Ball
Jesper Tegnér
+ Multiscale Simulations of Complex Systems by Learning their Effective Dynamics 2020 Pantelis R. Vlachas
Georgios Arampatzis
Caroline Uhler
Petros Koumoutsakos
+ PDF Chat Statistical Mechanics of Dynamical System Identification 2024 Andrei A. Klishin
Joseph Bakarji
J. Nathan Kutz
Krithika Manohar
+ Learning Deep Neural Network Representations for Koopman Operators of Nonlinear Dynamical Systems 2017 Enoch Yeung
Soumya Kundu
Nathan O. Hodas
+ Learning Deep Neural Network Representations for Koopman Operators of Nonlinear Dynamical Systems 2017 Enoch Yeung
Soumya Kundu
Nathan O. Hodas
+ Generative learning for nonlinear dynamics 2023 William Gilpin
+ Recurrences reveal shared causal drivers of complex time series 2023 William Gilpin

Works That Cite This (0)

Action Title Year Authors