Measures of path-based nonlinear expansion rates and Lagrangian uncertainty in stochastic flows
October 17, 2018 ยท Declared Dead ยท + Add venue
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Michal Branicki, Kenneth Uda
arXiv ID
1810.07567
Category
math.DS
Cross-listed
cs.IT,
math.PR
Citations
0
Last Checked
2 months ago
Abstract
We develop a probabilistic characterisation of trajectorial expansion rates in non-autonomous stochastic dynamical systems that can be defined over a finite time interval and used for the subsequent uncertainty quantification in Lagrangian (trajectory-based) predictions. These expansion rates are quantified via certain divergences (pre-metrics) between probability measures induced by the laws of the stochastic flow associated with the underlying dynamics. We construct scalar fields of finite-time divergence/expansion rates, show their existence and space-time continuity for general stochastic flows. Combining these divergence rate fields with our 'information inequalities' derived in allows for quantification and mitigation of the uncertainty in path-based observables estimated from simplified models in a way that is amenable to algorithmic implementations, and it can be utilised in information-geometric analysis of statistical estimation and inference, as well as in a data-driven machine/deep learning of coarse-grained models. We also derive a link between the divergence rates and finite-time Lyapunov exponents for probability measures and for path-based observables.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ math.DS
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Linearly-Recurrent Autoencoder Networks for Learning Dynamics
R.I.P.
๐ป
Ghosted
Gradient Descent Only Converges to Minimizers: Non-Isolated Critical Points and Invariant Regions
R.I.P.
๐ป
Ghosted
Eigendecompositions of Transfer Operators in Reproducing Kernel Hilbert Spaces
R.I.P.
๐ป
Ghosted
From rate distortion theory to metric mean dimension: variational principle
R.I.P.
๐ป
Ghosted
Double variational principle for mean dimension
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted