Optimal Work Extraction and the Minimum Description Length Principle
June 08, 2020 Β· Declared Dead Β· π Journal of Statistical Mechanics: Theory and Experiment
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
LΓ©o Touzo, Matteo Marsili, Neri Merhav, Γdgar RoldΓ‘n
arXiv ID
2006.04544
Category
cond-mat.stat-mech
Cross-listed
cs.IT
Citations
7
Venue
Journal of Statistical Mechanics: Theory and Experiment
Last Checked
2 months ago
Abstract
We discuss work extraction from classical information engines (e.g., SzilΓ‘rd) with $N$-particles, $q$ partitions, and initial arbitrary non-equilibrium states. In particular, we focus on their {\em optimal} behaviour, which includes the measurement of a set of quantities $Ξ¦$ with a feedback protocol that extracts the maximal average amount of work. We show that the optimal non-equilibrium state to which the engine should be driven before the measurement is given by the normalised maximum-likelihood probability distribution of a statistical model that admits $Ξ¦$ as sufficient statistics. Furthermore, we show that the minimax universal code redundancy $\mathcal{R}^*$ associated to this model, provides an upper bound to the work that the demon can extract on average from the cycle, in units of $k_{\rm B}T$. We also find that, in the limit of $N$ large, the maximum average extracted work cannot exceed $H[Ξ¦]/2$, i.e. one half times the Shannon entropy of the measurement. Our results establish a connection between optimal work extraction in stochastic thermodynamics and optimal universal data compression, providing design principles for optimal information engines. In particular, they suggest that: (i) optimal coding is thermodynamically efficient, and (ii) it is essential to drive the system into a critical state in order to achieve optimal performance.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β cond-mat.stat-mech
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Unsupervised learning of phase transitions: from principal component analysis to variational autoencoders
π
π
Old Age
Unsupervised Generative Modeling Using Matrix Product States
R.I.P.
π»
Ghosted
Solving Statistical Mechanics Using Variational Autoregressive Networks
R.I.P.
π»
Ghosted
Learning Thermodynamics with Boltzmann Machines
R.I.P.
π»
Ghosted
Information Flows? A Critique of Transfer Entropies
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted