1
|
Poole W, Ouldridge TE, Gopalkrishnan M. Autonomous learning of generative models with chemical reaction network ensembles. J R Soc Interface 2025; 22:20240373. [PMID: 39837479 PMCID: PMC11771824 DOI: 10.1098/rsif.2024.0373] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Revised: 08/19/2024] [Accepted: 10/29/2024] [Indexed: 01/23/2025] Open
Abstract
Can a micron-sized sack of interacting molecules autonomously learn an internal model of a complex and fluctuating environment? We draw insights from control theory, machine learning theory, chemical reaction network theory and statistical physics to develop a general architecture whereby a broad class of chemical systems can autonomously learn complex distributions. Our construction takes the form of a chemical implementation of machine learning's optimization workhorse: gradient descent on the relative entropy cost function, which we demonstrate can be viewed as a form of integral feedback control. We show how this method can be applied to optimize any detailed balanced chemical reaction network and that the construction is capable of using hidden units to learn complex distributions.
Collapse
Affiliation(s)
- William Poole
- California Institute of Technology, Pasadena, CA, USA
| | | | | |
Collapse
|
2
|
Lipka-Bartosik P, Perarnau-Llobet M, Brunner N. Thermodynamic computing via autonomous quantum thermal machines. SCIENCE ADVANCES 2024; 10:eadm8792. [PMID: 39231232 PMCID: PMC11758477 DOI: 10.1126/sciadv.adm8792] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/10/2023] [Accepted: 07/30/2024] [Indexed: 09/06/2024]
Abstract
We develop a physics-based model for classical computation based on autonomous quantum thermal machines. These machines consist of few interacting quantum bits (qubits) connected to several environments at different temperatures. Heat flows through the machine are here exploited for computing. The process starts by setting the temperatures of the environments according to the logical input. The machine evolves, eventually reaching a nonequilibrium steady state, from which the output of the computation can be determined via the temperature of an auxilliary finite-size reservoir. Such a machine, which we term a "thermodynamic neuron," can implement any linearly separable function, and we discuss explicitly the cases of NOT, 3-MAJORITY, and NOR gates. In turn, we show that a network of thermodynamic neurons can perform any desired function. We discuss the close connection between our model and artificial neurons (perceptrons) and argue that our model provides an alternative physics-based analog implementation of neural networks, and more generally a platform for thermodynamic computing.
Collapse
Affiliation(s)
| | | | - Nicolas Brunner
- Department of Applied Physics, University of Geneva, 1211 Geneva, Switzerland
| |
Collapse
|
3
|
Badias A, Banerjee AG. Neural Network Layer Algebra: A Framework to Measure Capacity and Compression in Deep Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:10380-10393. [PMID: 37027551 DOI: 10.1109/tnnls.2023.3241100] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
We present a new framework to measure the intrinsic properties of (deep) neural networks. While we focus on convolutional networks, our framework can be extrapolated to any network architecture. In particular, we evaluate two network properties, namely, capacity, which is related to expressivity, and compression, which is related to learnability. Both these properties depend only on the network structure and are independent of the network parameters. To this end, we propose two metrics: the first one, called layer complexity, captures the architectural complexity of any network layer; and, the second one, called layer intrinsic power, encodes how data are compressed along the network. The metrics are based on the concept of layer algebra, which is also introduced in this article. This concept is based on the idea that the global properties depend on the network topology, and the leaf nodes of any neural network can be approximated using local transfer functions, thereby allowing a simple computation of the global metrics. We show that our global complexity metric can be calculated and represented more conveniently than the widely used Vapnik-Chervonenkis (VC) dimension. We also compare the properties of various state-of-the-art architectures using our metrics and use the properties to analyze their accuracy on benchmark image classification datasets.
Collapse
|
4
|
Ingrosso A, Panizon E. Machine learning at the mesoscale: A computation-dissipation bottleneck. Phys Rev E 2024; 109:014132. [PMID: 38366483 DOI: 10.1103/physreve.109.014132] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Accepted: 12/05/2023] [Indexed: 02/18/2024]
Abstract
The cost of information processing in physical systems calls for a trade-off between performance and energetic expenditure. Here we formulate and study a computation-dissipation bottleneck in mesoscopic systems used as input-output devices. Using both real data sets and synthetic tasks, we show how nonequilibrium leads to enhanced performance. Our framework sheds light on a crucial compromise between information compression, input-output computation and dynamic irreversibility induced by nonreciprocal interactions.
Collapse
Affiliation(s)
- Alessandro Ingrosso
- Quantitative Life Sciences, Abdus Salam International Centre for Theoretical Physics, 34151 Trieste, Italy
| | - Emanuele Panizon
- Quantitative Life Sciences, Abdus Salam International Centre for Theoretical Physics, 34151 Trieste, Italy
| |
Collapse
|
5
|
Meyer-Ortmanns H. Heteroclinic networks for brain dynamics. FRONTIERS IN NETWORK PHYSIOLOGY 2023; 3:1276401. [PMID: 38020242 PMCID: PMC10663269 DOI: 10.3389/fnetp.2023.1276401] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/11/2023] [Accepted: 10/16/2023] [Indexed: 12/01/2023]
Abstract
Heteroclinic networks are a mathematical concept in dynamic systems theory that is suited to describe metastable states and switching events in brain dynamics. The framework is sensitive to external input and, at the same time, reproducible and robust against perturbations. Solutions of the corresponding differential equations are spatiotemporal patterns that are supposed to encode information both in space and time coordinates. We focus on the concept of winnerless competition as realized in generalized Lotka-Volterra equations and report on results for binding and chunking dynamics, synchronization on spatial grids, and entrainment to heteroclinic motion. We summarize proposals of how to design heteroclinic networks as desired in view of reproducing experimental observations from neuronal networks and discuss the subtle role of noise. The review is on a phenomenological level with possible applications to brain dynamics, while we refer to the literature for a rigorous mathematical treatment. We conclude with promising perspectives for future research.
Collapse
Affiliation(s)
- Hildegard Meyer-Ortmanns
- School of Science, Constructor University, Bremen, Germany
- Complexity Science Hub Vienna, Vienna, Austria
| |
Collapse
|
6
|
Milburn GJ, Shrapnel S, Evans PW. Physical Grounds for Causal Perspectivalism. ENTROPY (BASEL, SWITZERLAND) 2023; 25:1190. [PMID: 37628219 PMCID: PMC10453189 DOI: 10.3390/e25081190] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 07/27/2023] [Accepted: 07/28/2023] [Indexed: 08/27/2023]
Abstract
We ground the asymmetry of causal relations in the internal physical states of a special kind of open and irreversible physical system, a causal agent. A causal agent is an autonomous physical system, maintained in a steady state, far from thermal equilibrium, with special subsystems: sensors, actuators, and learning machines. Using feedback, the learning machine, driven purely by thermodynamic constraints, changes its internal states to learn probabilistic functional relations inherent in correlations between sensor and actuator records. We argue that these functional relations just are causal relations learned by the agent, and so such causal relations are simply relations between the internal physical states of a causal agent. We show that learning is driven by a thermodynamic principle: the error rate is minimised when the dissipated power is minimised. While the internal states of a causal agent are necessarily stochastic, the learned causal relations are shared by all machines with the same hardware embedded in the same environment. We argue that this dependence of causal relations on such 'hardware' is a novel demonstration of causal perspectivalism.
Collapse
Affiliation(s)
- Gerard J. Milburn
- Centre for Engineered Quantum Systems, School of Mathematics and Physics, The University of Queensland, St. Lucia, QLD 4072, Australia;
| | - Sally Shrapnel
- Centre for Engineered Quantum Systems, School of Mathematics and Physics, The University of Queensland, St. Lucia, QLD 4072, Australia;
| | - Peter W. Evans
- School of Historical and Philosophical Inquiry, The University of Queensland, St. Lucia, QLD 4072, Australia;
| |
Collapse
|
7
|
Pacheco P, Mera E, Fuentes V. Intensive Urbanization, Urban Meteorology and Air Pollutants: Effects on the Temperature of a City in a Basin Geography. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2023; 20:3941. [PMID: 36900952 PMCID: PMC10001953 DOI: 10.3390/ijerph20053941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/13/2023] [Revised: 02/11/2023] [Accepted: 02/16/2023] [Indexed: 06/18/2023]
Abstract
A qualitative study of thermal transfers is carried out from a record of measurements (time series) of meteorological variables (temperature, relative humidity and magnitude of wind speeds) and pollutants (PM10, PM2.5 and CO) in six localities located at different heights in the geographic basin of Santiago de Chile. The measurements were made in two periods, 2010-2013 and 2017-2020 (a total of 2,049,336 data), the last period coinciding with a process of intense urbanization, especially high-rise construction. The measurements, in the form of hourly time series, are analyzed on the one hand according to the theory of thermal conduction discretizing the differential equation of the temporal variation in the temperature and, on the other hand, through the theory of chaos that provides the entropies (S). Both procedures demonstrate, comparatively, that the last period of intense urbanization presents an increase in thermal transfers and temperature, which affects urban meteorology and makes it more complex. As shown by the chaotic analysis, there is a faster loss of information for the period 2017-2020. The consequences of the increase in temperature on human health and learning processes are studied.
Collapse
|
8
|
Thermodynamic fluctuation theorems govern human sensorimotor learning. Sci Rep 2023; 13:869. [PMID: 36650215 PMCID: PMC9845310 DOI: 10.1038/s41598-023-27736-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2022] [Accepted: 01/06/2023] [Indexed: 01/18/2023] Open
Abstract
The application of thermodynamic reasoning in the study of learning systems has a long tradition. Recently, new tools relating perfect thermodynamic adaptation to the adaptation process have been developed. These results, known as fluctuation theorems, have been tested experimentally in several physical scenarios and, moreover, they have been shown to be valid under broad mathematical conditions. Hence, although not experimentally challenged yet, they are presumed to apply to learning systems as well. Here we address this challenge by testing the applicability of fluctuation theorems in learning systems, more specifically, in human sensorimotor learning. In particular, we relate adaptive movement trajectories in a changing visuomotor rotation task to fully adapted steady-state behavior of individual participants. We find that human adaptive behavior in our task is generally consistent with fluctuation theorem predictions and discuss the merits and limitations of the approach.
Collapse
|
9
|
Hack P, Gottwald S, Braun DA. Jarzyski's Equality and Crooks' Fluctuation Theorem for General Markov Chains with Application to Decision-Making Systems. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1731. [PMID: 36554136 PMCID: PMC9777588 DOI: 10.3390/e24121731] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/07/2022] [Revised: 11/19/2022] [Accepted: 11/24/2022] [Indexed: 06/17/2023]
Abstract
We define common thermodynamic concepts purely within the framework of general Markov chains and derive Jarzynski's equality and Crooks' fluctuation theorem in this setup. In particular, we regard the discrete-time case, which leads to an asymmetry in the definition of work that appears in the usual formulation of Crooks' fluctuation theorem. We show how this asymmetry can be avoided with an additional condition regarding the energy protocol. The general formulation in terms of Markov chains allows transferring the results to other application areas outside of physics. Here, we discuss how this framework can be applied in the context of decision-making. This involves the definition of the relevant quantities, the assumptions that need to be made for the different fluctuation theorems to hold, as well as the consideration of discrete trajectories instead of the continuous trajectories, which are relevant in physics.
Collapse
Affiliation(s)
- Pedro Hack
- Institute of Neural Information Processing, Ulm University, 89081 Ulm, Germany
| | | | | |
Collapse
|
10
|
Optimal Population Coding for Dynamic Input by Nonequilibrium Networks. ENTROPY 2022; 24:e24050598. [PMID: 35626482 PMCID: PMC9140425 DOI: 10.3390/e24050598] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/10/2022] [Revised: 04/07/2022] [Accepted: 04/19/2022] [Indexed: 12/04/2022]
Abstract
The efficient coding hypothesis states that neural response should maximize its information about the external input. Theoretical studies focus on optimal response in single neuron and population code in networks with weak pairwise interactions. However, more biological settings with asymmetric connectivity and the encoding for dynamical stimuli have not been well-characterized. Here, we study the collective response in a kinetic Ising model that encodes the dynamic input. We apply gradient-based method and mean-field approximation to reconstruct networks given the neural code that encodes dynamic input patterns. We measure network asymmetry, decoding performance, and entropy production from networks that generate optimal population code. We analyze how stimulus correlation, time scale, and reliability of the network affect optimal encoding networks. Specifically, we find network dynamics altered by statistics of the dynamic input, identify stimulus encoding strategies, and show optimal effective temperature in the asymmetric networks. We further discuss how this approach connects to the Bayesian framework and continuous recurrent neural networks. Together, these results bridge concepts of nonequilibrium physics with the analyses of dynamics and coding in networks.
Collapse
|
11
|
Boussard A, Fessel A, Oettmeier C, Briard L, Döbereiner HG, Dussutour A. Adaptive behaviour and learning in slime moulds: the role of oscillations. Philos Trans R Soc Lond B Biol Sci 2021; 376:20190757. [PMID: 33487112 PMCID: PMC7935053 DOI: 10.1098/rstb.2019.0757] [Citation(s) in RCA: 31] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/27/2020] [Indexed: 12/11/2022] Open
Abstract
The slime mould Physarum polycephalum, an aneural organism, uses information from previous experiences to adjust its behaviour, but the mechanisms by which this is accomplished remain unknown. This article examines the possible role of oscillations in learning and memory in slime moulds. Slime moulds share surprising similarities with the network of synaptic connections in animal brains. First, their topology derives from a network of interconnected, vein-like tubes in which signalling molecules are transported. Second, network motility, which generates slime mould behaviour, is driven by distinct oscillations that organize into spatio-temporal wave patterns. Likewise, neural activity in the brain is organized in a variety of oscillations characterized by different frequencies. Interestingly, the oscillating networks of slime moulds are not precursors of nervous systems but, rather, an alternative architecture. Here, we argue that comparable information-processing operations can be realized on different architectures sharing similar oscillatory properties. After describing learning abilities and oscillatory activities of P. polycephalum, we explore the relation between network oscillations and learning, and evaluate the organism's global architecture with respect to information-processing potential. We hypothesize that, as in the brain, modulation of spontaneous oscillations may sustain learning in slime mould. This article is part of the theme issue 'Basal cognition: conceptual tools and the view from the single cell'.
Collapse
Affiliation(s)
- Aurèle Boussard
- Research Centre on Animal Cognition (CRCA), Centre for Integrative Biology (CBI), Toulouse University, CNRS, UPS, Toulouse 31062, France
| | - Adrian Fessel
- Institut für Biophysik, Universität Bremen, Otto-Hahn-Allee 1, 28359 Bremen, Germany
| | - Christina Oettmeier
- Institut für Biophysik, Universität Bremen, Otto-Hahn-Allee 1, 28359 Bremen, Germany
| | - Léa Briard
- Research Centre on Animal Cognition (CRCA), Centre for Integrative Biology (CBI), Toulouse University, CNRS, UPS, Toulouse 31062, France
| | | | - Audrey Dussutour
- Research Centre on Animal Cognition (CRCA), Centre for Integrative Biology (CBI), Toulouse University, CNRS, UPS, Toulouse 31062, France
| |
Collapse
|
12
|
Energetics of stochastic BCM type synaptic plasticity and storing of accurate information. J Comput Neurosci 2021; 49:71-106. [PMID: 33528721 PMCID: PMC8046702 DOI: 10.1007/s10827-020-00775-0] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2019] [Revised: 10/04/2020] [Accepted: 12/13/2020] [Indexed: 11/10/2022]
Abstract
Excitatory synaptic signaling in cortical circuits is thought to be metabolically expensive. Two fundamental brain functions, learning and memory, are associated with long-term synaptic plasticity, but we know very little about energetics of these slow biophysical processes. This study investigates the energy requirement of information storing in plastic synapses for an extended version of BCM plasticity with a decay term, stochastic noise, and nonlinear dependence of neuron’s firing rate on synaptic current (adaptation). It is shown that synaptic weights in this model exhibit bistability. In order to analyze the system analytically, it is reduced to a simple dynamic mean-field for a population averaged plastic synaptic current. Next, using the concepts of nonequilibrium thermodynamics, we derive the energy rate (entropy production rate) for plastic synapses and a corresponding Fisher information for coding presynaptic input. That energy, which is of chemical origin, is primarily used for battling fluctuations in the synaptic weights and presynaptic firing rates, and it increases steeply with synaptic weights, and more uniformly though nonlinearly with presynaptic firing. At the onset of synaptic bistability, Fisher information and memory lifetime both increase sharply, by a few orders of magnitude, but the plasticity energy rate changes only mildly. This implies that a huge gain in the precision of stored information does not have to cost large amounts of metabolic energy, which suggests that synaptic information is not directly limited by energy consumption. Interestingly, for very weak synaptic noise, such a limit on synaptic coding accuracy is imposed instead by a derivative of the plasticity energy rate with respect to the mean presynaptic firing, and this relationship has a general character that is independent of the plasticity type. An estimate for primate neocortex reveals that a relative metabolic cost of BCM type synaptic plasticity, as a fraction of neuronal cost related to fast synaptic transmission and spiking, can vary from negligible to substantial, depending on the synaptic noise level and presynaptic firing.
Collapse
|
13
|
Herpich T, Shayanfard K, Esposito M. Effective thermodynamics of two interacting underdamped Brownian particles. Phys Rev E 2020; 101:022116. [PMID: 32168555 DOI: 10.1103/physreve.101.022116] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2019] [Accepted: 01/25/2020] [Indexed: 06/10/2023]
Abstract
Starting from the stochastic thermodynamics description of two coupled underdamped Brownian particles, we showcase and compare three different coarse-graining schemes leading to an effective thermodynamic description for the first of the two particles: marginalization over one particle, bipartite structure with information flows, and the Hamiltonian of mean force formalism. In the limit of time-scale separation where the second particle with a fast relaxation time scale locally equilibrates with respect to the coordinates of the first slowly relaxing particle, the effective thermodynamics resulting from the first and third approach are shown to capture the full thermodynamics and to coincide with each other. In the bipartite approach, the slow part does not, in general, allow for an exact thermodynamic description as the entropic exchange between the particles is ignored. Physically, the second particle effectively becomes part of the heat reservoir. In the limit where the second particle becomes heavy and thus deterministic, the effective thermodynamics of the first two coarse-graining methods coincide with the full one. The Hamiltonian of mean force formalism, however, is shown to be incompatible with that limit. Physically, the second particle becomes a work source. These theoretical results are illustrated using an exactly solvable harmonic model.
Collapse
Affiliation(s)
- Tim Herpich
- Complex Systems and Statistical Mechanics, Physics and Materials Science Research Unit, University of Luxembourg, L-1511 Luxembourg, Luxembourg
| | - Kamran Shayanfard
- Complex Systems and Statistical Mechanics, Physics and Materials Science Research Unit, University of Luxembourg, L-1511 Luxembourg, Luxembourg
| | - Massimiliano Esposito
- Complex Systems and Statistical Mechanics, Physics and Materials Science Research Unit, University of Luxembourg, L-1511 Luxembourg, Luxembourg
| |
Collapse
|
14
|
Prediction and Dissipation in Nonequilibrium Molecular Sensors: Conditionally Markovian Channels Driven by Memoryful Environments. Bull Math Biol 2020; 82:25. [PMID: 31993762 DOI: 10.1007/s11538-020-00694-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2018] [Accepted: 12/31/2019] [Indexed: 10/25/2022]
Abstract
Biological sensors must often predict their input while operating under metabolic constraints. However, determining whether or not a particular sensor is evolved or designed to be accurate and efficient is challenging. This arises partly from the functional constraints being at cross purposes and partly since quantifying the prediction performance of even in silico sensors can require prohibitively long simulations, especially when highly complex environments drive sensors out of equilibrium. To circumvent these difficulties, we develop new expressions for the prediction accuracy and thermodynamic costs of the broad class of conditionally Markovian sensors subject to complex, correlated (unifilar hidden semi-Markov) environmental inputs in nonequilibrium steady state. Predictive metrics include the instantaneous memory and the total predictable information (the mutual information between present sensor state and input future), while dissipation metrics include power extracted from the environment and the nonpredictive information rate. Success in deriving these formulae relies on identifying the environment's causal states, the input's minimal sufficient statistics for prediction. Using these formulae, we study large random channels and the simplest nontrivial biological sensor model-that of a Hill molecule, characterized by the number of ligands that bind simultaneously-the sensor's cooperativity. We find that the seemingly impoverished Hill molecule can capture an order of magnitude more predictable information than large random channels.
Collapse
|
15
|
Abstract
In order to respond to environmental signals, cells often use small molecular circuits to transmit information about their surroundings. Recently, motivated by specific examples in signaling and gene regulation, a body of work has focused on the properties of circuits that function out of equilibrium and dissipate energy. We briefly review the probabilistic measures of information and dissipation and use simple models to discuss and illustrate trade-offs between information and dissipation in biological circuits. We find that circuits with non-steady state initial conditions can transmit more information at small readout delays than steady state circuits. The dissipative cost of this additional information proves marginal compared to the steady state dissipation. Feedback does not significantly increase the transmitted information for out of steady state circuits but does decrease dissipative costs. Lastly, we discuss the case of bursty gene regulatory circuits that, even in the fast switching limit, function out of equilibrium.
Collapse
|
16
|
Van Vu T, Hasegawa Y. Uncertainty relations for time-delayed Langevin systems. Phys Rev E 2019; 100:012134. [PMID: 31499914 DOI: 10.1103/physreve.100.012134] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2018] [Indexed: 06/10/2023]
Abstract
The thermodynamic uncertainty relation, which establishes a universal trade-off between nonequilibrium current fluctuations and dissipation, has been found for various Markovian systems. However, this relation has not been revealed for non-Markovian systems; therefore, we investigate the thermodynamic uncertainty relation for time-delayed Langevin systems. We prove that the fluctuation of arbitrary dynamical observables is constrained by the Kullback-Leibler divergence between the distributions of the forward path and its reversed counterpart. Specifically, for observables that are antisymmetric under time reversal, the fluctuation is bounded from below by a function of a quantity that can be identified as a generalization of the total entropy production in Markovian systems. We also provide a lower bound for arbitrary observables that are odd under position reversal. The term in this bound reflects the extent to which the position symmetry has been broken in the system and can be positive even in equilibrium. Our results hold for finite observation times and a large class of time-delayed systems because detailed underlying dynamics are not required for the derivation. We numerically verify the derived uncertainty relations using two single time-delay systems and one distributed time-delay system.
Collapse
Affiliation(s)
- Tan Van Vu
- Department of Information and Communication Engineering, Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
| | - Yoshihiko Hasegawa
- Department of Information and Communication Engineering, Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
| |
Collapse
|
17
|
Hasegawa Y, Van Vu T. Uncertainty relations in stochastic processes: An information inequality approach. Phys Rev E 2019; 99:062126. [PMID: 31330674 DOI: 10.1103/physreve.99.062126] [Citation(s) in RCA: 46] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2018] [Indexed: 06/10/2023]
Abstract
The thermodynamic uncertainty relation is an inequality stating that it is impossible to attain higher precision than the bound defined by entropy production. In statistical inference theory, information inequalities assert that it is infeasible for any estimator to achieve an error smaller than the prescribed bound. Inspired by the similarity between the thermodynamic uncertainty relation and the information inequalities, we apply the latter to systems described by Langevin equations, and we derive the bound for the fluctuation of thermodynamic quantities. When applying the Cramér-Rao inequality, the obtained inequality reduces to the fluctuation-response inequality. We find that the thermodynamic uncertainty relation is a particular case of the Cramér-Rao inequality, in which the Fisher information is the total entropy production. Using the equality condition of the Cramér-Rao inequality, we find that the stochastic total entropy production is the only quantity that can attain equality in the thermodynamic uncertainty relation. Furthermore, we apply the Chapman-Robbins inequality and obtain a relation for the lower bound of the ratio between the variance and the sensitivity of systems in response to arbitrary perturbations.
Collapse
Affiliation(s)
- Yoshihiko Hasegawa
- Department of Information and Communication Engineering, Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
| | - Tan Van Vu
- Department of Information and Communication Engineering, Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
| |
Collapse
|
18
|
Ito S. Stochastic Thermodynamic Interpretation of Information Geometry. PHYSICAL REVIEW LETTERS 2018; 121:030605. [PMID: 30085772 DOI: 10.1103/physrevlett.121.030605] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/06/2018] [Indexed: 06/08/2023]
Abstract
In recent years, the unified theory of information and thermodynamics has been intensively discussed in the context of stochastic thermodynamics. The unified theory reveals that information theory would be useful to understand nonstationary dynamics of systems far from equilibrium. In this Letter, we have found a new link between stochastic thermodynamics and information theory well-known as information geometry. By applying this link, an information geometric inequality can be interpreted as a thermodynamic uncertainty relationship between speed and thermodynamic cost. We have numerically applied an information geometric inequality to a thermodynamic model of a biochemical enzyme reaction.
Collapse
Affiliation(s)
- Sosuke Ito
- RIES, Hokkaido University, N20 W10, Kita-ku, Sapporo, Hokkaido 001-0020, Japan
| |
Collapse
|
19
|
Salazar DSP. Nonequilibrium thermodynamics of restricted Boltzmann machines. Phys Rev E 2017; 96:022131. [PMID: 28950614 DOI: 10.1103/physreve.96.022131] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2017] [Indexed: 06/07/2023]
Abstract
In this work, we analyze the nonequilibrium thermodynamics of a class of neural networks known as restricted Boltzmann machines (RBMs) in the context of unsupervised learning. We show how the network is described as a discrete Markov process and how the detailed balance condition and the Maxwell-Boltzmann equilibrium distribution are sufficient conditions for a complete thermodynamics description, including nonequilibrium fluctuation theorems. Numerical simulations in a fully trained RBM are performed and the heat exchange fluctuation theorem is verified with excellent agreement to the theory. We observe how the contrastive divergence functional, mostly used in unsupervised learning of RBMs, is closely related to nonequilibrium thermodynamic quantities. We also use the framework to interpret the estimation of the partition function of RBMs with the annealed importance sampling method from a thermodynamics standpoint. Finally, we argue that unsupervised learning of RBMs is equivalent to a work protocol in a system driven by the laws of thermodynamics in the absence of labeled data.
Collapse
Affiliation(s)
- Domingos S P Salazar
- Unidade de Educação a Distância e Tecnologia, Universidade Federal Rural de Pernambuco, Recife, Pernambuco 52171-900, Brazil
| |
Collapse
|