1
|
Pinchas A, Ben-Gal I, Painsky A. A Comparative Analysis of Discrete Entropy Estimators for Large-Alphabet Problems. ENTROPY (BASEL, SWITZERLAND) 2024; 26:369. [PMID: 38785618 PMCID: PMC11120205 DOI: 10.3390/e26050369] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Revised: 04/25/2024] [Accepted: 04/25/2024] [Indexed: 05/25/2024]
Abstract
This paper presents a comparative study of entropy estimation in a large-alphabet regime. A variety of entropy estimators have been proposed over the years, where each estimator is designed for a different setup with its own strengths and caveats. As a consequence, no estimator is known to be universally better than the others. This work addresses this gap by comparing twenty-one entropy estimators in the studied regime, starting with the simplest plug-in estimator and leading up to the most recent neural network-based and polynomial approximate estimators. Our findings show that the estimators' performance highly depends on the underlying distribution. Specifically, we distinguish between three types of distributions, ranging from uniform to degenerate distributions. For each class of distribution, we recommend the most suitable estimator. Further, we propose a sample-dependent approach, which again considers three classes of distribution, and report the top-performing estimators in each class. This approach provides a data-dependent framework for choosing the desired estimator in practical setups.
Collapse
Affiliation(s)
- Assaf Pinchas
- School of Electrical Engineering, The Iby and Aladar Fleischman Faculty of Engineering, Tel Aviv University, Tel Aviv 6997801, Israel
| | - Irad Ben-Gal
- Industrial Engineering Department, The Iby and Aladar Fleischman Faculty of Engineering, Tel Aviv University, Tel Aviv 6997801, Israel; (I.B.-G.); (A.P.)
| | - Amichai Painsky
- Industrial Engineering Department, The Iby and Aladar Fleischman Faculty of Engineering, Tel Aviv University, Tel Aviv 6997801, Israel; (I.B.-G.); (A.P.)
| |
Collapse
|
2
|
Camaglia F, Nemenman I, Mora T, Walczak AM. Bayesian estimation of the Kullback-Leibler divergence for categorical systems using mixtures of Dirichlet priors. Phys Rev E 2024; 109:024305. [PMID: 38491647 DOI: 10.1103/physreve.109.024305] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2023] [Accepted: 01/18/2024] [Indexed: 03/18/2024]
Abstract
In many applications in biology, engineering, and economics, identifying similarities and differences between distributions of data from complex processes requires comparing finite categorical samples of discrete counts. Statistical divergences quantify the difference between two distributions. However, their estimation is very difficult and empirical methods often fail, especially when the samples are small. We develop a Bayesian estimator of the Kullback-Leibler divergence between two probability distributions that makes use of a mixture of Dirichlet priors on the distributions being compared. We study the properties of the estimator on two examples: probabilities drawn from Dirichlet distributions and random strings of letters drawn from Markov chains. We extend the approach to the squared Hellinger divergence. Both estimators outperform other estimation techniques, with better results for data with a large number of categories and for higher values of divergences.
Collapse
Affiliation(s)
- Francesco Camaglia
- Laboratoire de physique de l'École normale supérieure, CNRS, PSL University, Sorbonne Université and Université de Paris, 75005 Paris, France
| | - Ilya Nemenman
- Department of Physics, Department of Biology, and Initiative for Theory and Modeling of Living Systems, Emory University, Atlanta, Georgia 30322, USA
| | - Thierry Mora
- Laboratoire de physique de l'École normale supérieure, CNRS, PSL University, Sorbonne Université and Université de Paris, 75005 Paris, France
| | - Aleksandra M Walczak
- Laboratoire de physique de l'École normale supérieure, CNRS, PSL University, Sorbonne Université and Université de Paris, 75005 Paris, France
| |
Collapse
|
3
|
De Gregorio J, Sánchez D, Toral R. Entropy Estimators for Markovian Sequences: A Comparative Analysis. ENTROPY (BASEL, SWITZERLAND) 2024; 26:79. [PMID: 38248204 PMCID: PMC11154276 DOI: 10.3390/e26010079] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Revised: 12/21/2023] [Accepted: 01/16/2024] [Indexed: 01/23/2024]
Abstract
Entropy estimation is a fundamental problem in information theory that has applications in various fields, including physics, biology, and computer science. Estimating the entropy of discrete sequences can be challenging due to limited data and the lack of unbiased estimators. Most existing entropy estimators are designed for sequences of independent events and their performances vary depending on the system being studied and the available data size. In this work, we compare different entropy estimators and their performance when applied to Markovian sequences. Specifically, we analyze both binary Markovian sequences and Markovian systems in the undersampled regime. We calculate the bias, standard deviation, and mean squared error for some of the most widely employed estimators. We discuss the limitations of entropy estimation as a function of the transition probabilities of the Markov processes and the sample size. Overall, this paper provides a comprehensive comparison of entropy estimators and their performance in estimating entropy for systems with memory, which can be useful for researchers and practitioners in various fields.
Collapse
Affiliation(s)
| | - David Sánchez
- Institute for Cross-Disciplinary Physics and Complex Systems IFISC (UIB-CSIC), Campus Universitat de les Illes Balears, E-07122 Palma de Mallorca, Spain; (J.D.G.); (R.T.)
| | | |
Collapse
|
4
|
Hernández DG, Roman A, Nemenman I. Low-probability states, data statistics, and entropy estimation. Phys Rev E 2023; 108:014101. [PMID: 37583218 DOI: 10.1103/physreve.108.014101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2022] [Accepted: 05/30/2023] [Indexed: 08/17/2023]
Abstract
A fundamental problem in the analysis of complex systems is getting a reliable estimate of the entropy of their probability distributions over the state space. This is difficult because unsampled states can contribute substantially to the entropy, while they do not contribute to the maximum likelihood estimator of entropy, which replaces probabilities by the observed frequencies. Bayesian estimators overcome this obstacle by introducing a model of the low-probability tail of the probability distribution. Which statistical features of the observed data determine the model of the tail, and hence the output of such estimators, remains unclear. Here we show that well-known entropy estimators for probability distributions on discrete state spaces model the structure of the low-probability tail based largely on a few statistics of the data: the sample size, the maximum likelihood estimate, the number of coincidences among the samples, and the dispersion of the coincidences. We derive approximate analytical entropy estimators for undersampled distributions based on these statistics, and we use the results to propose an intuitive understanding of how the Bayesian entropy estimators work.
Collapse
Affiliation(s)
- Damián G Hernández
- Department of Physics, Emory University, Atlanta, Georgia, USA
- Department of Medical Physics, Centro Atómico Bariloche and Instituto Balseiro, 8400 San Carlos de Bariloche, Argentina
| | - Ahmed Roman
- Department of Physics, Emory University, Atlanta, Georgia, USA
| | - Ilya Nemenman
- Department of Physics, Emory University, Atlanta, Georgia, USA
- Department of Biology, Emory University, Atlanta, Georgia, USA
- Initiative for Theory and Modeling of Living Systems, Emory University, Atlanta, Georgia, USA
| |
Collapse
|
5
|
Sowinski DR, Carroll-Nellenback J, DeSilva J, Frank A, Ghoshal G, Gleiser M. The Consensus Problem in Polities of Agents with Dissimilar Cognitive Architectures. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1378. [PMID: 37420398 DOI: 10.3390/e24101378] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/15/2022] [Revised: 09/09/2022] [Accepted: 09/19/2022] [Indexed: 07/09/2023]
Abstract
Agents interacting with their environments, machine or otherwise, arrive at decisions based on their incomplete access to data and their particular cognitive architecture, including data sampling frequency and memory storage limitations. In particular, the same data streams, sampled and stored differently, may cause agents to arrive at different conclusions and to take different actions. This phenomenon has a drastic impact on polities-populations of agents predicated on the sharing of information. We show that, even under ideal conditions, polities consisting of epistemic agents with heterogeneous cognitive architectures might not achieve consensus concerning what conclusions to draw from datastreams. Transfer entropy applied to a toy model of a polity is analyzed to showcase this effect when the dynamics of the environment is known. As an illustration where the dynamics is not known, we examine empirical data streams relevant to climate and show the consensus problem manifest.
Collapse
Affiliation(s)
| | | | - Jeremy DeSilva
- Department of Anthropology, Dartmouth College, Hanover, NH 03755, USA
| | - Adam Frank
- Department of Physics and Astronomy, University of Rochester, Rochester, NY 14627, USA
| | - Gourab Ghoshal
- Department of Physics and Astronomy, University of Rochester, Rochester, NY 14627, USA
| | - Marcelo Gleiser
- Department of Physics and Astronomy, Dartmouth College, Hanover, NH 03755, USA
| |
Collapse
|
6
|
Granado M, Collavini S, Baravalle R, Martinez N, Montemurro MA, Rosso OA, Montani F. High-frequency oscillations in the ripple bands and amplitude information coding: Toward a biomarker of maximum entropy in the preictal signals. CHAOS (WOODBURY, N.Y.) 2022; 32:093151. [PMID: 36182366 DOI: 10.1063/5.0101220] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/30/2022] [Accepted: 08/29/2022] [Indexed: 06/16/2023]
Abstract
Intracranial electroencephalography (iEEG) can directly record local field potentials (LFPs) from a large set of neurons in the vicinity of the electrode. To search for possible epileptic biomarkers and to determine the epileptogenic zone that gives rise to seizures, we investigated the dynamics of basal and preictal signals. For this purpose, we explored the dynamics of the recorded time series for different frequency bands considering high-frequency oscillations (HFO) up to 240 Hz. We apply a Hilbert transform to study the amplitude and phase of the signals. The dynamics of the different frequency bands in the time causal entropy-complexity plane, H × C, is characterized by comparing the dynamical evolution of the basal and preictal time series. As the preictal states evolve closer to the time in which the epileptic seizure starts, the, H × C, dynamics changes for the higher frequency bands. The complexity evolves to very low values and the entropy becomes nearer to its maximal value. These quasi-stable states converge to equiprobable states when the entropy is maximal, and the complexity is zero. We could, therefore, speculate that in this case, it corresponds to the minimization of Gibbs free energy. In this case, the maximum entropy is equivalent to the principle of minimum consumption of resources in the system. We can interpret this as the nature of the system evolving temporally in the preictal state in such a way that the consumption of resources by the system is minimal for the amplitude in frequencies between 220-230 and 230-240 Hz.
Collapse
Affiliation(s)
- Mauro Granado
- Instituto de Física de La Plata (IFLP), Universidad Nacional de La Plata, CONICET CCT-La Plata, Diagonal 113 entre 63 y 64, La Plata 1900, Buenos Aires, Argentina
| | - Santiago Collavini
- Instituto de Electrónica Industrial, Control y Procesamiento de Se nales (LEICI), Facultad de Ingeniería, Universidad Nacional de La Plata (UNLP-CONICET), La Plata 1900, Buenos Aires, Argentina
| | - Roman Baravalle
- Instituto de Física de La Plata (IFLP), Universidad Nacional de La Plata, CONICET CCT-La Plata, Diagonal 113 entre 63 y 64, La Plata 1900, Buenos Aires, Argentina
| | - Nataniel Martinez
- Instituto de Física de Mar del Plata, Universidad Nacional de Mar del Plata & CONICET, Mar del Plata 7600, Buenos Aires, Argentina
| | - Marcelo A Montemurro
- School of Mathematics & Statistics, Faculty of Science, Technology, Engineering & Mathematics, The Open University, Walton Hall, Milton Keynes MK7 6AA, United Kingdom
| | - Osvaldo A Rosso
- Instituto de Física de La Plata (IFLP), Universidad Nacional de La Plata, CONICET CCT-La Plata, Diagonal 113 entre 63 y 64, La Plata 1900, Buenos Aires, Argentina
| | - Fernando Montani
- Instituto de Física de La Plata (IFLP), Universidad Nacional de La Plata, CONICET CCT-La Plata, Diagonal 113 entre 63 y 64, La Plata 1900, Buenos Aires, Argentina
| |
Collapse
|
7
|
Grassberger P. On Generalized Schürmann Entropy Estimators. ENTROPY 2022; 24:e24050680. [PMID: 35626564 PMCID: PMC9141067 DOI: 10.3390/e24050680] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/08/2022] [Revised: 04/26/2022] [Accepted: 05/09/2022] [Indexed: 02/04/2023]
Abstract
We present a new class of estimators of Shannon entropy for severely undersampled discrete distributions. It is based on a generalization of an estimator proposed by T. Schürmann, which itself is a generalization of an estimator proposed by myself.For a special set of parameters, they are completely free of bias and have a finite variance, something which is widely believed to be impossible. We present also detailed numerical tests, where we compare them with other recent estimators and with exact results, and point out a clash with Bayesian estimators for mutual information.
Collapse
Affiliation(s)
- Peter Grassberger
- Jülich Supercomputing Center, Jülich Research Center, D-52425 Jülich, Germany
| |
Collapse
|
8
|
Gökmen DE, Ringel Z, Huber SD, Koch-Janusz M. Symmetries and phase diagrams with real-space mutual information neural estimation. Phys Rev E 2022; 104:064106. [PMID: 35030903 DOI: 10.1103/physreve.104.064106] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Accepted: 10/05/2021] [Indexed: 11/07/2022]
Abstract
Real-space mutual information (RSMI) was shown to be an important quantity, formally and from a numerical standpoint, in finding coarse-grained descriptions of physical systems. It very generally quantifies spatial correlations and can give rise to constructive algorithms extracting relevant degrees of freedom. Efficient and reliable estimation or maximization of RSMI is, however, numerically challenging. A recent breakthrough in theoretical machine learning has been the introduction of variational lower bounds for mutual information, parametrized by neural networks. Here we describe in detail how these results can be combined with differentiable coarse-graining operations to develop a single unsupervised neural-network-based algorithm, the RSMI-NE, efficiently extracting the relevant degrees of freedom in the form of the operators of effective field theories, directly from real-space configurations. We study the information contained in the statistical ensemble of constructed coarse-graining transformations and its recovery from partial input data using a secondary machine learning analysis applied to this ensemble. In particular, we show how symmetries, also emergent, can be identified. We demonstrate the extraction of the phase diagram and the order parameters for equilibrium systems and consider also an example of a nonequilibrium problem.
Collapse
Affiliation(s)
- Doruk Efe Gökmen
- Institute for Theoretical Physics, ETH Zurich, 8093 Zurich, Switzerland
| | - Zohar Ringel
- Racah Institute of Physics, The Hebrew University of Jerusalem, Jerusalem 9190401, Israel
| | - Sebastian D Huber
- Institute for Theoretical Physics, ETH Zurich, 8093 Zurich, Switzerland
| | - Maciej Koch-Janusz
- Institute for Theoretical Physics, ETH Zurich, 8093 Zurich, Switzerland.,Department of Physics, University of Zurich, 8057 Zurich, Switzerland.,James Franck Institute, The University of Chicago, Chicago, Illinois 60637, USA
| |
Collapse
|
9
|
Nayeem R, Bazzi S, Sadeghi M, Hogan N, Sternad D. Preparing to move: Setting initial conditions to simplify interactions with complex objects. PLoS Comput Biol 2021; 17:e1009597. [PMID: 34919539 PMCID: PMC8683040 DOI: 10.1371/journal.pcbi.1009597] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2021] [Accepted: 10/28/2021] [Indexed: 12/15/2022] Open
Abstract
Humans dexterously interact with a variety of objects, including those with complex internal dynamics. Even in the simple action of carrying a cup of coffee, the hand not only applies a force to the cup, but also indirectly to the liquid, which elicits complex reaction forces back on the hand. Due to underactuation and nonlinearity, the object's dynamic response to an action sensitively depends on its initial state and can display unpredictable, even chaotic behavior. With the overarching hypothesis that subjects strive for predictable object-hand interactions, this study examined how subjects explored and prepared the dynamics of an object for subsequent execution of the target task. We specifically hypothesized that subjects find initial conditions that shorten the transients prior to reaching a stable and predictable steady state. Reaching a predictable steady state is desirable as it may reduce the need for online error corrections and facilitate feed forward control. Alternative hypotheses were that subjects seek to reduce effort, increase smoothness, and reduce risk of failure. Motivated by the task of 'carrying a cup of coffee', a simplified cup-and-ball model was implemented in a virtual environment. Human subjects interacted with this virtual object via a robotic manipulandum that provided force feedback. Subjects were encouraged to first explore and prepare the cup-and-ball before initiating a rhythmic movement at a specified frequency between two targets without losing the ball. Consistent with the hypotheses, subjects increased the predictability of interaction forces between hand and object and converged to a set of initial conditions followed by significantly decreased transients. The three alternative hypotheses were not supported. Surprisingly, the subjects' strategy was more effortful and less smooth, unlike the observed behavior in simple reaching movements. Inverse dynamics of the cup-and-ball system and forward simulations with an impedance controller successfully described subjects' behavior. The initial conditions chosen by the subjects in the experiment matched those that produced the most predictable interactions in simulation. These results present first support for the hypothesis that humans prepare the object to minimize transients and increase stability and, overall, the predictability of hand-object interactions.
Collapse
Affiliation(s)
- Rashida Nayeem
- Department of Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States of America
| | - Salah Bazzi
- Department of Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States of America
- Department of Biology, Northeastern University, Boston, Massachusetts, United States of America
- Institute for Experiential Robotics, Northeastern University, Boston, Massachusetts, United States of America
| | - Mohsen Sadeghi
- Department of Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States of America
- Department of Biology, Northeastern University, Boston, Massachusetts, United States of America
| | - Neville Hogan
- Departments of Mechanical Engineering and Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts, United States of America
| | - Dagmar Sternad
- Department of Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States of America
- Department of Biology, Northeastern University, Boston, Massachusetts, United States of America
- Institute for Experiential Robotics, Northeastern University, Boston, Massachusetts, United States of America
- Department of Physics, Northeastern University, Boston, Massachusetts, United States of America
| |
Collapse
|
10
|
Jing Y, Widmer P, Bickel B. Word Order Variation is Partially Constrained by Syntactic Complexity. Cogn Sci 2021; 45:e13056. [PMID: 34758151 PMCID: PMC9287024 DOI: 10.1111/cogs.13056] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2020] [Revised: 07/21/2021] [Accepted: 09/14/2021] [Indexed: 12/02/2022]
Abstract
Previous work suggests that when speakers linearize syntactic structures, they place longer and more complex dependents further away from the head word to which they belong than shorter and simpler dependents, and that they do so with increasing rigidity the longer expressions get, for example, longer objects tend to be placed further away from their verb, and with less variation. Current theories of sentence processing furthermore make competing predictions on whether longer expressions are preferentially placed as early or as late as possible. Here we test these predictions using hierarchical distributional regression models that allow estimates of word order and word order variation at the level of individual dependencies in corpora from 71 languages, while controlling for confounding effects from the type of dependency (e.g., subject vs. object), and the type of clause (main vs. subordinate) involved as well as from trends that are characteristic of individual languages, language families, and language contact areas. Our results show the expected correlations of length with position and variation only for two out of six dependency types (obliques and nominal modifiers) and no difference between clause types. These findings challenge received theories of across‐the‐board effects of complexity on word order and word order variation and call for theoretical models that relativize effects to specific kinds of syntactic structures and dependencies.
Collapse
Affiliation(s)
- Yingqi Jing
- Department of Comparative Language Science, University of Zurich.,Center for the Interdisciplinary Study of Language Evolution, University of Zurich.,Department of Linguistics and Philology, Uppsala University
| | - Paul Widmer
- Department of Comparative Language Science, University of Zurich.,Center for the Interdisciplinary Study of Language Evolution, University of Zurich
| | - Balthasar Bickel
- Department of Comparative Language Science, University of Zurich.,Center for the Interdisciplinary Study of Language Evolution, University of Zurich
| |
Collapse
|
11
|
Entropy Estimation Using a Linguistic Zipf-Mandelbrot-Li Model for Natural Sequences. ENTROPY 2021; 23:e23091100. [PMID: 34573725 PMCID: PMC8468050 DOI: 10.3390/e23091100] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/15/2021] [Revised: 08/14/2021] [Accepted: 08/19/2021] [Indexed: 11/17/2022]
Abstract
Entropy estimation faces numerous challenges when applied to various real-world problems. Our interest is in divergence and entropy estimation algorithms which are capable of rapid estimation for natural sequence data such as human and synthetic languages. This typically requires a large amount of data; however, we propose a new approach which is based on a new rank-based analytic Zipf–Mandelbrot–Li probabilistic model. Unlike previous approaches, which do not consider the nature of the probability distribution in relation to language; here, we introduce a novel analytic Zipfian model which includes linguistic constraints. This provides more accurate distributions for natural sequences such as natural or synthetic emergent languages. Results are given which indicates the performance of the proposed ZML model. We derive an entropy estimation method which incorporates the linguistic constraint-based Zipf–Mandelbrot–Li into a new non-equiprobable coincidence counting algorithm which is shown to be effective for tasks such as entropy rate estimation with limited data.
Collapse
|
12
|
Alternative Dirichlet Priors for Estimating Entropy via a Power Sum Functional. MATHEMATICS 2021. [DOI: 10.3390/math9131493] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Entropy is a functional of probability and is a measurement of information contained in a system; however, the practical problem of estimating entropy in applied settings remains a challenging and relevant problem. The Dirichlet prior is a popular choice in the Bayesian framework for estimation of entropy when considering a multinomial likelihood. In this work, previously unconsidered Dirichlet type priors are introduced and studied. These priors include a class of Dirichlet generators as well as a noncentral Dirichlet construction, and in both cases includes the usual Dirichlet as a special case. These considerations allow for flexible behaviour and can account for negative and positive correlation. Resultant estimators for a particular functional, the power sum, under these priors and assuming squared error loss, are derived and represented in terms of the product moments of the posterior. This representation facilitates closed-form estimators for the Tsallis entropy, and thus expedite computations of this generalised Shannon form. Select cases of these proposed priors are considered to investigate the impact and effect on the estimation of Tsallis entropy subject to different parameter scenarios.
Collapse
|
13
|
Rudelt L, González Marx D, Wibral M, Priesemann V. Embedding optimization reveals long-lasting history dependence in neural spiking activity. PLoS Comput Biol 2021; 17:e1008927. [PMID: 34061837 PMCID: PMC8205186 DOI: 10.1371/journal.pcbi.1008927] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2020] [Revised: 06/15/2021] [Accepted: 03/31/2021] [Indexed: 11/19/2022] Open
Abstract
Information processing can leave distinct footprints on the statistics of neural spiking. For example, efficient coding minimizes the statistical dependencies on the spiking history, while temporal integration of information may require the maintenance of information over different timescales. To investigate these footprints, we developed a novel approach to quantify history dependence within the spiking of a single neuron, using the mutual information between the entire past and current spiking. This measure captures how much past information is necessary to predict current spiking. In contrast, classical time-lagged measures of temporal dependence like the autocorrelation capture how long-potentially redundant-past information can still be read out. Strikingly, we find for model neurons that our method disentangles the strength and timescale of history dependence, whereas the two are mixed in classical approaches. When applying the method to experimental data, which are necessarily of limited size, a reliable estimation of mutual information is only possible for a coarse temporal binning of past spiking, a so-called past embedding. To still account for the vastly different spiking statistics and potentially long history dependence of living neurons, we developed an embedding-optimization approach that does not only vary the number and size, but also an exponential stretching of past bins. For extra-cellular spike recordings, we found that the strength and timescale of history dependence indeed can vary independently across experimental preparations. While hippocampus indicated strong and long history dependence, in visual cortex it was weak and short, while in vitro the history dependence was strong but short. This work enables an information-theoretic characterization of history dependence in recorded spike trains, which captures a footprint of information processing that is beyond time-lagged measures of temporal dependence. To facilitate the application of the method, we provide practical guidelines and a toolbox.
Collapse
Affiliation(s)
- Lucas Rudelt
- Max Planck Institute for Dynamics and Self-Organization, Göttingen, Germany
| | | | - Michael Wibral
- Campus Institute for Dynamics of Biological Networks, University of Göttingen, Göttingen, Germany
| | - Viola Priesemann
- Max Planck Institute for Dynamics and Self-Organization, Göttingen, Germany
- Bernstein Center for Computational Neuroscience, Göttingen, Germany
| |
Collapse
|
14
|
Contreras Rodríguez L, Madarro-Capó EJ, Legón-Pérez CM, Rojas O, Sosa-Gómez G. Selecting an Effective Entropy Estimator for Short Sequences of Bits and Bytes with Maximum Entropy. ENTROPY (BASEL, SWITZERLAND) 2021; 23:561. [PMID: 33946438 PMCID: PMC8147137 DOI: 10.3390/e23050561] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/05/2021] [Revised: 04/26/2021] [Accepted: 04/28/2021] [Indexed: 11/22/2022]
Abstract
Entropy makes it possible to measure the uncertainty about an information source from the distribution of its output symbols. It is known that the maximum Shannon's entropy of a discrete source of information is reached when its symbols follow a Uniform distribution. In cryptography, these sources have great applications since they allow for the highest security standards to be reached. In this work, the most effective estimator is selected to estimate entropy in short samples of bytes and bits with maximum entropy. For this, 18 estimators were compared. Results concerning the comparisons published in the literature between these estimators are discussed. The most suitable estimator is determined experimentally, based on its bias, the mean square error short samples of bytes and bits.
Collapse
Affiliation(s)
- Lianet Contreras Rodríguez
- Facultad de Matemática y Computación, Instituto de Criptografía, Universidad de la Habana, Habana 10400, Cuba; (L.C.R.); (E.J.M.-C.); (C.M.L.-P.)
| | - Evaristo José Madarro-Capó
- Facultad de Matemática y Computación, Instituto de Criptografía, Universidad de la Habana, Habana 10400, Cuba; (L.C.R.); (E.J.M.-C.); (C.M.L.-P.)
| | - Carlos Miguel Legón-Pérez
- Facultad de Matemática y Computación, Instituto de Criptografía, Universidad de la Habana, Habana 10400, Cuba; (L.C.R.); (E.J.M.-C.); (C.M.L.-P.)
| | - Omar Rojas
- Facultad de Ciencias Económicas y Empresariales, Universidad Panamericana, Álvaro del Portillo 49, Zapopan, Jalisco 45010, Mexico;
| | - Guillermo Sosa-Gómez
- Facultad de Ciencias Económicas y Empresariales, Universidad Panamericana, Álvaro del Portillo 49, Zapopan, Jalisco 45010, Mexico;
| |
Collapse
|
15
|
Smiljanić J, Edler D, Rosvall M. Mapping flows on sparse networks with missing links. Phys Rev E 2020; 102:012302. [PMID: 32794952 DOI: 10.1103/physreve.102.012302] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2019] [Accepted: 06/09/2020] [Indexed: 11/07/2022]
Abstract
Unreliable network data can cause community-detection methods to overfit and highlight spurious structures with misleading information about the organization and function of complex systems. Here we show how to detect significant flow-based communities in sparse networks with missing links using the map equation. Since the map equation builds on Shannon entropy estimation, it assumes complete data such that analyzing undersampled networks can lead to overfitting. To overcome this problem, we incorporate a Bayesian approach with assumptions about network uncertainties into the map equation framework. Results in both synthetic and real-world networks show that the Bayesian estimate of the map equation provides a principled approach to revealing significant structures in undersampled networks.
Collapse
Affiliation(s)
- Jelena Smiljanić
- Integrated Science Lab, Department of Physics, Umeå University, SE-901 87 Umeå, Sweden.,Scientific Computing Laboratory, Center for the Study of Complex Systems, Institute of Physics Belgrade, University of Belgrade, Pregrevica 118, 11080 Belgrade, Serbia
| | - Daniel Edler
- Integrated Science Lab, Department of Physics, Umeå University, SE-901 87 Umeå, Sweden.,Gothenburg Global Biodiversity Centre, Box 461, SE-405 30 Gothenburg, Sweden.,Department of Biological and Environmental Sciences, University of Gothenburg, Carl Skottsbergs gata 22B, Gothenburg 41319, Sweden
| | - Martin Rosvall
- Integrated Science Lab, Department of Physics, Umeå University, SE-901 87 Umeå, Sweden
| |
Collapse
|
16
|
Verdú S. Empirical Estimation of Information Measures: A Literature Guide. ENTROPY 2019; 21:e21080720. [PMID: 33267434 PMCID: PMC7515235 DOI: 10.3390/e21080720] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/16/2019] [Revised: 06/10/2019] [Accepted: 06/11/2019] [Indexed: 11/23/2022]
Abstract
We give a brief survey of the literature on the empirical estimation of entropy, differential entropy, relative entropy, mutual information and related information measures. While those quantities are of central importance in information theory, universal algorithms for their estimation are increasingly important in data science, machine learning, biology, neuroscience, economics, language, and other experimental sciences.
Collapse
Affiliation(s)
- Sergio Verdú
- Independent Researcher, Princeton, NJ 08540, USA
| |
Collapse
|
17
|
van Dam A. Diversity and its decomposition into variety, balance and disparity. ROYAL SOCIETY OPEN SCIENCE 2019; 6:190452. [PMID: 31417744 PMCID: PMC6689592 DOI: 10.1098/rsos.190452] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/12/2019] [Accepted: 06/24/2019] [Indexed: 05/29/2023]
Abstract
Diversity is a central concept in many fields. Despite its importance, there is no unified methodological framework to measure diversity and its three components of variety, balance and disparity. Current approaches take into account disparity of the types by considering their pairwise similarities. Pairwise similarities between types may not adequately capture total disparity, since they do not take into account in which way pairs are similar. Hence, pairwise similarities do not discriminate between similarities of types in terms of the same feature and similarities in which all pairs share different features. This paper presents an alternative approach which is based on the overlap of features over the whole set of types. This results in a measure of diversity that takes into account the aspects of variety, balance and disparity. Based on this measure, the 'ABC decomposition' is introduced, which provides separate measures for the variety, balance and disparity, allowing them to enter analysis separately. The method is illustrated by analysing the industrial diversity from 1850 to present while taking into account the overlap in occupations they employ. Finally, the framework is extended to take into account disparity considering multiple features, providing a helpful tool in analysis of high-dimensional data.
Collapse
|
18
|
Hernández DG, Samengo I. Estimating the Mutual Information between Two Discrete, Asymmetric Variables with Limited Samples. ENTROPY 2019; 21:e21060623. [PMID: 33267337 PMCID: PMC7515115 DOI: 10.3390/e21060623] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/03/2019] [Revised: 06/11/2019] [Accepted: 06/13/2019] [Indexed: 11/27/2022]
Abstract
Determining the strength of nonlinear, statistical dependencies between two variables is a crucial matter in many research fields. The established measure for quantifying such relations is the mutual information. However, estimating mutual information from limited samples is a challenging task. Since the mutual information is the difference of two entropies, the existing Bayesian estimators of entropy may be used to estimate information. This procedure, however, is still biased in the severely under-sampled regime. Here, we propose an alternative estimator that is applicable to those cases in which the marginal distribution of one of the two variables—the one with minimal entropy—is well sampled. The other variable, as well as the joint and conditional distributions, can be severely undersampled. We obtain a consistent estimator that presents very low bias, outperforming previous methods even when the sampled data contain few coincidences. As with other Bayesian estimators, our proposal focuses on the strength of the interaction between the two variables, without seeking to model the specific way in which they are related. A distinctive property of our method is that the main data statistics determining the amount of mutual information is the inhomogeneity of the conditional distribution of the low-entropy variable in those states in which the large-entropy variable registers coincidences.
Collapse
|
19
|
Mahmud M, Vassanelli S. Open-Source Tools for Processing and Analysis of In Vitro Extracellular Neuronal Signals. ADVANCES IN NEUROBIOLOGY 2019; 22:233-250. [PMID: 31073939 DOI: 10.1007/978-3-030-11135-9_10] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/03/2022]
Abstract
The recent years have seen unprecedented growth in the manufacturing of neurotechnological tools. The latest technological advancements presented the neuroscientific community with neuronal probes containing thousands of recording sites. These next-generation probes are capable of simultaneously recording neuronal signals from a large number of channels. Numerically, a simple 128-channel neuronal data acquisition system equipped with a 16 bits A/D converter digitizing the acquired analog waveforms at a sampling frequency of 20 kHz will generate approximately 17 GB uncompressed data per hour. Today's biggest challenge is to mine this staggering amount of data and find useful information which can later be used in decoding brain functions, diagnosing diseases, and devising treatments. To this goal, many automated processing and analysis tools have been developed and reported in the literature. A good amount of them are also available as open source for others to adapt them to individual needs. Focusing on extracellularly recorded neuronal signals in vitro, this chapter provides an overview of the popular open-source tools applicable on these signals for spike trains and local field potentials analysis, and spike sorting. Towards the end, several future research directions have also been outlined.
Collapse
Affiliation(s)
- Mufti Mahmud
- Computing and Technology, School of Science and Technology, Nottingham Trent University, Nottingham, UK.
| | - Stefano Vassanelli
- NeuroChip Lab, Department of Biomedical Sciences, University of Padova, Padova, Italy
| |
Collapse
|
20
|
Assessing the Relevance of Specific Response Features in the Neural Code. ENTROPY 2018; 20:e20110879. [PMID: 33266602 PMCID: PMC7512461 DOI: 10.3390/e20110879] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/01/2018] [Revised: 11/12/2018] [Accepted: 11/13/2018] [Indexed: 11/27/2022]
Abstract
The study of the neural code aims at deciphering how the nervous system maps external stimuli into neural activity—the encoding phase—and subsequently transforms such activity into adequate responses to the original stimuli—the decoding phase. Several information-theoretical methods have been proposed to assess the relevance of individual response features, as for example, the spike count of a given neuron, or the amount of correlation in the activity of two cells. These methods work under the premise that the relevance of a feature is reflected in the information loss that is induced by eliminating the feature from the response. The alternative methods differ in the procedure by which the tested feature is removed, and the algorithm with which the lost information is calculated. Here we compare these methods, and show that more often than not, each method assigns a different relevance to the tested feature. We demonstrate that the differences are both quantitative and qualitative, and connect them with the method employed to remove the tested feature, as well as the procedure to calculate the lost information. By studying a collection of carefully designed examples, and working on analytic derivations, we identify the conditions under which the relevance of features diagnosed by different methods can be ranked, or sometimes even equated. The condition for equality involves both the amount and the type of information contributed by the tested feature. We conclude that the quest for relevant response features is more delicate than previously thought, and may yield to multiple answers depending on methodological subtleties.
Collapse
|
21
|
Baravalle R, Rosso OA, Montani F. Rhythmic activities of the brain: Quantifying the high complexity of beta and gamma oscillations during visuomotor tasks. CHAOS (WOODBURY, N.Y.) 2018; 28:075513. [PMID: 30070505 DOI: 10.1063/1.5025187] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/07/2018] [Accepted: 06/11/2018] [Indexed: 06/08/2023]
Abstract
Electroencephalography (EEG) signals depict the electrical activity that takes place at the surface of the brain and provide an important tool for understanding a variety of cognitive processes. The EEG is the product of synchronized activity of the brain, and variations in EEG oscillations patterns reflect the underlying changes in neuronal synchrony. Our aim is to characterize the complexity of the EEG rhythmic oscillations bands when the subjects perform a visuomotor or imagined cognitive tasks (imagined movement), providing a causal mapping of the dynamical rhythmic activities of the brain as a measure of attentional investment. We estimate the intrinsic correlational structure of the signals within the causality entropy-complexity plane H×C, where the enhanced complexity in the gamma 1, gamma 2, and beta 1 bands allows us to distinguish motor-visual memory tasks from control conditions. We identify the dynamics of the gamma 1, gamma 2, and beta 1 rhythmic oscillations within the zone of a chaotic dissipative behavior, whereas in contrast the beta 2 band shows a much higher level of entropy and a significant low level of complexity that correspond to a non-invertible cubic map. Our findings enhance the importance of the gamma band during attention in perceptual feature binding during the visuomotor/imagery tasks.
Collapse
Affiliation(s)
- Roman Baravalle
- IFLYSIB, CONICET & Universidad Nacional de La Plata, Calle 59-789, 1900 La Plata, Argentina
| | - Osvaldo A Rosso
- Departamento de Informática en Salud, Hospital Italiano de Buenos Aires & CONICET, C1199ABB Ciudad Autónoma de Buenos Aires, Argentina
| | - Fernando Montani
- IFLYSIB, CONICET & Universidad Nacional de La Plata, Calle 59-789, 1900 La Plata, Argentina
| |
Collapse
|
22
|
Timme NM, Lapish C. A Tutorial for Information Theory in Neuroscience. eNeuro 2018; 5:ENEURO.0052-18.2018. [PMID: 30211307 PMCID: PMC6131830 DOI: 10.1523/eneuro.0052-18.2018] [Citation(s) in RCA: 92] [Impact Index Per Article: 15.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2018] [Revised: 04/10/2018] [Accepted: 05/30/2018] [Indexed: 11/21/2022] Open
Abstract
Understanding how neural systems integrate, encode, and compute information is central to understanding brain function. Frequently, data from neuroscience experiments are multivariate, the interactions between the variables are nonlinear, and the landscape of hypothesized or possible interactions between variables is extremely broad. Information theory is well suited to address these types of data, as it possesses multivariate analysis tools, it can be applied to many different types of data, it can capture nonlinear interactions, and it does not require assumptions about the structure of the underlying data (i.e., it is model independent). In this article, we walk through the mathematics of information theory along with common logistical problems associated with data type, data binning, data quantity requirements, bias, and significance testing. Next, we analyze models inspired by canonical neuroscience experiments to improve understanding and demonstrate the strengths of information theory analyses. To facilitate the use of information theory analyses, and an understanding of how these analyses are implemented, we also provide a free MATLAB software package that can be applied to a wide range of data from neuroscience experiments, as well as from other fields of study.
Collapse
Affiliation(s)
- Nicholas M Timme
- Department of Psychology, Indiana University - Purdue University Indianapolis, 402 N. Blackford St, Indianapolis, IN 46202
| | - Christopher Lapish
- Department of Psychology, Indiana University - Purdue University Indianapolis, 402 N. Blackford St, Indianapolis, IN 46202
| |
Collapse
|
23
|
Dercle L, Ammari S, Bateson M, Durand PB, Haspinger E, Massard C, Jaudet C, Varga A, Deutsch E, Soria JC, Ferté C. Limits of radiomic-based entropy as a surrogate of tumor heterogeneity: ROI-area, acquisition protocol and tissue site exert substantial influence. Sci Rep 2017; 7:7952. [PMID: 28801575 PMCID: PMC5554130 DOI: 10.1038/s41598-017-08310-5] [Citation(s) in RCA: 56] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2016] [Accepted: 07/10/2017] [Indexed: 01/19/2023] Open
Abstract
Entropy is a promising quantitative imaging biomarker for characterizing cancer imaging phenotype. Entropy has been associated with tumor gene expression, tumor metabolism, tumor stage, patient prognosis, and treatment response. Our hypothesis states that tumor-specific biomarkers such as entropy should be correlated between synchronous metastases. Therefore, a significant proportion of the variance of entropy should be attributed to the malignant process. We analyzed 112 patients with matched/paired synchronous metastases (SM#1 and SM#2) prospectively enrolled in the MOSCATO-01 clinical trial. Imaging features were extracted from Regions Of Interest (ROI) delineated on CT-scan using TexRAD software. We showed that synchronous metastasis entropy was correlated across 5 Spatial Scale Filters: Spearman's Rho ranged between 0.41 and 0.59 (P = 0.0001, Bonferroni correction). Multivariate linear analysis revealed that entropy in SM#1 is significantly associated with (i) primary tumor type; (ii) entropy in SM#2 (same malignant process); (iii) ROI area size; (iv) metastasis site; and (v) entropy in the psoas muscle (reference tissue). Entropy was a logarithmic function of ROI area in normal control tissues (aorta, psoas) and in mathematical models (P < 0.01). We concluded that entropy is a tumor-specific metric only if confounding factors are corrected.
Collapse
Affiliation(s)
- Laurent Dercle
- INSERM U1015, Equipe Labellisée Ligue Nationale Contre le Cancer, Gustave Roussy Cancer Campus, Villejuif, France.
- Département de l'imagerie médicale, Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France.
- Department of Radiology, Columbia University Medical Center, New York, New York, USA.
| | - Samy Ammari
- Département de l'imagerie médicale, Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France
- Département d'Innovation Thérapeutique et des Essais Précoces (DITEP), Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France
| | | | - Paul Blanc Durand
- Département d'Innovation Thérapeutique et des Essais Précoces (DITEP), Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France
| | - Eva Haspinger
- Département d'Innovation Thérapeutique et des Essais Précoces (DITEP), Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France
| | - Christophe Massard
- Département d'Innovation Thérapeutique et des Essais Précoces (DITEP), Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France
| | - Cyril Jaudet
- Department of Radiotherapy, UZ Brussel, Brussels, Belgium
| | - Andrea Varga
- Département d'Innovation Thérapeutique et des Essais Précoces (DITEP), Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France
| | - Eric Deutsch
- Département de radiothérapie, Gustave Roussy Cancer Campus, Université Paris Saclay, F-94805, Villejuif, France
- INSERM U981, Biomarqueurs prédictifs et nouvelles stratégies en oncologie, Université Paris Sud, Gustave Roussy, Villejuif, France
| | - Jean-Charles Soria
- Département d'Innovation Thérapeutique et des Essais Précoces (DITEP), Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France
- INSERM U981, Biomarqueurs prédictifs et nouvelles stratégies en oncologie, Université Paris Sud, Gustave Roussy, Villejuif, France
- INSERM U1030, Paris Sud University, Gustave Roussy, Villejuif, France
| | - Charles Ferté
- Département d'Innovation Thérapeutique et des Essais Précoces (DITEP), Gustave Roussy, Université Paris Saclay, F-94805, Villejuif, France.
- INSERM U981, Biomarqueurs prédictifs et nouvelles stratégies en oncologie, Université Paris Sud, Gustave Roussy, Villejuif, France.
- INSERM U1030, Paris Sud University, Gustave Roussy, Villejuif, France.
| |
Collapse
|
24
|
Wollstadt P, Sellers KK, Rudelt L, Priesemann V, Hutt A, Fröhlich F, Wibral M. Breakdown of local information processing may underlie isoflurane anesthesia effects. PLoS Comput Biol 2017; 13:e1005511. [PMID: 28570661 PMCID: PMC5453425 DOI: 10.1371/journal.pcbi.1005511] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2016] [Accepted: 04/11/2017] [Indexed: 02/07/2023] Open
Abstract
The disruption of coupling between brain areas has been suggested as the mechanism underlying loss of consciousness in anesthesia. This hypothesis has been tested previously by measuring the information transfer between brain areas, and by taking reduced information transfer as a proxy for decoupling. Yet, information transfer is a function of the amount of information available in the information source—such that transfer decreases even for unchanged coupling when less source information is available. Therefore, we reconsidered past interpretations of reduced information transfer as a sign of decoupling, and asked whether impaired local information processing leads to a loss of information transfer. An important prediction of this alternative hypothesis is that changes in locally available information (signal entropy) should be at least as pronounced as changes in information transfer. We tested this prediction by recording local field potentials in two ferrets after administration of isoflurane in concentrations of 0.0%, 0.5%, and 1.0%. We found strong decreases in the source entropy under isoflurane in area V1 and the prefrontal cortex (PFC)—as predicted by our alternative hypothesis. The decrease in source entropy was stronger in PFC compared to V1. Information transfer between V1 and PFC was reduced bidirectionally, but with a stronger decrease from PFC to V1. This links the stronger decrease in information transfer to the stronger decrease in source entropy—suggesting reduced source entropy reduces information transfer. This conclusion fits the observation that the synaptic targets of isoflurane are located in local cortical circuits rather than on the synapses formed by interareal axonal projections. Thus, changes in information transfer under isoflurane seem to be a consequence of changes in local processing more than of decoupling between brain areas. We suggest that source entropy changes must be considered whenever interpreting changes in information transfer as decoupling. Currently we do not understand how anesthesia leads to loss of consciousness (LOC). One popular idea is that we loose consciousness when brain areas lose their ability to communicate with each other–as anesthetics might interrupt transmission on nerve fibers coupling them. This idea has been tested by measuring the amount of information transferred between brain areas, and taking this transfer to reflect the coupling itself. Yet, information that isn’t available in the source area can’t be transferred to a target. Hence, the decreases in information transfer could be related to less information being available in the source, rather than to a decoupling. We tested this possibility measuring the information available in source brain areas and found that it decreased under isoflurane anesthesia. In addition, a stronger decrease in source information lead to a stronger decrease of the information transfered. Thus, the input to the connection between brain areas determined the communicated information, not the strength of the coupling (which would result in a stronger decrease in the target). We suggest that interrupted information processing within brain areas has an important contribution to LOC, and should be focused on more in attempts to understand loss of consciousness under anesthesia.
Collapse
Affiliation(s)
- Patricia Wollstadt
- MEG Unit, Brain Imaging Center, Goethe University, Frankfurt/Main, Germany
- * E-mail: (PW); (VP)
| | - Kristin K. Sellers
- Department of Psychiatry, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
- Neurobiology Curriculum, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Lucas Rudelt
- Max Planck Institute for Dynamics and Self-Organization, Göttingen, Germany
| | - Viola Priesemann
- Max Planck Institute for Dynamics and Self-Organization, Göttingen, Germany
- Bernstein Center for Computational Neuroscience, BCCN, Göttingen, Germany
- * E-mail: (PW); (VP)
| | - Axel Hutt
- Deutscher Wetterdienst, Section FE 12 - Data Assimilation, Offenbach/Main, Germany
- Department of Mathematics and Statistics, University of Reading, Reading, United Kingdom
| | - Flavio Fröhlich
- Department of Psychiatry, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
- Neurobiology Curriculum, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
- Department of Cell Biology and Physiology, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
- Department of Biomedical Engineering, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
- Neuroscience Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
- Department of Neurology, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Michael Wibral
- MEG Unit, Brain Imaging Center, Goethe University, Frankfurt/Main, Germany
| |
Collapse
|
25
|
Minimum Sample Size for Reliable Causal Inference Using Transfer Entropy. ENTROPY 2017. [DOI: 10.3390/e19040150] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
26
|
Ferrari A. Modeling Information Content Via Dirichlet-Multinomial Regression Analysis. MULTIVARIATE BEHAVIORAL RESEARCH 2017; 52:259-270. [PMID: 28207283 DOI: 10.1080/00273171.2017.1279957] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Shannon entropy is being increasingly used in biomedical research as an index of complexity and information content in sequences of symbols, e.g. languages, amino acid sequences, DNA methylation patterns and animal vocalizations. Yet, distributional properties of information entropy as a random variable have seldom been the object of study, leading to researchers mainly using linear models or simulation-based analytical approach to assess differences in information content, when entropy is measured repeatedly in different experimental conditions. Here a method to perform inference on entropy in such conditions is proposed. Building on results coming from studies in the field of Bayesian entropy estimation, a symmetric Dirichlet-multinomial regression model, able to deal efficiently with the issue of mean entropy estimation, is formulated. Through a simulation study the model is shown to outperform linear modeling in a vast range of scenarios and to have promising statistical properties. As a practical example, the method is applied to a data set coming from a real experiment on animal communication.
Collapse
Affiliation(s)
- Alberto Ferrari
- a Department of Brain and Behavioural Sciences , University of Pavia
| |
Collapse
|
27
|
Abstract
For many organisms, the number of sensory neurons is largely determined during development, before strong environmental cues are present. This is despite the fact that environments can fluctuate drastically both from generation to generation and within an organism's lifetime. How can organisms get by by hard coding the number of sensory neurons? We approach this question using rate-distortion theory. A combination of simulation and theory suggests that when environments are large, the rate-distortion function-a proxy for material costs, timing delays, and energy requirements-depends only on coarse-grained environmental statistics that are expected to change on evolutionary, rather than ontogenetic, time scales.
Collapse
Affiliation(s)
- Sarah Marzen
- Department of Physics, Redwood Center for Theoretical Neuroscience, University of California at Berkeley, Berkeley, California 94720, USA.,Physics of Living Systems, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA
| | - Simon DeDeo
- Center for Complex Networks and Systems Research, Department of Informatics, Indiana University, 919 East 10th Street, Bloomington, Indiana 47408, USA.,Department of Social and Decision Sciences, 5000 Forbes Avenue, BP 208, Pittsburgh, Pennsylvania 15213, USA.,Santa Fe Institute, 1399 Hyde Park Road, Santa Fe, New Mexico 87501, USA
| |
Collapse
|
28
|
O'Neill PK, Erill I. Parametric bootstrapping for biological sequence motifs. BMC Bioinformatics 2016; 17:406. [PMID: 27716039 PMCID: PMC5052923 DOI: 10.1186/s12859-016-1246-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2016] [Accepted: 09/08/2016] [Indexed: 11/10/2022] Open
Abstract
Background Biological sequence motifs drive the specific interactions of proteins and nucleic acids. Accordingly, the effective computational discovery and analysis of such motifs is a central theme in bioinformatics. Many practical questions about the properties of motifs can be recast as random sampling problems. In this light, the task is to determine for a given motif whether a certain feature of interest is statistically unusual among relevantly similar alternatives. Despite the generality of this framework, its use has been frustrated by the difficulties of defining an appropriate reference class of motifs for comparison and of sampling from it effectively. Results We define two distributions over the space of all motifs of given dimension. The first is the maximum entropy distribution subject to mean information content, and the second is the truncated uniform distribution over all motifs having information content within a given interval. We derive exact sampling algorithms for each. As a proof of concept, we employ these sampling methods to analyze a broad collection of prokaryotic and eukaryotic transcription factor binding site motifs. In addition to positional information content, we consider the informational Gini coefficient of the motif, a measure of the degree to which information is evenly distributed throughout a motif’s positions. We find that both prokaryotic and eukaryotic motifs tend to exhibit higher informational Gini coefficients (IGC) than would be expected by chance under either reference distribution. As a second application, we apply maximum entropy sampling to the motif p-value problem and use it to give elementary derivations of two new estimators. Conclusions Despite the historical centrality of biological sequence motif analysis, this study constitutes to our knowledge the first use of principled null hypotheses for sequence motifs given information content. Through their use, we are able to characterize for the first time differerences in global motif statistics between biological motifs and their null distributions. In particular, we observe that biological sequence motifs show an unusual distribution of IGC, presumably due to biochemical constraints on the mechanisms of direct read-out. Electronic supplementary material The online version of this article (doi:10.1186/s12859-016-1246-8) contains supplementary material, which is available to authorized users.
Collapse
Affiliation(s)
- Patrick K O'Neill
- Department of Biological Sciences, University of Maryland, Baltimore County, 1000 Hilltop Circle, Baltimore, 21250, US
| | - Ivan Erill
- Department of Biological Sciences, University of Maryland, Baltimore County, 1000 Hilltop Circle, Baltimore, 21250, US.
| |
Collapse
|
29
|
Mahmud M, Vassanelli S. Processing and Analysis of Multichannel Extracellular Neuronal Signals: State-of-the-Art and Challenges. Front Neurosci 2016; 10:248. [PMID: 27313507 PMCID: PMC4889584 DOI: 10.3389/fnins.2016.00248] [Citation(s) in RCA: 32] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2016] [Accepted: 05/19/2016] [Indexed: 12/02/2022] Open
Abstract
In recent years multichannel neuronal signal acquisition systems have allowed scientists to focus on research questions which were otherwise impossible. They act as a powerful means to study brain (dys)functions in in-vivo and in in-vitro animal models. Typically, each session of electrophysiological experiments with multichannel data acquisition systems generate large amount of raw data. For example, a 128 channel signal acquisition system with 16 bits A/D conversion and 20 kHz sampling rate will generate approximately 17 GB data per hour (uncompressed). This poses an important and challenging problem of inferring conclusions from the large amounts of acquired data. Thus, automated signal processing and analysis tools are becoming a key component in neuroscience research, facilitating extraction of relevant information from neuronal recordings in a reasonable time. The purpose of this review is to introduce the reader to the current state-of-the-art of open-source packages for (semi)automated processing and analysis of multichannel extracellular neuronal signals (i.e., neuronal spikes, local field potentials, electroencephalogram, etc.), and the existing Neuroinformatics infrastructure for tool and data sharing. The review is concluded by pinpointing some major challenges that are being faced, which include the development of novel benchmarking techniques, cloud-based distributed processing and analysis tools, as well as defining novel means to share and standardize data.
Collapse
Affiliation(s)
- Mufti Mahmud
- NeuroChip Laboratory, Department of Biomedical Sciences, University of Padova Padova, Italy
| | - Stefano Vassanelli
- NeuroChip Laboratory, Department of Biomedical Sciences, University of Padova Padova, Italy
| |
Collapse
|
30
|
High Stimulus-Related Information in Barrel Cortex Inhibitory Interneurons. PLoS Comput Biol 2015; 11:e1004121. [PMID: 26098109 PMCID: PMC4476555 DOI: 10.1371/journal.pcbi.1004121] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2014] [Accepted: 01/11/2015] [Indexed: 01/28/2023] Open
Abstract
The manner in which populations of inhibitory (INH) and excitatory (EXC) neocortical neurons collectively encode stimulus-related information is a fundamental, yet still unresolved question. Here we address this question by simultaneously recording with large-scale multi-electrode arrays (of up to 128 channels) the activity of cell ensembles (of up to 74 neurons) distributed along all layers of 3–4 neighboring cortical columns in the anesthetized adult rat somatosensory barrel cortex in vivo. Using two different whisker stimulus modalities (location and frequency) we show that individual INH neurons – classified as such according to their distinct extracellular spike waveforms – discriminate better between restricted sets of stimuli (≤6 stimulus classes) than EXC neurons in granular and infra-granular layers. We also demonstrate that ensembles of INH cells jointly provide as much information about such stimuli as comparable ensembles containing the ~20% most informative EXC neurons, however presenting less information redundancy – a result which was consistent when applying both theoretical information measurements and linear discriminant analysis classifiers. These results suggest that a consortium of INH neurons dominates the information conveyed to the neocortical network, thereby efficiently processing incoming sensory activity. This conclusion extends our view on the role of the inhibitory system to orchestrate cortical activity. Perception of the environment relies on neuronal computation in the cerebral cortex. However, the exact algorithms by which cortical neuronal networks process relevant information from the inputs of sensory organs are only poorly understood. To address this problem we stimulated distinct whiskers and recorded the neuronal responses from identified cortical whisker representations of the rat using multi-site electrodes. For rodents the whisker system is one main sensory input channel, offering the unique property that for each whisker an identified cortical area ("barrel-related column") represents its main cortical input station. In the present study we were able to demonstrate that the action potential firing of single inhibitory neurons provides more information about behaviorally relevant qualities of whisker stimulation (identity of the stimulated whisker and frequency of stimulation) than excitatory neurons. In addition, information about stimulation qualities was encoded with less redundancy in inhibitory neurons. In summary, the results of our study suggest that inhibitory neurons carry substantial information about the sensory environment and can thereby adequately orchestrate neuronal activity in sensory cortices.
Collapse
|
31
|
Jiao J, Venkat K, Han Y, Weissman T. Minimax Estimation of Functionals of Discrete Distributions. IEEE TRANSACTIONS ON INFORMATION THEORY 2015; 61:2835-2885. [PMID: 29375152 PMCID: PMC5786426 DOI: 10.1109/tit.2015.2412945] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
We propose a general methodology for the construction and analysis of essentially minimax estimators for a wide class of functionals of finite dimensional parameters, and elaborate on the case of discrete distributions, where the support size S is unknown and may be comparable with or even much larger than the number of observations n. We treat the respective regions where the functional is nonsmooth and smooth separately. In the nonsmooth regime, we apply an unbiased estimator for the best polynomial approximation of the functional whereas, in the smooth regime, we apply a bias-corrected version of the maximum likelihood estimator (MLE). We illustrate the merit of this approach by thoroughly analyzing the performance of the resulting schemes for estimating two important information measures: 1) the entropy [Formula: see text] and 2) [Formula: see text], α > 0. We obtain the minimax L2 rates for estimating these functionals. In particular, we demonstrate that our estimator achieves the optimal sample complexity n ≍ S/ln S for entropy estimation. We also demonstrate that the sample complexity for estimating Fα (P), 0 < α < 1, is n ≍ S1/α /ln S, which can be achieved by our estimator but not the MLE. For 1 < α < 3/2, we show the minimax L2 rate for estimating Fα (P) is (n ln n)-2(α-1) for infinite support size, while the maximum L2 rate for the MLE is n-2(α-1). For all the above cases, the behavior of the minimax rate-optimal estimators with n samples is essentially that of the MLE (plug-in rule) with n ln n samples, which we term "effective sample size enlargement." We highlight the practical advantages of our schemes for the estimation of entropy and mutual information. We compare our performance with various existing approaches, and demonstrate that our approach reduces running time and boosts the accuracy. Moreover, we show that the minimax rate-optimal mutual information estimator yielded by our framework leads to significant performance boosts over the Chow-Liu algorithm in learning graphical models. The wide use of information measure estimation suggests that the insights and estimators obtained in this paper could be broadly applicable.
Collapse
Affiliation(s)
- Jiantao Jiao
- Department of Electrical Engineering, Stanford University, Stanford, CA 94305 USA
| | - Kartik Venkat
- Department of Electrical Engineering, Stanford University, Stanford, CA 94305 USA
| | - Yanjun Han
- Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
| | - Tsachy Weissman
- Department of Electrical Engineering, Stanford University, Stanford, CA 94305 USA
| |
Collapse
|
32
|
|
33
|
Assecondi S, Ostwald D, Bagshaw AP. Reliability of information-based integration of EEG and fMRI data: a simulation study. Neural Comput 2014; 27:281-305. [PMID: 25514112 DOI: 10.1162/neco_a_00695] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Most studies involving simultaneous electroencephalographic (EEG) and functional magnetic resonance imaging (fMRI) data rely on the first-order, affine-linear correlation of EEG and fMRI features within the framework of the general linear model. An alternative is the use of information-based measures such as mutual information and entropy, which can also detect higher-order correlations present in the data. The estimate of information-theoretic quantities might be influenced by several parameters, such as the numerosity of the sample, the amount of correlation between variables, and the discretization (or binning) strategy of choice. While these issues have been investigated for invasive neurophysiological data and a number of bias-correction estimates have been developed, there has been no attempt to systematically examine the accuracy of information estimates for the multivariate distributions arising in the context of EEG-fMRI recordings. This is especially important given the differences between electrophysiological and EEG-fMRI recordings. In this study, we drew random samples from simulated bivariate and trivariate distributions, mimicking the statistical properties of EEG-fMRI data. We compared the estimated information shared by simulated random variables with its numerical value and found that the interaction between the binning strategy and the estimation method influences the accuracy of the estimate. Conditional on the simulation assumptions, we found that the equipopulated binning strategy yields the best and most consistent results across distributions and bias correction methods. We also found that within bias correction techniques, the asymptotically debiased (TPMC), the jackknife debiased (JD), and the best upper bound (BUB) approach give similar results, and those are consistent across distributions.
Collapse
Affiliation(s)
- Sara Assecondi
- School of Psychology, University of Birmingham, Birmingham, B17 2TT, U.K.
| | | | | |
Collapse
|
34
|
Papapetrou M, Kugiumtzis D. Investigating long range correlation in DNA sequences using significance tests of conditional mutual information. Comput Biol Chem 2014; 53 Pt A:32-42. [PMID: 25205032 DOI: 10.1016/j.compbiolchem.2014.08.007] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/11/2014] [Indexed: 11/29/2022]
Abstract
This study exploits the use of Markov chain order estimation from symbol sequences of systems exhibiting long memory or long range correlations (LRC), such as DNA sequences. In the presence of limited sequence length, LRC chain can be approximated by a high order Markov chain. For the order estimation, the parametric significance test of conditional mutual information IC(m) is applied, found in an earlier work to be suitable for high order estimation. Here, it is computationally optimized applying an iterative algorithm for calculating IC(m) at increasing order m, enabling the analysis of long symbol sequences of high Markov chain order or LRC. The simulation study shows that when the true order is reasonably small, the estimated order saturates at the true order with the increase of the symbol sequence length, while when the true order is very large or the chain has LRC, the estimated order increases logarithmically with the symbol sequence length. The order estimation shows a different dependence on the DNA sequence length for bacteria, the plant Arabidopsis thaliana and the human chromosome, indicating a different long memory structure in their DNA.
Collapse
Affiliation(s)
- Maria Papapetrou
- Department of Electrical and Computer Engineering, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece.
| | - Dimitris Kugiumtzis
- Department of Electrical and Computer Engineering, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece.
| |
Collapse
|
35
|
Strelioff CC, Crutchfield JP. Bayesian structural inference for hidden processes. PHYSICAL REVIEW. E, STATISTICAL, NONLINEAR, AND SOFT MATTER PHYSICS 2014; 89:042119. [PMID: 24827205 DOI: 10.1103/physreve.89.042119] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/10/2013] [Indexed: 06/03/2023]
Abstract
We introduce a Bayesian approach to discovering patterns in structurally complex processes. The proposed method of Bayesian structural inference (BSI) relies on a set of candidate unifilar hidden Markov model (uHMM) topologies for inference of process structure from a data series. We employ a recently developed exact enumeration of topological ε-machines. (A sequel then removes the topological restriction.) This subset of the uHMM topologies has the added benefit that inferred models are guaranteed to be ε-machines, irrespective of estimated transition probabilities. Properties of ε-machines and uHMMs allow for the derivation of analytic expressions for estimating transition probabilities, inferring start states, and comparing the posterior probability of candidate model topologies, despite process internal structure being only indirectly present in data. We demonstrate BSI's effectiveness in estimating a process's randomness, as reflected by the Shannon entropy rate, and its structure, as quantified by the statistical complexity. We also compare using the posterior distribution over candidate models and the single, maximum a posteriori model for point estimation and show that the former more accurately reflects uncertainty in estimated values. We apply BSI to in-class examples of finite- and infinite-order Markov processes, as well to an out-of-class, infinite-state hidden process.
Collapse
Affiliation(s)
- Christopher C Strelioff
- Complexity Sciences Center and Physics Department, University of California at Davis, One Shields Avenue, Davis, California 95616, USA
| | - James P Crutchfield
- Complexity Sciences Center and Physics Department, University of California at Davis, One Shields Avenue, Davis, California 95616, USA and Santa Fe Institute, 1399 Hyde Park Road, Santa Fe, New Mexico 87501, USA
| |
Collapse
|
36
|
Estimating Functions of Distributions Defined over Spaces of Unknown Size. ENTROPY 2013. [DOI: 10.3390/e15114668] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
37
|
Bootstrap Methods for the Empirical Study of Decision-Making and Information Flows in Social Systems. ENTROPY 2013. [DOI: 10.3390/e15062246] [Citation(s) in RCA: 30] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
38
|
Yousefi S, Kehtarnavaz N, Cao Y. Computationally tractable stochastic image modeling based on symmetric Markov mesh random fields. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2013; 22:2192-2206. [PMID: 23412614 DOI: 10.1109/tip.2013.2246516] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
In this paper, the properties of a new class of causal Markov random fields, named symmetric Markov mesh random field, are initially discussed. It is shown that the symmetric Markov mesh random fields from the upper corners are equivalent to the symmetric Markov mesh random fields from the lower corners. Based on this new random field, a symmetric, corner-independent, and isotropic image model is then derived which incorporates the dependency of a pixel on all its neighbors. The introduced image model comprises the product of several local 1D density and 2D joint density functions of pixels in an image thus making it computationally tractable and practically feasible by allowing the use of histogram and joint histogram approximations to estimate the model parameters. An image restoration application is also presented to confirm the effectiveness of the model developed. The experimental results demonstrate that this new model provides an improved tool for image modeling purposes compared to the conventional Markov random field models.
Collapse
Affiliation(s)
- Siamak Yousefi
- Department of Electrical Engineering, University of Texas at Dallas, Richardson, TX 75080, USA.
| | | | | |
Collapse
|
39
|
Bayesian and Quasi-Bayesian Estimators for Mutual Information from Discrete Data. ENTROPY 2013. [DOI: 10.3390/e15051738] [Citation(s) in RCA: 32] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
40
|
Abstract
Abstract
We examine the role of information-based measures in detecting and analysing phase transitions. We contend that phase transitions have a general character, visible in transitions in systems as diverse as classical flocking models, human expertise, and social networks. Information-based measures such as mutual information and transfer entropy are particularly suited to detecting the change in scale and range of coupling in systems that herald a phase transition in progress, but their use is not necessarily straightforward, possessing difficulties in accurate estimation due to limited sample sizes and the complexities of analysing non-stationary time series. These difficulties are surmountable with careful experimental choices. Their effectiveness in revealing unexpected connections between diverse systems makes them a promising tool for future research.
Collapse
|
41
|
Vinck M, Battaglia FP, Balakirsky VB, Vinck AJH, Pennartz CMA. Estimation of the entropy based on its polynomial representation. PHYSICAL REVIEW. E, STATISTICAL, NONLINEAR, AND SOFT MATTER PHYSICS 2012; 85:051139. [PMID: 23004735 DOI: 10.1103/physreve.85.051139] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/19/2012] [Indexed: 06/01/2023]
Abstract
Estimating entropy from empirical samples of finite size is of central importance for information theory as well as the analysis of complex statistical systems. Yet, this delicate task is marred by intrinsic statistical bias. Here we decompose the entropy function into a polynomial approximation function and a remainder function. The approximation function is based on a Taylor expansion of the logarithm. Given n observations, we give an unbiased, linear estimate of the first n power series terms based on counting sets of k coincidences. For the remainder function we use nonlinear Bayesian estimation with a nearly flat prior distribution on the entropy that was developed by Nemenman, Shafee, and Bialek. Our simulations show that the combined entropy estimator has reduced bias in comparison to other available estimators.
Collapse
Affiliation(s)
- Martin Vinck
- Cognitive and Systems Neuroscience Group, Center for Neuroscience, University of Amsterdam, The Netherlands
| | | | | | | | | |
Collapse
|
42
|
Coincidences and Estimation of Entropies of Random Variables with Large Cardinalities. ENTROPY 2011. [DOI: 10.3390/e13122013] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
43
|
VanderKraats ND, Banerjee A. A finite-sample, distribution-free, probabilistic lower bound on mutual information. Neural Comput 2011; 23:1862-98. [PMID: 21492010 DOI: 10.1162/neco_a_00144] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
For any memoryless communication channel with a binary-valued input and a one-dimensional real-valued output, we introduce a probabilistic lower bound on the mutual information given empirical observations on the channel. The bound is built on the Dvoretzky-Kiefer-Wolfowitz inequality and is distribution free. A quadratic time algorithm is described for computing the bound and its corresponding class-conditional distribution functions. We compare our approach to existing techniques and show the superiority of our bound to a method inspired by Fano's inequality where the continuous random variable is discretized.
Collapse
Affiliation(s)
- Nathan D VanderKraats
- Computer and Information Science and Engineering, University of Florida, Gainesville, FL 32611, USA.
| | | |
Collapse
|
44
|
Goldberg DH, Victor JD, Gardner EP, Gardner D. Spike train analysis toolkit: enabling wider application of information-theoretic techniques to neurophysiology. Neuroinformatics 2009; 7:165-78. [PMID: 19475519 DOI: 10.1007/s12021-009-9049-y] [Citation(s) in RCA: 40] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2009] [Accepted: 04/30/2009] [Indexed: 02/03/2023]
Abstract
Conventional methods widely available for the analysis of spike trains and related neural data include various time- and frequency-domain analyses, such as peri-event and interspike interval histograms, spectral measures, and probability distributions. Information theoretic methods are increasingly recognized as significant tools for the analysis of spike train data. However, developing robust implementations of these methods can be time-consuming, and determining applicability to neural recordings can require expertise. In order to facilitate more widespread adoption of these informative methods by the neuroscience community, we have developed the Spike Train Analysis Toolkit. STAToolkit is a software package which implements, documents, and guides application of several information-theoretic spike train analysis techniques, thus minimizing the effort needed to adopt and use them. This implementation behaves like a typical Matlab toolbox, but the underlying computations are coded in C for portability, optimized for efficiency, and interfaced with Matlab via the MEX framework. STAToolkit runs on any of three major platforms: Windows, Mac OS, and Linux. The toolkit reads input from files with an easy-to-generate text-based, platform-independent format. STAToolkit, including full documentation and test cases, is freely available open source via http://neuroanalysis.org , maintained as a resource for the computational neuroscience and neuroinformatics communities. Use cases drawn from somatosensory and gustatory neurophysiology, and community use of STAToolkit, demonstrate its utility and scope.
Collapse
Affiliation(s)
- David H Goldberg
- Laboratory of Neuroinformatics-D-404 and Department of Physiology, Weill Medical College of Cornell University, 1300 York Avenue, New York, NY 10065, USA
| | | | | | | |
Collapse
|
45
|
|
46
|
Dorval AD. Probability distributions of the logarithm of inter-spike intervals yield accurate entropy estimates from small datasets. J Neurosci Methods 2008; 173:129-39. [PMID: 18620755 DOI: 10.1016/j.jneumeth.2008.05.013] [Citation(s) in RCA: 46] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2007] [Revised: 05/08/2008] [Accepted: 05/09/2008] [Indexed: 11/17/2022]
Abstract
The maximal information that the spike train of any neuron can pass on to subsequent neurons can be quantified as the neuronal firing pattern entropy. Difficulties associated with estimating entropy from small datasets have proven an obstacle to the widespread reporting of firing pattern entropies and more generally, the use of information theory within the neuroscience community. In the most accessible class of entropy estimation techniques, spike trains are partitioned linearly in time and entropy is estimated from the probability distribution of firing patterns within a partition. Ample previous work has focused on various techniques to minimize the finite dataset bias and standard deviation of entropy estimates from under-sampled probability distributions on spike timing events partitioned linearly in time. In this manuscript we present evidence that all distribution-based techniques would benefit from inter-spike intervals being partitioned in logarithmic time. We show that with logarithmic partitioning, firing rate changes become independent of firing pattern entropy. We delineate the entire entropy estimation process with two example neuronal models, demonstrating the robust improvements in bias and standard deviation that the logarithmic time method yields over two widely used linearly partitioned time approaches.
Collapse
Affiliation(s)
- Alan D Dorval
- Department of Biomedical Engineering, Duke University, Durham, NC 27708, United States.
| |
Collapse
|
47
|
Zhao W, Serpedin E, Dougherty ER. Inferring connectivity of genetic regulatory networks using information-theoretic criteria. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2008; 5:262-274. [PMID: 18451435 DOI: 10.1109/tcbb.2007.1067] [Citation(s) in RCA: 39] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
Recently, the concept of mutual information has been proposed for inferring the structure of genetic regulatory networks from gene expression profiling. After analyzing the limitations of mutual information in inferring the gene-to-gene interactions, this paper introduces the concept of conditional mutual information and based on it proposes two novel algorithms to infer the connectivity structure of genetic regulatory networks. One of the proposed algorithms exhibits a better accuracy while the other algorithm excels in simplicity and flexibility. By exploiting the mutual information and conditional mutual information, a practical metric is also proposed to assess the likeliness of direct connectivity between genes. This novel metric resolves a common limitation associated with the current inference algorithms, namely the situations where the gene connectivity is established in terms of the dichotomy of being either connected or disconnected. Based on the data sets generated by synthetic networks, the performance of the proposed algorithms is compared favorably relative to existing state-of-the-art schemes. The proposed algorithms are also applied on realistic biological measurements, such as the cutaneous melanoma data set, and biological meaningful results are inferred.
Collapse
Affiliation(s)
- Wentao Zhao
- Department of Electrical and Computer Engineering, Texas A & M University, College Station, TX 77843-3128, USA.
| | | | | |
Collapse
|
48
|
Nemenman I, Lewen GD, Bialek W, de Ruyter van Steveninck RR. Neural coding of natural stimuli: information at sub-millisecond resolution. PLoS Comput Biol 2008; 4:e1000025. [PMID: 18369423 PMCID: PMC2265477 DOI: 10.1371/journal.pcbi.1000025] [Citation(s) in RCA: 78] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2007] [Accepted: 01/10/2008] [Indexed: 11/19/2022] Open
Abstract
Sensory information about the outside world is encoded by neurons in sequences of discrete, identical pulses termed action potentials or spikes. There is persistent controversy about the extent to which the precise timing of these spikes is relevant to the function of the brain. We revisit this issue, using the motion-sensitive neurons of the fly visual system as a test case. Our experimental methods allow us to deliver more nearly natural visual stimuli, comparable to those which flies encounter in free, acrobatic flight. New mathematical methods allow us to draw more reliable conclusions about the information content of neural responses even when the set of possible responses is very large. We find that significant amounts of visual information are represented by details of the spike train at millisecond and sub-millisecond precision, even though the sensory input has a correlation time of approximately 55 ms; different patterns of spike timing represent distinct motion trajectories, and the absolute timing of spikes points to particular features of these trajectories with high precision. Finally, the efficiency of our entropy estimator makes it possible to uncover features of neural coding relevant for natural visual stimuli: first, the system's information transmission rate varies with natural fluctuations in light intensity, resulting from varying cloud cover, such that marginal increases in information rate thus occur even when the individual photoreceptors are counting on the order of one million photons per second. Secondly, we see that the system exploits the relatively slow dynamics of the stimulus to remove coding redundancy and so generate a more efficient neural code.
Collapse
Affiliation(s)
- Ilya Nemenman
- Computer, Computational, and Statistical Sciences Division and Center for Nonlinear Studies, Los Alamos National Laboratory, Los Alamos, New Mexico, United States of America.
| | | | | | | |
Collapse
|
49
|
Feature selection, mutual information, and the classification of high-dimensional patterns. Pattern Anal Appl 2008. [DOI: 10.1007/s10044-008-0107-0] [Citation(s) in RCA: 52] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
50
|
Abstract
Data on 'neural coding' have frequently been analyzed using information-theoretic measures. These formulations involve the fundamental and generally difficult statistical problem of estimating entropy. We review briefly several methods that have been advanced to estimate entropy and highlight a method, the coverage-adjusted entropy estimator (CAE), due to Chao and Shen that appeared recently in the environmental statistics literature. This method begins with the elementary Horvitz-Thompson estimator, developed for sampling from a finite population, and adjusts for the potential new species that have not yet been observed in the sample-these become the new patterns or 'words' in a spike train that have not yet been observed. The adjustment is due to I. J. Good, and is called the Good-Turing coverage estimate. We provide a new empirical regularization derivation of the coverage-adjusted probability estimator, which shrinks the maximum likelihood estimate. We prove that the CAE is consistent and first-order optimal, with rate O(P)(1/log n), in the class of distributions with finite entropy variance and that, within the class of distributions with finite qth moment of the log-likelihood, the Good-Turing coverage estimate and the total probability of unobserved words converge at rate O(P)(1/(log n)(q)). We then provide a simulation study of the estimator with standard distributions and examples from neuronal data, where observations are dependent. The results show that, with a minor modification, the CAE performs much better than the MLE and is better than the best upper bound estimator, due to Paninski, when the number of possible words m is unknown or infinite.
Collapse
Affiliation(s)
- Vincent Q Vu
- Department of Statistics, University of California, Berkeley, CA 94720-3860, USA.
| | | | | |
Collapse
|