Contrast Information Dynamics: A Novel Information Measure for Cognitive Modelling
Abstract
:1. Introduction
2. Theoretical Framework
2.1. Motivation: Cognitive Modelling
2.2. Predictive Cognition
2.3. The Information Dynamics of Music and Language
2.4. Sequence Segmentation and Boundary Entropy
2.5. The Information Dynamics of Thinking
2.6. A Simple Taxonomy of Information Measures
- Global information measures characterise the expected behaviour of the process as a whole, independent of time. For instance, the entropy rate [41] (Ch.4) of a stochastic process is the mean rate of growth in entropy as the length of the sequence tends to infinity (). It provides a global measure of information production associated with the entire stochastic process.
- Process information measures summarise the expected behaviour of a stochastic process relative to specific points in time. For instance, in nonstationary processes, the value of the entropy will depend on the provided time index n. It is a process measure because it tracks the expected information produced by a stochastic process over time.
- Dynamic information measures measure the behaviour of specific realisations of a process at specific points in time. For instance, the value of the information content, , will depend not only on the time index n, but also the value that the process takes on at that time. It is a dynamic measure tracking the information provided by a specific sequence of a stochastic process over time.
2.7. Other Information Measures Related to Information Dynamics
3. Information Dynamics in Continuous Systems
3.1. Motivation
- Dynamic Measure The measure must quantify information associated with a specific observation. Since we are concerned with specific realisations of a stochastic process at specific points in time, the information measure must be a dynamic measure.
- Uniform Application The measure must be able to work for both discrete-time and continuous-time stochastic processes over both discrete and continuous state spaces in a uniform manner. When the measure is applied uniformly, we can compare different types of processes in a meaningful way.
- Coordinate Invariance The measure must be independent of the coordinate system used to describe the process. Put another way, different but equivalent descriptions of a process must provide the same amount of information.
3.2. Coordinate Invariance
3.3. A Specific Information Measure of Information Dynamics
4. Contrast Information
4.1. Defining Contrast Information
4.2. Expected Contrast Information
4.3. Relationship with Information Content and Shannon Entropy
5. Temporal Variants of Contrast Information
5.1. Temporal Regimes
5.2. Predictive Contrast Information
5.3. Connective Contrast Information
5.4. Reflective Contrast Information
5.5. Backward Temporal Variants
5.6. Terminology and Novelty
6. Contrast Information in IDyOMS
6.1. Information Dynamics of Multidimensional Sequences (IDyOMS)
6.2. Method
6.3. Results
7. Constrast Information of Some Stochastic Processes
7.1. Contrast Information of a Discrete-Time Markov Process
7.2. Contrast Information of a Continuous-Time Markov Process
7.3. Discrete-Time Gaussian Process
7.4. Continuous-Time Gaussian Process
8. Discussion
8.1. Comparison with Other Information Measures
8.2. Contributions
8.3. Future Work
- Boundary Entropy The first application of the new measures will be in replicating prior segmentation work in music. This will require the adaptation of the information profile peak-picking algorithms for use with contrast information. Subsequently, we will test the continuous measures on speech using the TIMIT dataset (https://paperswithcode.com/dataset/timit; accessed 23 July 2024).
- Continuous-state IDyOMS We will extend our IDyOMS software to allow for continuous states, thus extending its representational reach in music and other domains. This will require substituting the PPM algorithm with models of continuous feature dimensions, as well as methods for combining viewpoint models based on contrast information rather than entropy.
- Neural correlates We aim to collaborate with colleagues in neuroscience to investigate whether and how the neural correlates of perceived sound correspond with IDyOT representations, with the aim to make the system more human-like.
- Spectral Knowledge Representation We will further develop the idea of Spectral Knowledge Representation [40] to allow our system to reason using the symbols identified by segmentation.
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
IDyOT | Information Dynamics of Thinking; |
IDyOM | Information Dynamics of Music; |
IDyOMS | Information Dynamics of Multidimensional Sequences; |
PPM | Prediction by Partial Match; |
DTDS | Discrete-Time, Discrete-State (Stochastic Process); |
CTDS | Continuous-Time, Discrete-State (Stochastic Process); |
DTCS | Discrete-Time, Continuous-State (Stochastic Process); |
CTCS | Continuous-Time, Continuous-State (Stochastic Process); |
MAE | Mean Absolute Error; |
DMTC | Discrete-Time Markov Chain; |
CTMC | Continuous-Time Markov Chain; |
DTGP | Discrete-Time Gaussian Process; |
CTGP | Continuous-Time Gaussian Process. |
Appendix A. Proofs of Coordinate (In)variance
Appendix B. Proofs of Properties of Contrast Information
References
- Shannon, C. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27. 379–423, 623–656. [Google Scholar] [CrossRef]
- Pearce, M.T. The Construction and Evaluation of Statistical Models of Melodic Structure in Music Perception and Composition. Ph.D. Thesis, Department of Computing, City University, London, UK, 2005. [Google Scholar]
- Pearce, M.T.; Wiggins, G.A. Auditory Expectation: The Information Dynamics of Music Perception and Cognition. Top. Cogn. Sci. 2012, 4, 625–652. [Google Scholar] [CrossRef]
- Agres, K.; Abdallah, S.; Pearce, M. Information-Theoretic Properties of Auditory Sequences Dynamically Influence Expectation and Memory. Cogn. Sci. 2018, 42, 43–76. [Google Scholar] [CrossRef] [PubMed]
- Abdallah, S.A.; Plumbley, M.D. A Measure of Statistical Complexity Based on Predictive Information. arXiv 2010, arXiv:1012.1890. [Google Scholar]
- MacKay, D.J.C. Information Theory, Inference, and Learning Algorithms; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
- Friston, K. The free-energy principle: A unified brain theory? Nat. Rev. Neurosci. 2010, 11, 127–138. [Google Scholar] [CrossRef] [PubMed]
- Schmidhuber, J. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010). Auton. Ment. Dev. IEEE Trans. 2010, 2, 230–247. [Google Scholar] [CrossRef]
- Clark, A. Whatever next? Predictive brains, situated agents, and the future of cognitive science. Behav. Brain Sci. 2013, 36, 181–204. [Google Scholar] [CrossRef]
- Wiggins, G.A. Creativity, Information, and Consciousness: The Information Dynamics of Thinking. Phys. Life Rev. 2020, 34–35, 1–39. [Google Scholar] [CrossRef] [PubMed]
- Wiggins, G.A. Artificial Musical Intelligence: Computational creativity in a closed cognitive world. In Artificial Intelligence and the Arts: Computational Creativity in the Visual Arts, Music, 3D, Games, and Artistic Perspectives; Computational Synthesis and Creative Systems; Springer International Publishing: Heidelberg, Germany, 2021. [Google Scholar]
- Huron, D. Sweet Anticipation: Music and the Psychology of Expectation; Bradford Books; MIT Press: Cambridge, MA, USA, 2006. [Google Scholar]
- Conklin, D. Prediction and Entropy of Music. Master’s Thesis, Department of Computer Science, University of Calgary, Calgary, AB, Canada, 1990. [Google Scholar]
- Conklin, D.; Witten, I.H. Multiple Viewpoint Systems for Music Prediction. J. New Music. Res. 1995, 24, 51–73. [Google Scholar] [CrossRef]
- Pearce, M.T. Statistical learning and probabilistic prediction in music cognition: Mechanisms of stylistic enculturation. Ann. N. Y. Acad. Sci. 2018, 1423, 378–395. [Google Scholar] [CrossRef]
- Moffat, A. Implementing the PPM data compression scheme. IEEE Trans. Commun. 1990, 38, 1917–1921. [Google Scholar] [CrossRef]
- Bunton, S. Semantically Motivated Improvements for PPM Variants. Comput. J. 1997, 40, 76–93. [Google Scholar] [CrossRef]
- Pearce, M.T.; Conklin, D.; Wiggins, G.A. Methods for Combining Statistical Models of Music. In Computer Music Modelling and Retrieval; Wiil, U.K., Ed.; Springer: Heidelberg, Germany, 2005; pp. 295–312. [Google Scholar]
- Pearce, M.T.; Wiggins, G.A. Expectation in Melody: The Influence of Context and Learning. Music. Percept. 2006, 23, 377–405. [Google Scholar] [CrossRef]
- Pearce, M.T.; Herrojo Ruiz, M.; Kapasi, S.; Wiggins, G.A.; Bhattacharya, J. Unsupervised Statistical Learning Underpins Computational, Behavioural and Neural Manifestations of Musical Expectation. NeuroImage 2010, 50, 303–314. [Google Scholar] [CrossRef] [PubMed]
- Hansen, N.C.; Pearce, M.T. Predictive uncertainty in auditory sequence processing. Front. Psychol. 2014, 5. [Google Scholar] [CrossRef] [PubMed]
- Pearce, M.T.; Wiggins, G.A. Evaluating cognitive models of musical composition. In Proceedings of the 4th International Joint Workshop on Computational Creativity, London, UK, 17–19 June 2007; Cardoso, A., Wiggins, G.A., Eds.; pp. 73–80. [Google Scholar]
- Dubnov, S. Deep Music Information Dynamics. arXiv 2021, arXiv:2102.01133. [Google Scholar]
- Tishby, N.; Pereira, F.C.; Bialek, W. The Information Bottleneck Method. arXiv 2000, arXiv:physics/0004057. [Google Scholar]
- Pearce, M.T.; Müllensiefen, D.; Wiggins, G.A. The role of expectation and probabilistic learning in auditory boundary perception: A model comparison. Perception 2010, 39, 1367–1391. [Google Scholar] [CrossRef] [PubMed]
- Wiggins, G.A. Cue Abstraction, Paradigmatic Analysis and Information Dynamics: Towards Music Analysis by Cognitive Model. Music. Sci. 2010, 14, 307–322. [Google Scholar] [CrossRef]
- Miller, G.A. The Magical Number Seven, Plus or Minus Two: Some Limits on Our Capacity for Processing Information. Pyschological Rev. 1956, 63, 81–97. [Google Scholar] [CrossRef]
- Wiggins, G.A. “I let the music speak”: Cross-domain application of a cognitive model of musical learning. In Statistical Learning and Language Acquisition; Rebuschat, P., Williams, J., Eds.; Mouton De Gruyter: Amsterdam, The Netherlands, 2012; pp. 463–495. [Google Scholar]
- Griffiths, S.S.; McGinity, M.M.; Forth, J.; Purver, M.; Wiggins, G.A. Information-Theoretic Segmentation of Natural Language. In Proceedings of the Third International Workshop on Artificial Intelligence and Cognition, Turin, Italy, 28–29 September 2015. [Google Scholar]
- Tan, N.; Aiello, R.; Bever, T.G. Harmonic structure as a determinant of melodic organization. Mem. Cogn. 1981, 9, 533–539. [Google Scholar] [CrossRef]
- Chiappe, P.; Schmuckler, M.A. Phrasing influences the recognition of melodies. Psychon. Bull. Rev. 1997, 4, 254–259. [Google Scholar] [CrossRef] [PubMed]
- Wiggins, G.A.; Sanjekdar, A. Learning and consolidation as re-representation: Revising the meaning of memory. Front. Psychol. Cogn. Sci. 2019, 10, 802. [Google Scholar] [CrossRef] [PubMed]
- Agres, K.; Forth, J.; Wiggins, G.A. Evaluation of Musical Creativity and Musical Metacreation Systems. ACM Comput. Entertain. 2016, 14, 3:1–3:33. [Google Scholar] [CrossRef]
- Large, E.W. A generic nonlinear model for auditory perception. In Auditory Mechanisms: Processes and Models; Nuttall, A.L., Ren, T., Gillespie, P., Grosh, K., de Boer, E., Eds.; World Scientific: Singapore, 2006; pp. 516–517. [Google Scholar]
- Spivey, M. The Continuity of Mind; Oxford University Press: Oxford, UK, 2008. [Google Scholar]
- Forth, J.; Agres, K.; Purver, M.; Wiggins, G.A. Entraining IDyOT: Timing in the information dynamics of thinking. Front. Psychol. 2016, 7, 1575. [Google Scholar] [CrossRef] [PubMed]
- Kraus, N.; Nicol, T. Brainstem Encoding of Speech and Music Sounds in Humans. In The Oxford Handbook of the Auditory Brainstem; Oxford University Press: Oxford, UK, 2019; pp. 741–758. [Google Scholar] [CrossRef]
- Bellier, L.; Llorens, A.; Marciano, D.; Gunduz, A.; Schalk, G.; Brunner, P.; Knight, R.T. Music can be reconstructed from human auditory cortex activity using nonlinear decoding models. PLoS Biol. 2023, 21, e3002176. [Google Scholar] [CrossRef] [PubMed]
- Pasley, B.N.; David, S.V.; Mesgarani, N.; Flinker, A.; Shamma, S.A.; Crone, N.E.; Knight, R.T.; Chang, E.F. Reconstructing Speech from Human Auditory Cortex. PLoS Biol. 2023, 10, e1001251. [Google Scholar] [CrossRef] [PubMed]
- Homer, S.T.; Harley, N.; Wiggins, G.A. The Discrete Resonance Spectrogram: A novel method for precise determination of spectral content. in preparation.
- Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
- Ebrahimi, N.; Kirmani, S.; Soofi, E.S. Multivariate Dynamic Information. J. Multivar. Anal. 2007, 98, 328–349. [Google Scholar] [CrossRef]
- Pinto, T.; Morais, H.; Corchado, J.M. Adaptive Entropy-Based Learning with Dynamic Artificial Neural Network. Neurocomputing 2019, 338, 432–440. [Google Scholar] [CrossRef]
- Sinai, Y. On the Notion of Entropy of a Dynamical System. Dokl. Russ. Acad. Sci. 1959, 124, 768–771. [Google Scholar]
- Layek, G. An Introduction to Dynamical Systems and Chaos; Springer: New Delhi, India, 2015. [Google Scholar] [CrossRef]
- Brockwell, P.J.; Davis, R.A. Time Series: Theory and Methods; Springer Series in Statistics; Springer: New York, NY, USA, 1991. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762 [cs]. [Google Scholar]
- Gu, A.; Dao, T. Mamba: Linear-Time Sequence Modeling with Selective State Spaces. arXiv 2024, arXiv:2312.00752. [Google Scholar]
- Delgado-Bonal, A.; Marshak, A. Approximate Entropy and Sample Entropy: A Comprehensive Tutorial. Entropy 2019, 21, 541. [Google Scholar] [CrossRef]
- Pincus, S. Approximate Entropy (ApEn) as a Complexity Measure. Chaos Interdiscip. J. Nonlinear Sci. 1995, 5, 110–117. [Google Scholar] [CrossRef] [PubMed]
- Richman, J.S.; Moorman, J.R. Physiological Time-Series Analysis Using Approximate Entropy and Sample Entropy. Am. J. Physiol.-Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed]
- Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale Entropy Analysis of Biological Signals. Phys. Rev. E 2005, 71, 021906. [Google Scholar] [CrossRef]
- Dubnov, S. Spectral Anticipations. Comput. Music. J. 2006, 30, 63–83. [Google Scholar] [CrossRef]
- Bialek, W.; Tishby, N. Predictive Information. arXiv 1999, arXiv:cond-mat/9902341. [Google Scholar]
- Abdallah, S.A.; Plumbley, M.D. A Measure of Statistical Complexity Based on Predictive Information with Application to Finite Spin Systems. Phys. Lett. A 2012, 376, 275–281. [Google Scholar] [CrossRef]
- Abdallah, S.; Plumbley, M. Information Dynamics: Patterns of Expectation and Surprise in the Perception of Music. Connect. Sci. 2009, 21, 89–117. [Google Scholar] [CrossRef]
- Caticha, A. Relative Entropy and Inductive Inference. AIP Conf. Proc. 2004, 707, 75–96. [Google Scholar] [CrossRef]
- Amari, S.I. Information Geometry and Its Applications; Springer: Tokyo, Japan, 2016; Volume 194. [Google Scholar] [CrossRef]
- Eckhorn, R.; Pöpel, B. Rigorous and Extended Application of Information Theory to the Afferent Visual System of the Cat. I. Basic Concepts. Kybernetik 1974, 16, 191–200. [Google Scholar] [CrossRef]
- DeWeese, M.R.; Meister, M. How to Measure the Information Gained from One Symbol. Network Comput. Neural Syst. 1999, 10, 325. [Google Scholar] [CrossRef]
- Dubnov, S.; Assayag, G.; Cont, A. Audio Oracle Analysis of Musical Information Rate. In Proceedings of the 2011 IEEE Fifth International Conference on Semantic Computing, Palo Alto, CA, USA, 18–21 September 2011; pp. 567–571. [Google Scholar] [CrossRef]
- Good, I.J. Good Thinking: The Foundations of Probability and Its Applications; University of Minnesota Press: Minneapolis, MN, USA, 1983. [Google Scholar]
- Braverman, M.; Chen, X.; Kakade, S.; Narasimhan, K.; Zhang, C.; Zhang, Y. Calibration, Entropy Rates, and Memory in Language Models. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 13–18 July 2020; pp. 1089–1099. [Google Scholar]
- Cleary, J.; Witten, I. Data Compression Using Adaptive Coding and Partial String Matching. IEEE Trans. Commun. 1984, 32, 396–402. [Google Scholar] [CrossRef]
- Hedges, T.; Wiggins, G.A. The Prediction of Merged Attributes with Multiple Viewpoint Systems. J. New Music. Res. 2016, 45, 314–332. [Google Scholar] [CrossRef]
- Anderson, W.J. Continuous-Time Markov Chains; Springer Series in Statistics; Springer: New York, NY, USA, 1991. [Google Scholar] [CrossRef]
- Soch, J.; Maja; Monticone, P.; Faulkenberry, T.J.; Kipnis, A.; Petrykowski, K.; Allefeld, C.; Atze, H.; Knapp, A.; McInerney, C.D.; et al. StatProofBook/StatProofBook.github.io: StatProofBook 2023. 2023. Available online: https://zenodo.org/records/10495684 (accessed on 23 July 2024).
- Brockwell, P.; Davis, R.; Yang, Y. Continuous-Time Gaussian Autoregression. Stat. Sin. 2007, 17, 63–80. [Google Scholar]
- Kaiser, A.; Schreiber, T. Information Transfer in Continuous Processes. Phys. D Nonlinear Phenom. 2002, 166, 43–62. [Google Scholar] [CrossRef]
- Massey, J.L. Causality, Feedback, and Directed Information. Proc. Int. Symp. Inf. Theory Applic. (ISITA-90) 1990, 2. [Google Scholar]
- Alemi, A.A.; Fischer, I.; Dillon, J.V.; Murphy, K. Deep Variational Information Bottleneck. arXiv 2019, arXiv:1612.00410. [Google Scholar]
- Fischer, I. The Conditional Entropy Bottleneck. Entropy 2020, 22, 999. [Google Scholar] [CrossRef]
Regime | Continuous | Notation | Discrete | Notation |
---|---|---|---|---|
Past | X | X | ||
Near Past | ||||
Point Past | ||||
Present | Y | Y | ||
Point Future | ||||
Near Future | ||||
Future | Z | Z |
CPITCH | DUR | CPITCH × DUR | ||||
---|---|---|---|---|---|---|
Order | ||||||
0 | .86 | .88 | .89 | 1 | −.83 | −.82 |
1 | .85 | .86 | .72 | .51 | .34 | .38 |
2 | .76 | .77 | .70 | .66 | .45 | .47 |
3 | .73 | .76 | .70 | .69 | .23 | .28 |
4 | .75 | .81 | .71 | .72 | .19 | .33 |
5 | .77 | .86 | .73 | .78 | .23 | .39 |
6 | .78 | .88 | .74 | .80 | .32 | .45 |
7 | .80 | .90 | .75 | .82 | .45 | .50 |
8 | .82 | .91 | .75 | .84 | .58 | .54 |
9 | .84 | .92 | .76 | .86 | .66 | .57 |
10 | .85 | .92 | .77 | .87 | .69 | .60 |
CPITCH | DUR | CPITCH × DUR | ||||
---|---|---|---|---|---|---|
Order | ||||||
0 | 1 | – | 1 | – | 1 | – |
1 | .20 | −.06 | .36 | −.48 | −.28 | −.23 |
2 | .24 | .28 | .31 | −.14 | .06 | .33 |
3 | .51 | .50 | .40 | .29 | .27 | .36 |
4 | .67 | .68 | .44 | .29 | .34 | .46 |
5 | .73 | .81 | .50 | .42 | .37 | .53 |
6 | .75 | .89 | .55 | .49 | .42 | .59 |
7 | .77 | .93 | .60 | .60 | .54 | .64 |
8 | .80 | .95 | .64 | .70 | .68 | .68 |
9 | .83 | .96 | .66 | .76 | .77 | .71 |
10 | .85 | .97 | .67 | .80 | .81 | .73 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Homer, S.T.; Harley, N.; Wiggins, G.A. Contrast Information Dynamics: A Novel Information Measure for Cognitive Modelling. Entropy 2024, 26, 638. https://doi.org/10.3390/e26080638
Homer ST, Harley N, Wiggins GA. Contrast Information Dynamics: A Novel Information Measure for Cognitive Modelling. Entropy. 2024; 26(8):638. https://doi.org/10.3390/e26080638
Chicago/Turabian StyleHomer, Steven T., Nicholas Harley, and Geraint A. Wiggins. 2024. "Contrast Information Dynamics: A Novel Information Measure for Cognitive Modelling" Entropy 26, no. 8: 638. https://doi.org/10.3390/e26080638
APA StyleHomer, S. T., Harley, N., & Wiggins, G. A. (2024). Contrast Information Dynamics: A Novel Information Measure for Cognitive Modelling. Entropy, 26(8), 638. https://doi.org/10.3390/e26080638