no code implementations • 8 Nov 2023 • Kieran A. Murphy, Dani S. Bassett
Deterministic chaos permits a precise notion of a "perfect measurement" as one that, when obtained repeatedly, captures all of the information created by the system's evolution with minimal redundancy.
no code implementations • 6 Sep 2023 • Xiaohuan Xia, Andrei A. Klishin, Jennifer Stiso, Christopher W. Lynn, Ari E. Kahn, Lorenzo Caciagli, Dani S. Bassett
Using a serial response experiment with human participants (n=$100$), we replicate our predictions by detecting a surprisal effect at the finer-level of the hierarchy but not at the coarser-level of the hierarchy.
1 code implementation • 11 Jul 2023 • Shubhankar P. Patankar, Mathieu Ouellet, Juan Cervino, Alejandro Ribeiro, Kieran A. Murphy, Dani S. Bassett
The theories view curiosity as an intrinsic motivation to optimize for topological features of subgraphs induced by nodes visited in the environment.
1 code implementation • 10 Jul 2023 • Kieran A. Murphy, Dani S. Bassett
Guided by the distributed information bottleneck as a learning objective, the information decomposition identifies the variation in the measurements of the system state most relevant to specified macroscale behavior.
no code implementations • 30 Nov 2022 • Kieran A. Murphy, Dani S. Bassett
Borrowing from information theory, we use the Distributed Information Bottleneck to find optimal compressions of each feature that maximally preserve information about the output.
no code implementations • 29 Nov 2022 • Dale Zhou, Jason Z. Kim, Adam R. Pines, Valerie J. Sydnor, David R. Roalf, John A. Detre, Ruben C. Gur, Raquel E. Gur, Theodore D. Satterthwaite, Dani S. Bassett
Using a large sample of youth ($n=1, 040$), we test predictions in two ways: by measuring the dimensionality of spontaneous activity from sensorimotor to association cortex, and by assessing the representational capacity for 24 behaviors in neural circuits and 20 cognitive variables in recurrent neural networks.
1 code implementation • 25 Oct 2022 • Kieran A. Murphy, Dani S. Bassett
A hallmark of chaotic dynamics is the loss of information with time.
no code implementations • 25 Aug 2022 • Lia Papadopoulos, Demian Battaglia, Dani S. Bassett
Oscillatory synchrony is hypothesized to support the flow of information between brain regions, with different phase-locked configurations enabling activation of different effective interactions.
no code implementations • 24 Aug 2022 • Andrew C. Murphy, Romain Duprat, Theodore D. Satterthwaite, Desmond J. Oathes, Dani S. Bassett
For each individual, we measured the TMS-induced change in FC between the FPS and DMS (the FC network), and the structural coupling between the stimulated area and the FPS and DMS (the structural context network (SCN)).
no code implementations • 15 Apr 2022 • Kieran A. Murphy, Dani S. Bassett
The Distributed Information Bottleneck throttles the downstream complexity of interactions between the components of the input, deconstructing a relationship into meaningful approximations found through deep learning without requiring custom-made datasets or neural network architectures.
1 code implementation • 3 Apr 2022 • Shubhankar P. Patankar, Dale Zhou, Christopher W. Lynn, Jason Z. Kim, Mathieu Ouellet, Harang Ju, Perry Zurn, David M. Lydon-Staley, Dani S. Bassett
We formalize curiosity as the process of building a growing knowledge network to quantitatively investigate information gap theory, compression progress theory, and the conformational change theory of curiosity.
1 code implementation • 26 Oct 2021 • Leon Weninger, Pragya Srivastava, Dale Zhou, Jason Z. Kim, Eli J. Cornblath, Maxwell A. Bertolero, Ute Habel, Dorit Merhof, Dani S. Bassett
These activity patterns define global brain states and contain information in accordance with their expected probability of occurrence.
no code implementations • 16 Oct 2021 • Lindsay M. Smith, Jason Z. Kim, Zhixin Lu, Dani S. Bassett
Neural systems are well known for their ability to learn and store information as memories.