Preprint
Refine
Document Type
- Preprint (44) (remove)
Language
- English (44)
Has Fulltext
- yes (44)
Is part of the Bibliography
- no (44)
Keywords
- natural scenes (2)
- neuronal populations (2)
- primary visual cortex (2)
- stimulus encoding (2)
- visual attention (2)
- inter-individual variability (1)
- perceptual closure (1)
- precision weighting (1)
- predictive coding (1)
- schizophrenia (1)
Institute
- MPI für Hirnforschung (44) (remove)
Abstract Trial-to-trial variability and spontaneous activity of cortical recordings have been suggested to reflect intrinsic noise. This view is currently challenged by mounting evidence for structure in these phenomena: Trial-to-trial variability decreases following stimulus onset and can be predicted by previous spontaneous activity. This spontaneous activity is similar in magnitude and structure to evoked activity and can predict decisions. Allof the observed neuronal properties described above can be accounted for, at an abstract computational level, by the sampling-hypothesis, according to which response variability reflects stimulus uncertainty. However, a mechanistic explanation at the level of neural circuit dynamics is still missing.
In this study, we demonstrate that all of these phenomena can be accounted for by a noise-free self-organizing recurrent neural network model (SORN). It combines spike-timing dependent plasticity (STDP) and homeostatic mechanisms in a deterministic network of excitatory and inhibitory McCulloch-Pitts neurons. The network self-organizes to spatio-temporally varying input sequences.
We find that the key properties of neural variability mentioned above develop in this model as the network learns to perform sampling-like inference. Importantly, the model shows high trial-to-trial variability although it is fully deterministic. This suggests that the trial-to-trial variability in neural recordings may not reflect intrinsic noise. Rather, it may reflect a deterministic approximation of sampling-like learning and inference. The simplicity of the model suggests that these correlates of the sampling theory are canonical properties of recurrent networks that learn with a combination of STDP and homeostatic plasticity mechanisms.
Author Summary Neural recordings seem very noisy. If the exact same stimulus is shown to an animal multiple times, the neural response will vary. In fact, the activity of a single neuron shows many features of a stochastic process. Furthermore, in the absence of a sensory stimulus, cortical spontaneous activity has a magnitude comparable to the activity observed during stimulus presentation. These findings have led to a widespread belief that neural activity is indeed very noisy. However, recent evidence indicates that individual neurons can operate very reliably and that the spontaneous activity in the brain is highly structured, suggesting that much of the noise may in fact be signal. One hypothesis regarding this putative signal is that it reflects a form of probabilistic inference through sampling. Here we show that the key features of neural variability can be accounted for in a completely deterministic network model through self-organization. As the network learns a model of its sensory inputs, the deterministic dynamics give rise to sampling-like inference. Our findings show that the notorious variability in neural recordings does not need to be seen as evidence for a noisy brain. Instead it may reflect sampling-like inference emerging from a self-organized learning process.
Cross-frequency coupling (CFC) has been proposed to coordinate neural dynamics across spatial and temporal scales. Despite its potential relevance for understanding healthy and pathological brain function, the standard CFC analysis and physiological interpretation come with fundamental problems. For example, apparent CFC can appear because of spectral correlations due to common non-stationarities that may arise in the total absence of interactions between neural frequency components. To provide a road map towards an improved mechanistic understanding of CFC, we organize the available and potential novel statistical/modeling approaches according to their biophysical interpretability. While we do not provide solutions for all the problems described, we provide a list of practical recommendations to avoid common errors and to enhance the interpretability of CFC analysis.
How is semantic information stored in the human mind and brain? Some philosophers and cognitive scientists argue for vectorial representations of concepts, where the meaning of a word is represented as its position in a high-dimensional neural state space. At the intersection of natural language processing and artificial intelligence, a class of very successful distributional word vector models has developed that can account for classic EEG findings of language, i.e., the ease vs. difficulty of integrating a word with its sentence context. However, models of semantics have to account not only for context-based word processing, but should also describe how word meaning is represented. Here, we investigate whether distributional vector representations of word meaning can model brain activity induced by words presented without context. Using EEG activity (event-related brain potentials) collected while participants in two experiments (English, German) read isolated words, we encode and decode word vectors taken from the family of prediction-based word2vec algorithms. We find that, first, the position of a word in vector space allows the prediction of the pattern of corresponding neural activity over time, in particular during a time window of 300 to 500 ms after word onset. Second, distributional models perform better than a human-created taxonomic baseline model (WordNet), and this holds for several distinct vector-based models. Third, multiple latent semantic dimensions of word meaning can be decoded from brain activity. Combined, these results suggest that empiricist, prediction-based vectorial representations of meaning are a viable candidate for the representational architecture of human semantic knowledge.
Evading imminent predator threat is critical for survival. Effective defensive strategies can vary, even between closely related species. However, the neural basis of such species-specific behaviours is still poorly understood. Here we find that two sister species of deer mice (genus Peromyscus) show different responses to the same looming stimulus: P. maniculatus, which occupy densely vegetated habitats, predominantly dart to escape, while the open field specialist, P. polionotus, pause their movement. This difference arises from species-specific escape thresholds, is largely context-independent, and can be triggered by both visual and auditory threat stimuli. Using immunohistochemistry and electrophysiological recordings, we find that although visual threat activates the superior colliculus in both species, the role of the dorsal periaqueductal gray (dPAG) in driving behaviour differs. While dPAG activity scales with running speed and involves both excitatory and inhibitory neurons in P. maniculatus, the dPAG is largely silent in P. polionotus, even when darting is triggered. Moreover, optogenetic activation of excitatory dPAG neurons reliably elicits darting behaviour in P. maniculatus but not P. polionotus. Together, we trace the evolution of species-specific escape thresholds to a central circuit node, downstream of peripheral sensory neurons, localizing an ecologically relevant behavioural difference to a specific region of the complex mammalian brain.
To characterize the left-ventral occipito-temporal cortex (lvOT) role during reading in a quantitatively explicit and testable manner, we propose the lexical categorization model (LCM). The LCM assumes that lvOT optimizes linguistic processing by allowing fast meaning access when words are familiar and filter out orthographic strings without meaning. The LCM successfully simulates benchmark results from functional brain imaging. Empirically, using functional magnetic resonance imaging, we demonstrate that quantitative LCM simulations predict lvOT activation across three studies better than alternative models. Besides, we found that word-likeness, which is assumed as input to LCM, is represented posterior to lvOT. In contrast, a dichotomous word/non-word contrast, which is assumed as the LCM’s output, could be localized to upstream frontal brain regions. Finally, we found that training lexical categorization results in more efficient reading. Thus, we propose a ventral-visual-stream processing framework for reading involving word-likeness extraction followed by lexical categorization, before meaning extraction.
Ribosomes translate the genetic code into proteins. Recent technical advances have facilitated in situ structural analyses of ribosome functional states inside eukaryotic cells and the minimal bacterium Mycoplasma. However, such analyses of Gram-negative bacteria are lacking, despite their ribosomes being major antimicrobial drug targets. Here we compare two E. coli strains, a lab E. coli K-12 and human gut isolate E. coli ED1a, for which tetracycline exhibits bacteriostatic and bactericidal action, respectively. The in situ ribosome structures upon tetracycline treatment show a virtually identical drug binding-site in both strains, yet the distribution of ribosomal complexes clearly differs. While K-12 retains ribosomes in a translation competent state, tRNAs are lost in the vast majority of ED1a ribosomes. A differential response is also reflected in proteome-wide abundance and thermal stability assessment. Our study underlines the need to include molecular analyses and to consider gut bacteria when addressing antibiotic mode of action.
Neural computations emerge from recurrent neural circuits that comprise hundreds to a few thousand neurons. Continuous progress in connectomics, electrophysiology, and calcium imaging require tractable spiking network models that can consistently incorporate new information about the network structure and reproduce the recorded neural activity features. However, it is challenging to predict which spiking network connectivity configurations and neural properties can generate fundamental operational states and specific experimentally reported nonlinear cortical computations. Theoretical descriptions for the computational state of cortical spiking circuits are diverse, including the balanced state where excitatory and inhibitory inputs balance almost perfectly or the inhibition stabilized state (ISN) where the excitatory part of the circuit is unstable. It remains an open question whether these states can co-exist with experimentally reported nonlinear computations and whether they can be recovered in biologically realistic implementations of spiking networks. Here, we show how to identify spiking network connectivity patterns underlying diverse nonlinear computations such as XOR, bistability, inhibitory stabilization, supersaturation, and persistent activity. We established a mapping between the stabilized supralinear network (SSN) and spiking activity which allowed us to pinpoint the location in parameter space where these activity regimes occur. Notably, we found that biologically-sized spiking networks can have irregular asynchronous activity that does not require strong excitation-inhibition balance or large feedforward input and we showed that the dynamic firing rate trajectories in spiking networks can be precisely targeted without error-driven training algorithms.
The firing pattern of ventral midbrain dopamine neurons is controlled by afferent and intrinsic activity to generate prediction error signals that are essential for reward-based learning. Given the absence of intracellular in vivo recordings in the last three decades, the subthreshold membrane potential events that cause changes in dopamine neuron firing patterns remain unknown. By establishing stable in vivo whole-cell recordings of >100 spontaneously active midbrain dopamine neurons in anaesthetized mice, we identified the repertoire of subthreshold membrane potential signatures associated with distinct in vivo firing patterns. We demonstrate that dopamine neuron in vivo activity deviates from a single spike pacemaker pattern by eliciting transient increases in firing rate generated by at least two diametrically opposing biophysical mechanisms: a transient depolarization resulting in high frequency plateau bursts associated with a reactive, depolarizing shift in action potential threshold; and a prolonged hyperpolarization preceding slower rebound bursts characterized by a predictive, hyperpolarizing shift in action potential threshold. Our findings therefore illustrate a framework for the biophysical implementation of prediction error and sensory cue coding in dopamine neurons by tuning action potential threshold dynamics.
When a visual stimulus is repeated, average neuronal responses typically decrease, yet they might maintain or even increase their impact through increased synchronization. Previous work has found that many repetitions of a grating lead to increasing gamma-band synchronization. Here we show in awake macaque area V1 that both, repetition-related reductions in firing rate and increases in gamma are specific to the repeated stimulus. These effects showed some persistence on the timescale of minutes. Further, gamma increases were specific to the presented stimulus location. Importantly, repetition effects on gamma and on firing rates generalized to natural images. These findings suggest that gamma-band synchronization subserves the adaptive processing of repeated stimulus encounters, both for generating efficient stimulus responses and possibly for memory formation.
Human language relies on hierarchically structured syntax to facilitate efficient and robust communication. The correct processing of syntactic information is essential for successful communication between speakers. As an abstract level of language, syntax has often been studied separately from the physical form of the speech signal, thus often masking the interactions that can promote better syntactic processing in the human brain. We analyzed a MEG dataset to investigate how acoustic cues, specifically prosody, interact with syntactic operations. We examined whether prosody enhances the cortical encoding of syntactic representations. We decoded left-sided dependencies directly from brain activity and evaluated possible modulations of the decoding by the presence of prosodic boundaries. Our findings demonstrate that prosodic boundary presence improves the representation of left-sided dependencies, indicating the facilitative role of prosodic cues in processing abstract linguistic features. This study gives neurobiological evidence for the boosting of syntactic processing via interaction with prosody.