Frankfurt Institute for Advanced Studies (FIAS)
Refine
Year of publication
Document Type
- Preprint (821)
- Article (715)
- Conference Proceeding (27)
- Doctoral Thesis (18)
- Part of Periodical (6)
- Contribution to a Periodical (3)
- Part of a Book (2)
- Diploma Thesis (1)
- Review (1)
Has Fulltext
- yes (1594) (remove)
Is part of the Bibliography
- no (1594)
Keywords
- Heavy Ion Experiments (20)
- Hadron-Hadron Scattering (11)
- Hadron-Hadron scattering (experiments) (10)
- LHC (10)
- Heavy-ion collision (7)
- schizophrenia (7)
- Black holes (6)
- Equation of state (5)
- QCD (5)
- Quark-Gluon Plasma (5)
Institute
- Frankfurt Institute for Advanced Studies (FIAS) (1594)
- Physik (1298)
- Informatik (1000)
- Medizin (64)
- MPI für Hirnforschung (30)
- Ernst Strüngmann Institut (26)
- Biowissenschaften (22)
- Psychologie (13)
- Biochemie und Chemie (11)
- Helmholtz International Center for FAIR (7)
Gamma synchronization has generally been associated with grouping processes in the visual system. Here, we examine in monkey V1 whether gamma oscillations play a functional role in segmenting surfaces of plaid stimuli. Local field potentials (LFPs) and spiking activity were recorded simultaneously from multiple sites in the opercular and calcarine regions while the monkeys were presented with sequences of single and superimposed components of plaid stimuli. In accord with the previous studies, responses to the single components (gratings) exhibited strong and sustained gamma-band oscillations (30–65 Hz). The superposition of the second component, however, led to profound changes in the temporal structure of the responses, characterized by a drastic reduction of gamma oscillations in the spiking activity and systematic shifts to higher frequencies in the LFP (~10% increase). Comparisons between cerebral hemispheres and across monkeys revealed robust subject-specific spectral signatures. A possible interpretation of our results may be that single gratings induce strong cooperative interactions among populations of cells that share similar response properties, whereas plaids lead to competition. Overall, our results suggest that the functional architecture of the cortex is a major determinant of the neuronal synchronization dynamics in V1. Key words: attention , gamma , gratings , oscillation , visual cortex
Human Transformer2-beta (hTra2-beta) is an important member of the serine/arginine-rich protein family, and contains one RNA recognition motif (RRM). It controls the alternative splicing of several pre-mRNAs, including those of the calcitonin/calcitonin gene-related peptide (CGRP), the survival motor neuron 1 (SMN1) protein and the tau protein. Accordingly, the RRM of hTra2-beta specifically binds to two types of RNA sequences [the CAA and (GAA)2 sequences]. We determined the solution structure of the hTra2-beta RRM (spanning residues Asn110–Thr201), which not only has a canonical RRM fold, but also an unusual alignment of the aromatic amino acids on the beta-sheet surface. We then solved the complex structure of the hTra2-beta RRM with the (GAA)2 sequence, and found that the AGAA tetra-nucleotide was specifically recognized through hydrogen-bond formation with several amino acids on the N- and C-terminal extensions, as well as stacking interactions mediated by the unusually aligned aromatic rings on the beta-sheet surface. Further NMR experiments revealed that the hTra2-beta RRM recognizes the CAA sequence when it is integrated in the stem-loop structure. This study indicates that the hTra2-beta RRM recognizes two types of RNA sequences in different RNA binding modes.
The goal of this project is to develop a framework for a cell that takes in consideration its internal structure, using an agent-based approach. In this framework, a cell was simulated as many sub-particles interacting to each other. This sub-particles can, in principle, represent any internal structure from the cell (organelles, etc). In the model discussed here, two types of sub-particles were used: membrane sub-particles and cytosolic elements. A kinetic and dynamic Delaunay triangulation was used in order to define the neighborhood relations between the sub-particles. However, it was soon noted that the relations defined by the Delaunay triangulation were not suitable to define the interactions between membrane sub-particles. The cell membrane is a lipid bilayer, and does not present any long range interactions between their sub-particles. This means that the membrane particles should not be able to interact in a long range. Instead, their interactions should be confined to the two-dimensional surface supposedly formed by the membrane. A method to select, from the original three-dimensional triangulations, connections restricted to the two-dimensional surface formed by the cell membrane was then developed. The algorithm uses as starting point the three-dimensional Delaunay triangulation involving both internal and membrane sub-particles. From this triangulation, only the subset of connections between membrane sub-particles was considered. Since the cell is full of internal particles, the collection of the membrane particles' connections will resemble the surface to be obtained, even though it will still have many connections that do not belong to the restricted triangulation on the surface. This "thick surface" was called a quasi-surface. The following step was to refine the quasi-surface, cutting out some of the connections so that the ones left made a proper surface triangulation with the membrane points. For that, the quasi-surface was separated in clusters. Clusters are defined as areas on the quasi-surface that are not yet properly triangulated on a two-dimensional surface. Each of the clusters was then re-triangulated independently, using re-triangulation methods also developed during this work. The interactions between cytosolic elements was given by a Lennard-Jones potential, as well as the interactions between cytosolic elements and membrane particles. Between only membrane particles, the interactions were given by an elastic interaction. For each particle, the equation of motion was written. The algorithm chosen to solve the equations of motion was the Verlet algorithm. Since the cytosol can be approximated as a gel, it is reasonable to suppose that the sub-cellular particles are moving in an overdamped environment. Therefore, an overdamped approximation was used for all interactions. Additionally, an adaptive algorithm was used in order to define the size of the time step used in each interaction. After the method to re-triangulate the membrane points was implemented, the time needed to re-triangulate a single cluster was studied, followed by an analysis on how the time needed to re-triangulate each point in a cluster varied with the cluster size. The frequency of appearance for each cluster size was also compared, as this information is necessary to guarantee that the total time needed by to re-triangulate a cell is convergent. At last, the total time spent re-triangulating a surface was plotted, as well as a scaling for the total re-triangulation time with the variation. Even though there is still a lot to be done, the work presented here is an important step on the way to the main goal of this project: to create an agent-based framework that not only allows the simulation of any sub-cellular structure of interest but also provides meaningful interaction relations to particles belonging to the cell membrane.
Although models based on independent component analysis (ICA) have been successful in explaining various properties of sensory coding in the cortex, it remains unclear how networks of spiking neurons using realistic plasticity rules can realize such computation. Here, we propose a biologically plausible mechanism for ICA-like learning with spiking neurons. Our model combines spike-timing dependent plasticity and synaptic scaling with an intrinsic plasticity rule that regulates neuronal excitability to maximize information transmission. We show that a stochastically spiking neuron learns one independent component for inputs encoded either as rates or using spike-spike correlations. Furthermore, different independent components can be recovered, when the activity of different neurons is decorrelated by adaptive lateral inhibition.
Understanding the dynamics of recurrent neural networks is crucial for explaining how the brain processes information. In the neocortex, a range of different plasticity mechanisms are shaping recurrent networks into effective information processing circuits that learn appropriate representations for time-varying sensory stimuli. However, it has been difficult to mimic these abilities in artificial neural network models. Here we introduce SORN, a self-organizing recurrent network. It combines three distinct forms of local plasticity to learn spatio-temporal patterns in its input while maintaining its dynamics in a healthy regime suitable for learning. The SORN learns to encode information in the form of trajectories through its high-dimensional state space reminiscent of recent biological findings on cortical coding. All three forms of plasticity are shown to be essential for the network's success. Keywords: synaptic plasticity, intrinsic plasticity, recurrent neural networks, reservoir computing, time series prediction
In the present work, the problem of protein folding is addressed from the point of view of equilibrium thermodynamics. The conformation of a globular protein in solution at common temperatures is quite complicated without any geometrical symmetry, but it is an ordered state in the sense of its biological activity. This complicated conformation of a single protein molecule is destroyed upon increasing the temperature or by the addition of appropriate chemical agents, as is revealed by the loss of its activity and change of the physical properties, and so on. Once the complicated native structures having biological activity are lost, it would be natural to suppose that the native structure could hardly be restored. Nevertheless, pioneers, such as Anson and Mirsky, recognized as early as in 1925 that this was not always the case. If one defines the folded and unfolded states of a protein as two distinct phases of a system, then under the variation of temperature the system is transformed from one phase state into another and vice versa. The process of protein folding is accompanied by the release or absorption of a certain amount of energy, corresponding to the first-oder-type phase transitions in the bulk. Knowing the partition function of the system one can evaluate its energy and heat capacity under different temperatures. This task was performed in this work. The results of the developed statistical mechanics model were compared with the results of molecular dynamic simulations of alanine poylpeptides. In particular, the dependencies on temperature of the total energy of the system and heat capacity were compared for alanine polypeptides consisting of 21, 30, 40, 50 and 100 amino acids. The good correspondence of the results of the theoretical model with the results of molecular dynamics simulations allowed to validate the assumptions made about the system and to establish the accuracy range of the theory. In order to perform the comparison of the results of theoretical model and the molecular dynamics simulations it is necessary to perform the efficient analysis of the results of molecular dynamics simulations. This task was also addressed in the present work. In particular, different ways to obtain dependence of the heat capacity on temperature from molecular dynamics simulations are discussed and the most efficient one is proposed. The present thesis reports the result of molecular dynamic simulations for not only alanine polypeptides by also for valine and leucine polypeptides. In valine and leucine polypeptides, it is also possible to observe the helix↔random coil transitions with the increase of temperature. The current thesis presents a work that starts with the investigation of the fundamental degrees of freedom in polypeptides that are responsible for the conformational transitions. Then this knowledge is applied for the statistical mechanics description of helix↔coil transitions in polypeptides. Finally, the theoretical formalism is generalized for the case of proteins in water environment and the comparison of the results of the statistical mechanics model with the experimental measurements of the heat capacity on temperature dependencies for two globular proteins is performed. The presented formalism is based on fundamental physical properties of the system and provides the possibility to describe the folding↔unfolding transitions quantitatively. The combination of these two facts is the major novelty of the presented approach in comparison to the existing ones. The “transparent” physical nature of the formalism provides a possibility to further apply it to a large variety of systems and processes. For instance, it can be used for investigation of the influence of the mutations in the proteins on their stability. This task is of primary importance for design of novel proteins and drug delivering molecules in medicine. It can provide further insights into the problem of protein aggregation and formation of amyloids. The problem of protein aggregation is closely associated with various illnesses such as Alzheimer and mad cow disease. With certain modifications, the presented theoretical method can be applied to the description of the protein crystallization process, which is important for the determination of the structure of proteins with X-Rays. There many other possible applications of the ideas described in the thesis. For instance, the similar formalism can be developed for the description of melting and unzipping of DNA, growth of nanotubes, formation of fullerenes, etc.
Short-term memory requires the coordination of sub-processes like encoding, retention, retrieval and comparison of stored material to subsequent input. Neuronal oscillations have an inherent time structure, can effectively coordinate synaptic integration of large neuron populations and could therefore organize and integrate distributed sub-processes in time and space. We observed field potential oscillations (14–95 Hz) in ventral prefrontal cortex of monkeys performing a visual memory task. Stimulus-selective and performance-dependent oscillations occurred simultaneously at 65–95 Hz and 14–50 Hz, the latter being phase-locked throughout memory maintenance. We propose that prefrontal oscillatory activity may be instrumental for the dynamical integration of local and global neuronal processes underlying short-term memory.
Experience-driven formation of parts-based representations in a model of layered visual memory
(2009)
Growing neuropsychological and neurophysiological evidence suggests that the visual cortex uses parts-based representations to encode, store and retrieve relevant objects. In such a scheme, objects are represented as a set of spatially distributed local features, or parts, arranged in stereotypical fashion. To encode the local appearance and to represent the relations between the constituent parts, there has to be an appropriate memory structure formed by previous experience with visual objects. Here, we propose a model how a hierarchical memory structure supporting efficient storage and rapid recall of parts-based representations can be established by an experience-driven process of self-organization. The process is based on the collaboration of slow bidirectional synaptic plasticity and homeostatic unit activity regulation, both running at the top of fast activity dynamics with winner-take-all character modulated by an oscillatory rhythm. These neural mechanisms lay down the basis for cooperation and competition between the distributed units and their synaptic connections. Choosing human face recognition as a test task, we show that, under the condition of open-ended, unsupervised incremental learning, the system is able to form memory traces for individual faces in a parts-based fashion. On a lower memory layer the synaptic structure is developed to represent local facial features and their interrelations, while the identities of different persons are captured explicitly on a higher layer. An additional property of the resulting representations is the sparseness of both the activity during the recall and the synaptic patterns comprising the memory traces. Keywords: visual memory, self-organization, unsupervised learning, competitive learning, bidirectional plasticity, activity homeostasis, parts-based representation, cortical column
We suggest a new method to compute the spectrum and wave functions of excited states. We construct a stochastic basis of Bargmann link states, drawn from a physical probability density distribution and compute transition amplitudes between stochastic basis states. From such transition matrix we extract wave functions and the energy spectrum. We apply this method toU(1)2+1 lattice gauge theory. As a test we compute the energy spectrum, wave functions and thermodynamical functions of the electric Hamiltonian and compare it with analytical results. We find excellent agreement. We observe scaling of energies and wave functions in the variable of time. We also present first results on a small lattice for the full Hamiltonian including the magnetic term.
Poster Presentation from Nineteenth Annual Computational Neuroscience Meeting: CNS*2010 San Antonio, TX, USA. 24-30 July 2010 Statistical models of neural activity are at the core of the field of modern computational neuroscience. The activity of single neurons has been modeled to successfully explain dependencies of neural dynamics to its own spiking history, to external stimuli or other covariates [1]. Recently, there has been a growing interest in modeling spiking activity of a population of simultaneously recorded neurons to study the effects of correlations and functional connectivity on neural information processing (existing models include generalized linear models [2,3] or maximum-entropy approaches [4]). For point-process-based models of single neurons, the time-rescaling theorem has proven to be a useful toolbox to assess goodness-of-fit. In its univariate form, the time-rescaling theorem states that if the conditional intensity function of a point process is known, then its inter-spike intervals can be transformed or “rescaled” so that they are independent and exponentially distributed [5]. However, the theorem in its original form lacks sensitivity to detect even strong dependencies between neurons. Here, we present how the theorem can be extended to be applied to neural population models and we provide a step-by-step procedure to perform the statistical tests. We then apply both the univariate and multivariate tests to simplified toy models, but also to more complicated many-neuron models and to neuronal populations recorded in V1 of awake monkey during natural scenes stimulation. We demonstrate that important features of the population activity can only be detected using the multivariate extension of the test. ...