004 Datenverarbeitung; Informatik
Refine
Year of publication
Document Type
- Article (19) (remove)
Has Fulltext
- yes (19)
Is part of the Bibliography
- no (19)
Keywords
- synaptic plasticity (4)
- Fisher information (3)
- Hebbian learning (3)
- objective functions (3)
- deep learning (2)
- generating functionals (2)
- homeostatic adaption (2)
- CBM detector (1)
- CLVisc (1)
- Deep learning (1)
Institute
- Physik (19) (remove)
In this proceeding, we review our recent work using deep convolutional neural network (CNN) to identify the nature of the QCD transition in a hybrid modeling of heavy-ion collisions. Within this hybrid model, a viscous hydrodynamic model is coupled with a hadronic cascade “after-burner”. As a binary classification setup, we employ two different types of equations of state (EoS) of the hot medium in the hydrodynamic evolution. The resulting final-state pion spectra in the transverse momentum and azimuthal angle plane are fed to the neural network as the input data in order to distinguish different EoS. To probe the effects of the fluctuations in the event-by-event spectra, we explore different scenarios for the input data and make a comparison in a systematic way. We observe a clear hierarchy in the predictive power when the network is fed with the event-by-event, cascade-coarse-grained and event-fine-averaged spectra. The carefully-trained neural network can extract high-level features from pion spectra to identify the nature of the QCD transition in a realistic simulation scenario.
The Fisher information constitutes a natural measure for the sensitivity of a probability distribution with respect to a set of parameters. An implementation of the stationarity principle for synaptic learning in terms of the Fisher information results in a Hebbian self-limiting learning rule for synaptic plasticity. In the present work, we study the dependence of the solutions to this rule in terms of the moments of the input probability distribution and find a preference for non-Gaussian directions, making it a suitable candidate for independent component analysis (ICA). We confirm in a numerical experiment that a neuron trained under these rules is able to find the independent components in the non-linear bars problem. The specific form of the plasticity rule depends on the transfer function used, becoming a simple cubic polynomial of the membrane potential for the case of the rescaled error function. The cubic learning rule is also an excellent approximation for other transfer functions, as the standard sigmoidal, and can be used to show analytically that the proposed plasticity rules are selective for directions in the space of presynaptic neural activities characterized by a negative excess kurtosis.
We present an effective model for timing-dependent synaptic plasticity (STDP) in terms of two interacting traces, corresponding to the fraction of activated NMDA receptors and the concentration in the dendritic spine of the postsynaptic neuron. This model intends to bridge the worlds of existing simplistic phenomenological rules and highly detailed models, thus constituting a practical tool for the study of the interplay of neural activity and synaptic plasticity in extended spiking neural networks. For isolated pairs of pre- and postsynaptic spikes, the standard pairwise STDP rule is reproduced, with appropriate parameters determining the respective weights and timescales for the causal and the anticausal contributions. The model contains otherwise only three free parameters, which can be adjusted to reproduce triplet nonlinearities in hippocampal culture and cortical slices. We also investigate the transition from time-dependent to rate-dependent plasticity occurring for both correlated and uncorrelated spike patterns.
Generating functionals may guide the evolution of a dynamical system and constitute a possible route for handling the complexity of neural networks as relevant for computational intelligence.We propose and explore a new objective function, which allows to obtain plasticity rules for the afferent synaptic weights. The adaption rules are Hebbian, self-limiting, and result from the minimization of the Fisher information with respect to the synaptic flux. We perform a series of simulations examining the behavior of the new learning rules in various circumstances.The vector of synaptic weights aligns with the principal direction of input activities, whenever one is present. A linear discrimination is performed when there are two or more principal directions; directions having bimodal firing-rate distributions, being characterized by a negative excess kurtosis, are preferred. We find robust performance and full homeostatic adaption of the synaptic weights results as a by-product of the synaptic flux minimization. This self-limiting behavior allows for stable online learning for arbitrary durations.The neuron acquires new information when the statistics of input activities is changed at a certain point of the simulation, showing however, a distinct resilience to unlearn previously acquired knowledge. Learning is fast when starting with randomly drawn synaptic weights and substantially slower when the synaptic weights are already fully adapted.
Poster presentation: The brain is autonomously active and this self-sustained neural activity is in general modulated, but not driven, by the sensory input data stream [1,2]. Traditionally one has regarded this eigendynamics as resulting from inter-modular recurrent neural activity [3]. Understanding the basic modules for cognitive computation is, in this view, the primary focus of research and the overall neural dynamics would be determined by the the topology of the intermodular pathways. Here we examine an alternative point of view, asking whether certain aspects of the neural eigendynamics have a central functional role for overall cognitive computation [4,5]. Transiently stable neural activity is regularly observed on the cognitive time-scale of 80–100 ms, with indications that neural competition [6] plays an important role in the selection of the transiently stable neural ensembles [7], also denoted winning coalitions [8]. We report on a theory approach which implements these two principles, transient-state dynamics and neural competition, in terms of an associative neural network with clique encoding [9]. A cognitive system [10] with a non-trivial internal eigendynamics has two seemingly contrasting tasks to fulfill. The internal processes need to be regular and not chaotic on one side, but sensitive to the afferent sensory stimuli on the other side. We show, that these two contrasting demands can be reconciled within our approach based on competitive transient-state dynamics, when allowing the sensory stimuli to modulate the competition for the next winning coalition. By testing the system with the bars problem, we find an emerging cognitive capability. Only based on the two basic architectural principles, neural competition and transient-state dynamics, with no explicit algorithmic encoding, the system performs on its own a non-linear independent component analysis of input data stream. The system has rudimentary biological features. All learning is local Hebbian-style, unsupervised and online. It exhibits an ever-ongoing eigendynamics and at no time is the state or the value of synaptic strengths reset or the system restarted; there is no separation between training and performance. We believe that this kind of approach – cognitive computation with autonomously active neural networks – to be an emerging field, relevant both for system neuroscience and synthetic cognitive systems.
Understanding the dynamics of recurrent neural networks is crucial for explaining how the brain processes information. In the neocortex, a range of different plasticity mechanisms are shaping recurrent networks into effective information processing circuits that learn appropriate representations for time-varying sensory stimuli. However, it has been difficult to mimic these abilities in artificial neural network models. Here we introduce SORN, a self-organizing recurrent network. It combines three distinct forms of local plasticity to learn spatio-temporal patterns in its input while maintaining its dynamics in a healthy regime suitable for learning. The SORN learns to encode information in the form of trajectories through its high-dimensional state space reminiscent of recent biological findings on cortical coding. All three forms of plasticity are shown to be essential for the network's success. Keywords: synaptic plasticity, intrinsic plasticity, recurrent neural networks, reservoir computing, time series prediction
Recent experimental findings have reported the presence of unconventional charge orders in the enlarged (2 × 2) unit-cell of kagome metals AV3Sb5 (A = K, Rb, Cs) and hinted towards specific topological signatures. Motivated by these discoveries, we investigate the types of topological phases that can be realized in such kagome superlattices. In this context, we employ a recently introduced statistical method capable of constructing topological models for any generic lattice. By analyzing large data sets generated from symmetry-guided distributions of randomized tight-binding parameters, and labeled with the corresponding topological index, we extract physically meaningful information. We illustrate the possible real-space manifestations of charge and bond modulations and associated flux patterns for different topological classes, and discuss their relation to present theoretical predictions and experimental signatures for the AV3Sb5 family. Simultaneously, we predict higher-order topological phases that may be realized by appropriately manipulating the currently known systems.
In this talk we presented a novel technique, based on Deep Learning, to determine the impact parameter of nuclear collisions at the CBM experiment. PointNet based Deep Learning models are trained on UrQMD followed by CBMRoot simulations of Au+Au collisions at 10 AGeV to reconstruct the impact parameter of collisions from raw experimental data such as hits of the particles in the detector planes, tracks reconstructed from the hits or their combinations. The PointNet models can perform fast, accurate, event-by-event impact parameter determination in heavy ion collision experiments. They are shown to outperform a simple model which maps the track multiplicity to the impact parameter. While conventional methods for centrality classification merely provide an expected impact parameter distribution for a given centrality class, the PointNet models predict the impact parameter from 2–14 fm on an event-by-event basis with a mean error of −0.33 to 0.22 fm.
The state-of-the-art pattern recognition method in machine learning (deep convolution neural network) is used to identify the equation of state (EoS) employed in the relativistic hydrodynamic simulations of heavy ion collisions. High-level correlations of particle spectra in transverse momentum and azimuthal angle learned by the network act as an effective EoS-meter in deciphering the nature of the phase transition in QCD. The EoS-meter is model independent and insensitive to other simulation inputs including the initial conditions and shear viscosity for hydrodynamic simulations. Through this study we demonstrate that there is a traceable encoder of the dynamical information from the phase structure that survives the evolution and exists in the final snapshot of heavy ion collisions and one can exclusively and effectively decode these information from the highly complex final output with machine learning when traditional methods fail. Besides the deep neural network, the performance of traditional machine learning classifiers are also provided.
Cortical pyramidal neurons have a complex dendritic anatomy, whose function is an active research field. In particular, the segregation between its soma and the apical dendritic tree is believed to play an active role in processing feed-forward sensory information and top-down or feedback signals. In this work, we use a simple two-compartment model accounting for the nonlinear interactions between basal and apical input streams and show that standard unsupervised Hebbian learning rules in the basal compartment allow the neuron to align the feed-forward basal input with the top-down target signal received by the apical compartment. We show that this learning process, termed coincidence detection, is robust against strong distractions in the basal input space and demonstrate its effectiveness in a linear classification task.
Gasdermin-D (GSDMD) is the ultimate effector of pyroptosis, a form of programmed cell death associated with pathogen invasion and inflammation. After proteolytic cleavage by caspases, the GSDMD N-terminal domain (GSDMDNT) assembles on the inner leaflet of the plasma membrane and induces the formation of membrane pores. We use atomistic molecular dynamics simulations to study GSDMDNT monomers, oligomers, and rings in an asymmetric plasma membrane mimetic. We identify distinct interaction motifs of GSDMDNT with phosphatidylinositol-4,5-bisphosphate (PI(4,5)P2) and phosphatidylserine (PS) headgroups and describe their conformational dependence. Oligomers are stabilized by shared lipid binding sites between neighboring monomers acting akin to double-sided tape. We show that already small GSDMDNT oligomers support stable, water-filled, and ion-conducting membrane pores bounded by curled beta-sheets. In large-scale simulations, we resolve the process of pore formation from GSDMDNT arcs and lipid efflux from partial rings. We find that high-order GSDMDNT oligomers can crack under the line tension of 86 pN created by an open membrane edge to form the slit pores or closed GSDMDNT rings seen in atomic force microscopy experiments. Our simulations provide a detailed view of key steps in GSDMDNT-induced plasma membrane pore formation, including sublytic pores that explain nonselective ion flux during early pyroptosis.
TriMem: A parallelized hybrid Monte Carlo software for efficient simulations of lipid membranes
(2022)
Lipid membranes are integral building blocks of living cells and perform a multitude of biological functions. Currently, molecular simulations of cellular-scale membrane remodeling processes at atomic resolution are extremely difficult, due to their size, complexity, and the large times-scales on which these processes occur. Instead, elastic membrane models are used to simulate membrane shapes and transitions between them and to infer their properties and functions. Unfortunately, an efficiently parallelized open-source simulation code to do so has been lacking. Here, we present TriMem, a parallel hybrid Monte Carlo simulation engine for triangulated lipid membranes. The kernels are efficiently coded in C++ and wrapped with Python for ease-of-use. The parallel implementation of the energy and gradient calculations and of Monte Carlo flip moves of edges in the triangulated membrane enable us to simulate large and highly curved membrane structures. For validation, we reproduce phase diagrams of vesicles with varying surface-to-volume ratios and area difference. We also compute the density of states to verify correct Boltzmann sampling. The software can be used to tackle a range of large-scale membrane remodeling processes as a step toward cell-scale simulations. Additionally, extensive documentation make the software accessible to the broad biophysics and computational cell biology communities.
We introduce a novel technique that utilizes a physics-driven deep learning method to reconstruct the dense matter equation of state from neutron star observables, particularly the masses and radii. The proposed framework involves two neural networks: one to optimize the EoS using Automatic Differentiation in the unsupervised learning scheme; and a pre-trained network to solve the Tolman–Oppenheimer–Volkoff (TOV) equations. The gradient-based optimization process incorporates a Bayesian picture into the proposed framework. The reconstructed EoS is proven to be consistent with the results from conventional methods. Furthermore, the resulting tidal deformation is in agreement with the limits obtained from the gravitational wave event, GW170817.
Consequences of minimal length discretization on line element, metric tensor, and geodesic equation
(2021)
When minimal length uncertainty emerging from a generalized uncertainty principle (GUP) is thoughtfully implemented, it is of great interest to consider its impacts on gravitational Einstein field equations (gEFEs) and to try to assess consequential modifications in metric manifesting properties of quantum geometry due to quantum gravity. GUP takes into account the gravitational impacts on the noncommutation relations of length (distance) and momentum operators or time and energy operators and so on. On the other hand, gEFE relates classical geometry or general relativity gravity to the energy–momentum tensors, that is, proposing quantum equations of state. Despite the technical difficulties, we intend to insert GUP into the metric tensor so that the line element and the geodesic equation in flat and curved space are accordingly modified. The latter apparently encompasses acceleration, jerk, and snap (jounce) of a particle in the quasi-quantized gravitational field. Finite higher orders of acceleration apparently manifest phenomena such as accelerating expansion and transitions between different radii of curvature and so on.
Zellulare Nichtlineare Netzwerke (CNN) wurden 1988 von Chua und Yang (Chua und Yang, 1988) eingeführt. Diese Netzwerke sind dadurch gekennzeichnet, dass eine Zelle, die die kleinste Einheit eines CNN darstellt, nur mit Zellen innerhalb einer bestimmten Umgebung verbunden ist. üblicherweise sind Art und Stärke der Wechselwirkung zwischen zwei Zellen eines CNN translationsinvariant, d.h. sie hängen nur von der relativen Lage beider Zellen zueinander ab. Im Vordergrund aktueller Arbeiten stehen auf derartigen Netzwerken basierende schaltungstechnische Realisierungen mit bis zu 176x144 Zellen, die eine direkte Verbindung zu zweidimensionalen optischen Sensor-Anordnungen aufweisen. Über einen separaten Speicherbereich können die Zellkopplungen eines Netzwerks verändert werden, wodurch eine adaptive Verarbeitung von mehrdimensionalen Sensorsignalen ermöglicht wird. Das kürzlich vorgestellte so genannte EyeRis System (Anafocus Ltd.) enthält zusätzlich noch einen Standardprozessor und stellt (bei einer Größe vergleichbar mit der einer Kreditkarte) daher ein vollständiges superschnelles System zur Informationsverarbeitung dar. In diesem Beitrag sollen, nach einem kurzen Überblick über die Eigenschaften von CNN, aktuelle Realisierungen und exemplarisch eine neuere eigene Anwendung vorgestellt und besprochen werden.
Changes in the efficacies of synapses are thought to be the neurobiological basis of learning and memory. The efficacy of a synapse depends on its current number of neurotransmitter receptors. Recent experiments have shown that these receptors are highly dynamic, moving back and forth between synapses on time scales of seconds and minutes. This suggests spontaneous fluctuations in synaptic efficacies and a competition of nearby synapses for available receptors. Here we propose a mathematical model of this competition of synapses for neurotransmitter receptors from a local dendritic pool. Using minimal assumptions, the model produces a fast multiplicative scaling behavior of synapses. Furthermore, the model explains a transient form of heterosynaptic plasticity and predicts that its amount is inversely related to the size of the local receptor pool. Overall, our model reveals logistical tradeoffs during the induction of synaptic plasticity due to the rapid exchange of neurotransmitter receptors between synapses.
A deep convolutional neural network (CNN) is developed to study symmetry energy (Esym(ρ)) effects by learning the mapping between the symmetry energy and the two-dimensional (transverse momentum and rapidity) distributions of protons and neutrons in heavy-ion collisions. Supervised training is performed with labeled data-set from the ultrarelativistic quantum molecular dynamics (UrQMD) model simulation. It is found that, by using proton spectra on event-by-event basis as input, the accuracy for classifying the soft and stiff Esym(ρ) is about 60% due to large event-by-event fluctuations, while by setting event-summed proton spectra as input, the classification accuracy increases to 98%. The accuracies for 5-label (5 different Esym(ρ)) classification task are about 58% and 72% by using proton and neutron spectra, respectively. For the regression task, the mean absolute errors (MAE) which measure the average magnitude of the absolute differences between the predicted and actual L (the slope parameter of Esym(ρ)) are about 20.4 and 14.8 MeV by using proton and neutron spectra, respectively. Fingerprints of the density-dependent nuclear symmetry energy on the transverse momentum and rapidity distributions of protons and neutrons can be identified by convolutional neural network algorithm.