Frankfurt Institute for Advanced Studies (FIAS)
Refine
Year of publication
Document Type
- Preprint (889)
- Article (738)
- Conference Proceeding (27)
- Doctoral Thesis (18)
- Part of Periodical (6)
- Contribution to a Periodical (3)
- Part of a Book (2)
- Diploma Thesis (1)
- Master's Thesis (1)
- Review (1)
Has Fulltext
- yes (1686) (remove)
Is part of the Bibliography
- no (1686)
Keywords
- Heavy Ion Experiments (21)
- Hadron-Hadron Scattering (11)
- Hadron-Hadron scattering (experiments) (10)
- LHC (10)
- Heavy-ion collision (7)
- schizophrenia (7)
- Black holes (6)
- Equation of state (5)
- QCD (5)
- Quark-Gluon Plasma (5)
Institute
- Frankfurt Institute for Advanced Studies (FIAS) (1686)
- Physik (1311)
- Informatik (1002)
- Medizin (64)
- MPI für Hirnforschung (31)
- Ernst Strüngmann Institut (26)
- Biowissenschaften (22)
- Psychologie (13)
- Biochemie und Chemie (11)
- Helmholtz International Center for FAIR (7)
Glia, the helper cells of the brain, are essential in maintaining neural resilience across time and varying challenges: By reacting to changes in neuronal health glia carefully balance repair or disposal of injured neurons. Malfunction of these interactions is implicated in many neurodegenerative diseases. We present a reductionist model that mimics repair-or-dispose decisions to generate a hypothesis for the cause of disease onset. The model assumes four tissue states: healthy and challenged tissue, primed tissue at risk of acute damage propagation, and chronic neurodegeneration. We discuss analogies to progression stages observed in the most common neurodegenerative conditions and to experimental observations of cellular signaling pathways of glia-neuron crosstalk. The model suggests that the onset of neurodegeneration can result as a compromise between two conflicting goals: short-term resilience to stressors versus long-term prevention of tissue damage.
We derive the relation between cumulants of a conserved charge measured in a subvolume of a thermal system and the corresponding grand-canonical susceptibilities, taking into account exact global conservation of that charge. The derivation is presented for an arbitrary equation of state, with the assumption that the subvolume is sufficiently large to be close to the thermodynamic limit. Our framework – the subensemble acceptance method (SAM) – quantifies the effect of global conservation laws and is an important step toward a direct comparison between cumulants of conserved charges measured in central heavy ion collisions and theoretical calculations of grand-canonical susceptibilities, such as lattice QCD. As an example, we apply our formalism to net-baryon fluctuations at vanishing baryon chemical potentials as encountered in collisions at the LHC and RHIC.
EEG microstate periodicity explained by rotating phase patterns of resting-state alpha oscillations
(2020)
Spatio-temporal patterns in electroencephalography (EEG) can be described by microstate analysis, a discrete approximation of the continuous electric field patterns produced by the cerebral cortex. Resting-state EEG microstates are largely determined by alpha frequencies (8-12 Hz) and we recently demonstrated that microstates occur periodically with twice the alpha frequency.
To understand the origin of microstate periodicity, we analyzed the analytic amplitude and the analytic phase of resting-state alpha oscillations independently. In continuous EEG data we found rotating phase patterns organized around a small number of phase singularities which varied in number and location. The spatial rotation of phase patterns occurred with the underlying alpha frequency. Phase rotors coincided with periodic microstate motifs involving the four canonical microstate maps. The analytic amplitude showed no oscillatory behaviour and was almost static across time intervals of 1-2 alpha cycles, resulting in the global pattern of a standing wave.
In n=23 healthy adults, time-lagged mutual information analysis of microstate sequences derived from amplitude and phase signals of awake eyes-closed EEG records showed that only the phase component contributed to the periodicity of microstate sequences. Phase sequences showed mutual information peaks at multiples of 50 ms and the group average had a main peak at 100 ms (10 Hz), whereas amplitude sequences had a slow and monotonous information decay. This result was confirmed by an independent approach combining temporal principal component analysis (tPCA) and autocorrelation analysis.
We reproduced our observations in a generic model of EEG oscillations composed of coupled non-linear oscillators (Stuart-Landau model). Phase-amplitude dynamics similar to experimental EEG occurred when the oscillators underwent a supercritical Hopf bifurcation, a common feature of many computational models of the alpha rhythm.
These findings explain our previous description of periodic microstate recurrence and its relation to the time scale of alpha oscillations. Moreover, our results corroborate the predictions of computational models and connect experimentally observed EEG patterns to properties of critical oscillator networks.
p53 regulates the cellular response to genotoxic damage and prevents carcinogenic events. Theoretical and experimental studies state that the p53-Mdm2 network constitutes the core module of regulatory interactions activated by cellular stress induced by a variety of signaling pathways. In this paper, a strategy to control the p53-Mdm2 network regulated by p14ARF is developed, based on the pinning control technique, which consists into applying local feedback controllers to a small number of nodes (pinned ones) in the network. Pinned nodes are selected on the basis of their importance level in a topological hierarchy, their degree of connectivity within the network, and the biological role they perform. In this paper, two cases are considered. For the first case, the oscillatory pattern under gamma-radiation is recovered; afterward, as the second case, increased expression of p53 level is taken into account. For both cases, the control law is applied to p14ARF (pinned node based on a virtual leader methodology), and overexpressed Mdm2-mediated p53 degradation condition is considered as carcinogenic initial behavior. The approach in this paper uses a computational algorithm, which opens an alternative path to understand the cellular responses to stress, doing it possible to model and control the gene regulatory network dynamics in two different biological contexts. As the main result of the proposed control technique, the two mentioned desired behaviors are obtained.
A new method of event characterization based on Deep Learning is presented. The PointNet models can be used for fast, online event-by-event impact parameter determination at the CBM experiment. For this study, UrQMD and the CBM detector simulation are used to generate Au+Au collision events at 10 AGeV which are then used to train and evaluate PointNet based architectures. The models can be trained on features like the hit position of particles in the CBM detector planes, tracks reconstructed from the hits or combinations thereof. The Deep Learning models reconstruct impact parameters from 2-14 fm with a mean error varying from -0.33 to 0.22 fm. For impact parameters in the range of 5-14 fm, a model which uses the combination of hit and track information of particles has a relative precision of 4-9% and a mean error of -0.33 to 0.13 fm. In the same range of impact parameters, a model with only track information has a relative precision of 4-10% and a mean error of -0.18 to 0.22 fm. This new method of event-classification is shown to be more accurate and less model dependent than conventional methods and can utilize the performance boost of modern GPU processor units.
Summary
Wild relatives of crops thrive in habitats where environmental conditions can be restrictive for productivity and survival of cultivated species. The genetic basis of this variability, particularly for tolerance to high temperatures, is not well understood. We examined the capacity of wild and cultivated accessions to acclimate to rapid temperature elevations that cause heat stress (HS).
We investigated genotypic variation in thermotolerance of seedlings of wild and cultivated accessions. The contribution of polymorphisms associated with thermotolerance variation was examined regarding alterations in function of the identified gene.
We show that tomato germplasm underwent a progressive loss of acclimation to strong temperature elevations. Sensitivity is associated with intronic polymorphisms in the HS transcription factor HsfA2 which affect the splicing efficiency of its pre‐mRNA. Intron splicing in wild species results in increased synthesis of isoform HsfA2‐II, implicated in the early stress response, at the expense of HsfA2‐I which is involved in establishing short‐term acclimation and thermotolerance.
We propose that the selection for modern HsfA2 haplotypes reduced the ability of cultivated tomatoes to rapidly acclimate to temperature elevations, but enhanced their short‐term acclimation capacity. Hence, we provide evidence that alternative splicing has a central role in the definition of plant fitness plasticity to stressful conditions.
Our primary objective is to construct a plausible, unified model of inflation, dark energy and dark matter from a fundamental Lagrangian action first principle, wherein all fundamental ingredients are systematically dynamically generated starting from a very simple model of modified gravity interacting with a single scalar field employing the formalism of non-Riemannian spacetime volume-elements. The non-Riemannian volume element in the initial scalar field action leads to a hidden, nonlinear Noether symmetry which produces an energy-momentum tensor identified as the sum of a dynamically generated cosmological constant and dust-like dark matter. The non-Riemannian volume-element in the initial Einstein–Hilbert action upon passage to the physical Einstein-frame creates, dynamically, a second scalar field with a non-trivial inflationary potential and with an additional interaction with the dynamically generated dark matter. The resulting Einstein-frame action describes a fully dynamically generated inflationary model coupled to dark matter. Numerical results for observables such as the scalar power spectral index and the tensor-to-scalar ratio conform to the latest 2018 PLANCK data.
We estimate the feeddown contributions from decays of unstable A=4 and A=5 nuclei to the final yields of protons, deuterons, tritons, 3He, and 4He produced in relativistic heavy-ion collisions at sNN>2.4 GeV, using the statistical model. The feeddown contribution effects do not exceed 5% at LHC and top RHIC energies due to the large penalty factors involved, but are substantial at intermediate collision energies. We observe large feeddown contributions for tritons, 3He, and 4He at sNN≲10 GeV, where they may account for as much as 70% of the final yield at the lower end of the collision energies considered. Sizable (>10%) effects for deuteron yields are observed at sNN≲4 GeV. The results suggest that the excited nuclei feeddown cannot be neglected in the ongoing and future analysis of light nuclei production at intermediate collision energies, including HADES and CBM experiments at FAIR, NICA at JINR, RHIC beam energy scan and fixed-target programmes, and NA61/SHINE at CERN. We further show that the freeze-out curve in the T-μB plane itself is affected significantly by the light nuclei at high baryochemical potential.
In this paper, we discuss the damping of density oscillations in dense nuclear matter in the temperature range relevant to neutron star mergers. This damping is due to bulk viscosity arising from the weak interaction “Urca” processes of neutron decay and electron capture. The nuclear matter is modelled in the relativistic density functional approach. The bulk viscosity reaches a resonant maximum close to the neutrino trapping temperature, then drops rapidly as temperature rises into the range where neutrinos are trapped in neutron stars. We investigate the bulk viscous dissipation timescales in a post-merger object and identify regimes where these timescales are as short as the characteristic timescale ∼10 ms, and, therefore, might affect the evolution of the post-merger object. Our analysis indicates that bulk viscous damping would be important at not too high temperatures of the order of a few MeV and densities up to a few times saturation density.
We study D and DS mesons at finite temperature using an effective field theory based on chiral and heavy-quark spin-flavor symmetries within the imaginary-time formalism. Interactions with the light degrees of freedom are unitarized via a Bethe-Salpeter approach, and the D and self-energies are calculated self-consistently. We generate dynamically the e D∗0(2300)and Ds(2317)state, and study their possible identification as the chiral We study Dand Dsmesons at finite temperature using an effective field theory based on chiral and heavy-quark spin-flavor symmetries within the imaginary-time formalism. Interactions with the light degrees of freedom are unitarized via a Bethe-Salpeter approach, and the Dand Dsself-energies are calculated self-consistently. We generate dynamically the D∗0(2300)and Ds(2317)states, and study their possible identification as the chiral partners of the Dand Dsground states, respectively. We show the evolution of their masses and decay widths as functions of temperature, and provide an analysis of the chiral-symmetry restoration in the heavy-flavor sector below the transition temperature. In particular, we analyse the very special case of the D-meson, for which the chiral partner is associated to the double-pole structure of the D∗0(2300).
First, we propose a scale-invariant modified gravity interacting with a neutral scalar inflaton and a Higgs-like SU(2)×U(1) iso-doublet scalar field based on the formalism of non-Riemannian (metric-independent) spacetime volume-elements. This model describes, in the physical Einstein frame, a quintessential inflationary scenario driven by the “inflaton” together with the gravity-“inflaton” assisted dynamical spontaneous SU(2)×U(1) symmetry breaking in the post-inflationary universe, whereas the SU(2)×U(1) symmetry remains intact in the inflationary epoch. Next, we find the explicit representation of the latter quintessential inflationary model with a dynamical Higgs effect as an Eddington-type purely affine gravity.
Measurement of ϒ(1S) elliptic flow at forward rapidity in Pb-Pb collisions at √sNN = 5.02 TeV
(2019)
The first measurement of the ϒ(1S) elliptic flow coefficient (v2) is performed at forward rapidity (2.5 < y < 4) in Pb–Pb collisions at √sNN = 5.02 TeV with the ALICE detector at the LHC. The results are obtained with the scalar product method and are reported as a function of transverse momentum (pT) up to 15 GeV/c in the 5%–60% centrality interval. The measured Υ(1S)v2 is consistent with 0 and with the small positive values predicted by transport models within uncertainties. The v2 coefficient in 2 < pT < 15 GeV/c is lower than that of inclusive J/ψ mesons in the same pT interval by 2.6 standard deviations. These results, combined with earlier suppression measurements, are in agreement with a scenario in which the Υ(1S) production in Pb–Pb collisions at LHC energies is dominated by dissociation limited to the early stage of the collision, whereas in the J/ψ case there is substantial experimental evidence of an additional regeneration component.
Distillation of scalar exchange by coherent hypernucleus production in antiproton–nucleus collisions
(2017)
The total and angular differential cross sections of the coherent process p¯ + A Z → A (Z − 1) + ¯ are evaluated at the beam momenta 1.5 ÷ 20 GeV/c within the meson exchange model with bound proton and -hyperon wave functions. It is shown that the shape of the beam momentum dependence of the hypernucleus production cross sections with various discrete states is strongly sensitive to the presence of the scalar κ-meson exchange in the p p¯ → ¯ amplitude. This can be used as a clean test of the exchange by scalar π K correlation in coherent p A¯ reactions.
The study of hypernuclei in relativistic ion collisions open new opportunities for nuclear and particle physics. The main processes leading to the production of hypernuclei in these reactions are the disintegration of large excited hyper-residues (target- and projectile-like), and the coalescence of hyperons with other baryons into light clusters. We use the transport, coalescence and statistical models to describe the whole reaction, and demonstrate the effectiveness of this approach: These reactions lead to the abundant production of multi-strange nuclei and new hypernuclear states. A broad distribution of predicted hypernuclei in masses and isospin allows for investigating properties of exotic hypernuclei, as well as the hypermatter both at high and low temperatures. There is a saturation of the hypernuclei production at high energies, therefore, the optimal way to pursue this experimental research is to use the accelerator facilities of intermediate energies, like FAIR (Darmstadt) and NICA (Dubna).
Formation of hypermatter and hypernuclei within transport models in relativistic ion collisions
(2015)
Within a combined approach we investigate the main features of the production of hyper-fragments in relativistic heavy-ion collisions. The formation of hyperons is modeled within the UrQMD and HSD transport codes. To describe the hyperon capture by nucleons and nuclear residues a coalescence of baryons (CB) model was developed. We demonstrate that the origin of hypernuclei of various masses can be explained by typical baryon interactions, and that it is similar to processes leading to the production of conventional nuclei. At high beam energies we predict a saturation of the yields of all hyper-fragments, therefore, this kind of reactions can be studied with high yields even at the accelerators of moderate relativistic energies.
Observations show that, at the beginning of their existence, neutron stars are accelerated briskly to velocities of up to a thousand kilometers per second. We argue that this remarkable effect can be explained as a manifestation of quantum anomalies on astrophysical scales. To theoretically describe the early stage in the life of neutron stars we use hydrodynamics as a systematic effective-field-theory framework. Within this framework, anomalies of the Standard Model of particle physics as underlying microscopic theory imply the presence of a particular set of transport terms, whose form is completely fixed by theoretical consistency. The resulting chiral transport effects in proto-neutron stars enhance neutrino emission along the internal magnetic field, and the recoil can explain the order of magnitude of the observed kick velocities.
Unparticle Casimir effect
(2017)
In this paper we present the un-Casimir effect, namely the study of the Casimir energy in the presence of an unparticle component in addition to the electromagnetic field contribution. The distinctive feature of the un-Casimir effect is a fractalization of metallic plates. This result emerges through a new dependence of the Casimir energy on the plate separation that scales with a continuous power controlled by the unparticle dimension. As long as the perfect conductor approximation is valid, we find bounds on the unparticle scale that are independent of the effective coupling constant between the scale invariant sector and ordinary matter. We find regions of the parameter space such that for plate distances around 5 μm and larger the un-Casimir bound wins over the other bounds.
We calculate ratios of higher-order susceptibilities quantifying fluctuations in the number of net-protons and in the net-electric charge using the Hadron Resonance Gas (HRG) model. We take into account the effect of resonance decays, the kinematic acceptance cuts in rapidity, pseudo-rapidity and transverse momentum used in the experimental analysis, as well as a randomization of the isospin of nucleons in the hadronic phase. By comparing these results to the latest experimental data from the STAR Collaboration, we determine the freeze-out conditions from net-electric charge and net-proton distributions and discuss their consistency.
Motivated by a recent finding of an exact solution of the relativistic Boltzmann equation in a Friedmann–Robertson–Walker spacetime, we implement this metric into the newly developed transport approach Simulating Many Accelerated Strongly-interacting Hadrons (SMASH). We study the numerical solution of the transport equation and compare it to this exact solution for massless particles. We also compare a different initial condition, for which the transport equation can be independently solved numerically. Very nice agreement is observed in both cases. Having passed these checks for the SMASH code, we study a gas of massive particles within the same spacetime, where the particle decoupling is forced by the Hubble expansion. In this simple scenario we present an analysis of the freeze-out times, as function of the masses and cross sections of the particles. The results might be of interest for their potential application to relativistic heavy-ion collisions, for the characterization of the freeze-out process in terms of hadron properties.
We compare the reconstructed hadronization conditions in relativistic nuclear collisions in the nucleon–nucleon centre-of-mass energy range 4.7–2760 GeV in terms of temperature and baryon-chemical potential with lattice QCD calculations, by using hadronic multiplicities. We obtain hadronization temperatures and baryon chemical potentials with a fit to measured multiplicities by correcting for the effect of post-hadronization rescattering. The post-hadronization modification factors are calculated by means of a coupled hydrodynamical-transport model simulation under the same conditions of approximate isothermal and isochemical decoupling as assumed in the statistical hadronization model fits to the data. The fit quality is considerably better than without rescattering corrections, as already found in previous work. The curvature of the obtained “true” hadronization pseudo-critical line κ is found to be 0.0048 ± 0.0026, in agreement with lattice QCD estimates; the pseudo-critical temperature at vanishing is found to be 164.3 ± 1.8 MeV.
We discuss different models for the spin structure of the nonperturbative pomeron: scalar, vector, and rank-2 symmetric tensor. The ratio of single-helicity-flip to helicity-conserving amplitudes in polarised high-energy proton–proton elastic scattering, known as the complex r5 parameter, is calculated for these models. We compare our results to experimental data from the STAR experiment. We show that the spin-0 (scalar) pomeron model is clearly excluded by the data, while the vector pomeron is inconsistent with the rules of quantum field theory. The tensor pomeron is found to be perfectly consistent with the STAR data.
We study the production of entropy in the context of a nonequilibrium chiral phase transition. The dynamical symmetry breaking is modeled by a Langevin equation for the order parameter coupled to the Bjorken dynamics of a quark plasma. We investigate the impact of dissipation and noise on the entropy and explore the possibility of reheating for crossover and first-order phase transitions, depending on the expansion rate of the fluid. The relative increase in is estimated to range from 10% for a crossover to 100% for a first-order phase transition at low beam energies, which could be detected in the pion-to-proton ratio as a function of beam energy.
We compute the probability distribution P(N) of the net-baryon number at finite temperature and quark-chemical potential, μ, at a physical value of the pion mass in the quark-meson model within the functional renormalization group scheme. For μ/T < 1, the model exhibits the chiral crossover transition which belongs to the universality class of the O(4) spin system in three dimensions. We explore the influence of the chiral crossover transition on the properties of the net baryon number probability distribution, P(N). By considering ratios of P(N) to the Skellam function, with the same mean and variance, we unravel the characteristic features of the distribution that are related to O(4) criticality at the chiral crossover transition. We explore the corresponding ratios for data obtained at RHIC by the STAR Collaboration and discuss their implications. We also examine O(4) criticality in the context of binomial and negative-binomial distributions for the net proton number.
We present a systematic study of the normalized symmetric cumulants, NSC(n,m), at the eccentricity level in proton-proton interactions at within a wounded hot spot approach. We focus our attention on the influence of spatial correlations between the proton constituents, in our case gluonic hot spots, on this observable. We notice that the presence of short-range repulsive correlations between the hot spots systematically decreases the values of and in mid- to ultra-central collisions while increases them in peripheral interactions. In the case of we find that, as suggested by data, an anti-correlation of and in ultra-central collisions, i.e. , is possible within the correlated scenario while it never occurs without correlations when the number of gluonic hot spots is set to three. We attribute this fact to the decisive role of correlations on enlarging the probability of interaction topologies that reduce the value of and, eventually, make it negative. Further, we explore the dependence of our conclusions on the number of hot spots, the values of the hot spot radius and the repulsive core distance. Our results add evidence to the idea that considering spatial correlations between the subnucleonic degrees of freedom of the proton may have a strong impact on the initial state properties of proton-proton interactions [1].
In this paper we discuss to what extent one can infer details of the interior structure of a black hole based on its horizon. Recalling that black hole thermal properties are connected to the non-classical nature of gravity, we circumvent the restrictions of the no-hair theorem by postulating that the black hole interior is singularity free due to violations of the usual energy conditions. Further these conditions allow one to establish a one-to-one, holographic projection between Planckian areal “bits” on the horizon and “voxels”, representing the gravitational degrees of freedom in the black hole interior. We illustrate the repercussions of this idea by discussing an example of the black hole interior consisting of a de Sitter core postulated to arise from the local graviton quantum vacuum energy. It is shown that the black hole entropy can emerge as the statistical entropy of a gas of voxels.
We present an analysis of the role of the charge within the self-complete quantum gravity paradigm. By studying the classicalization of generic ultraviolet improved charged black hole solutions around the Planck scale, we showed that the charge introduces important differences with respect to the neutral case. First, there exists a family of black hole parameters fulfilling the particle-black hole condition. Second, there is no extremal particle-black hole solution but quasi extremal charged particle-black holes at the best. We showed that the Hawking emission disrupts the condition of particle-black hole. By analyzing the Schwinger pair production mechanism, the charge is quickly shed and the particle-black hole condition can ultimately be restored in a cooling down phase towards a zero temperature configuration, provided non-classical effects are taken into account.
Bardeen black hole chemistry
(2019)
In the present paper we try to connect the Bardeen black hole with the concept of the recently proposed black hole chemistry. We study thermodynamic properties of the regular black hole with an anti-deSitter background. The negative cosmological constant Λ plays the role of the positive thermodynamic pressure of the system. After studying the thermodynamic variables, we derive the corresponding equation of state and we show that a neutral Bardeen-anti-deSitter black hole has similar phenomenology to the chemical Van der Waals fluid. This is equivalent to saying that the system exhibits criticality and a first order small/large black hole phase transition reminiscent of the liquid/gas coexistence.
The properties of open strange meson K1± in nuclear matter are estimated in the QCD sum rule approach. We obtain a relation between the in-medium mass and width of K1− (K1+) in nuclear matter, and show that the upper limit of the mass shift is as large as −249 (−35) MeV. The spectral modification of the K1 meson is possible to be probed by using kaon beams at J-PARC. Such measurement together with that of K⁎ will shed light on how chiral symmetry is partially restored in nuclear matter.
The effect of a non-zero strangeness chemical potential on the strong interaction phase diagram has been studied within the framework of the SU(3) quark-hadron chiral parity-doublet model. Both, the nuclear liquid-gas and the chiral/deconfinement phase transitions are modified. The first-order line in the chiral phase transition is observed to vanish completely, with the entire phase boundary becoming a crossover. These changes in the nature of the phase transitions are expected to modify various susceptibilities, the effects of which might be detectable in particle-number distributions resulting from moderate-temperature and high-density heavy-ion collision experiments.
In this letter we present some stringy corrections to black hole spacetimes emerging from string T-duality. As a first step, we derive the static Newtonian potential by exploiting the relation between the T-duality and the path integral duality. We show that the intrinsic non-perturbative nature of stringy corrections introduces an ultraviolet cutoff known as zero-point length in the path integral duality literature. As a result, the static potential is found to be regular. We use this result to derive a consistent black hole metric for the spherically symmetric, electrically neutral case. It turns out that the new spacetime is regular and is formally equivalent to the Bardeen metric, apart from a different ultraviolet regulator. On the thermodynamics side, the Hawking temperature admits a maximum before a cooling down phase towards a thermodynamically stable end of the black hole evaporation process. The findings support the idea of universality of quantum black holes.
We consider a simple model of modified gravity interacting with a single scalar field ϕ with weakly coupled exponential potential within the framework of non-Riemannian spacetime volume-form formalism. The specific form of the action is fixed by the requirement of invariance under global Weyl-scale symmetry. Upon passing to the physical Einstein frame we show how the non-Riemannian volume elements create a second canonical scalar field u and dynamically generate a non-trivial two-scalar-field potential Ueff(u,ϕ) with two remarkable features: (i) it possesses a large flat region for large u describing a slow-roll inflation; (ii) it has a stable low-lying minimum w.r.t. (u,ϕ) representing the dark energy density in the “late universe”. We study the corresponding two-field slow-roll inflation and show that the pertinent slow-roll inflationary curve ϕ = ϕ(u) in the two-field space (u,ϕ) has a very small curvature, i.e., ϕ changes very little during the inflationary evolution of u on the flat region of Ueff(u,ϕ). Explicit expressions are found for the slow-roll parameters which differ from those in the single-field inflationary counterpart. Numerical solutions for the scalar spectral index and the tensor-to-scalar ratio are derived agreeing with the observational data.
Rethinking superdeterminism
(2020)
Quantum mechanics has irked physicists ever since its conception more than 100 years ago. While some of the misgivings, such as it being unintuitive, are merely aesthetic, quantum mechanics has one serious shortcoming: it lacks a physical description of the measurement process. This “measurement problem” indicates that quantum mechanics is at least an incomplete theory—good as far as it goes, but missing a piece—or, more radically, is in need of complete overhaul. Here we describe an approach which may provide this sought-for completion or replacement: Superdeterminism. A superdeterministic theory is one which violates the assumption of Statistical Independence (that distributions of hidden variables are independent of measurement settings). Intuition suggests that Statistical Independence is an essential ingredient of any theory of science (never mind physics), and for this reason Superdeterminism is typically discarded swiftly in any discussion of quantum foundations. The purpose of this paper is to explain why the existing objections to Superdeterminism are based on experience with classical physics and linear systems, but that this experience misleads us. Superdeterminism is a promising approach not only to solve the measurement problem, but also to understand the apparent non-locality of quantum physics. Most importantly, we will discuss how it may be possible to test this hypothesis in an (almost) model independent way.
In this work, we discuss the dense matter equation of state (EOS) for the extreme range of conditions encountered in neutron stars and their mergers. The calculation of the properties of such an EOS involves modeling different degrees of freedom (such as nuclei, nucleons, hyperons, and quarks), taking into account different symmetries, and including finite density and temperature effects in a thermodynamically consistent manner. We begin by addressing subnuclear matter consisting of nucleons and a small admixture of light nuclei in the context of the excluded volume approach. We then turn our attention to supranuclear homogeneous matter as described by the Chiral Mean Field (CMF) formalism. Finally, we present results from realistic neutron-star-merger simulations performed using the CMF model that predict signatures for deconfinement to quark matter in gravitational wave signals.
In power systems, flow allocation (FA) methods enable to allocate the usage and costs of the transmission grid to each single market participant. Based on predefined assumptions, the power flow is split into isolated generator-specific or producer-specific sub-flows. Two prominent FA methods, Marginal Participation (MP) and Equivalent Bilateral Exchanges (EBEs), build upon the linearized power flow and thus on the Power Transfer Distribution Factors (PTDFs). Despite their intuitive and computationally efficient concepts, they are restricted to networks with passive transmission elements only. As soon as a significant number of controllable transmission elements, such as high-voltage direct current (HVDC) lines, operate in the system, they lose their applicability. This work reformulates the two methods in terms of Virtual Injection Patterns (VIPs), which allows one to efficiently introduce a shift parameter q to tune contributions of net sources and net sinks in the network. In this work, major properties and differences in the methods are pointed out, and it is shown how the MP and EBE algorithms can be applied to generic meshed AC-DC electricity grids: by introducing a pseudo-impedance ω¯ , which reflects the operational state of controllable elements and allows one to extend the PTDF matrix under the assumption of knowing the current flow in the system. Basic properties from graph theory are used to solve for the pseudo-impedance in dependence of the position within the network. This directly enables, e.g., HVDC lines to be considered in the MP and EBE algorithms. The extended methods are applied to a low-carbon European network model (PyPSA-EUR) with a spatial resolution of 181 nodes and an 18% transmission expansion compared to today’s total transmission capacity volume. The allocations of MP and EBE show that countries with high wind potentials profit most from the transmission grid expansion. Based on the average usage of transmission system expansion, a method of distributing operational and capital expenditures is proposed. In addition, it is shown how injections from renewable resources strongly drive country-to-country allocations and thus cross-border electricity flows.
The Karl Schwarzschild Meeting 2017 (KSM2017) has been the third instalment of the conference dedicated to the great Frankfurter scientist, who derived the first black hole solution of Einstein's equations about 100 years ago.
The event has been a 5 day meeting in the field of black holes, AdS/CFT correspondence and gravitational physics. Like the two previous instalments, the conference continued to attract a stellar ensemble of participants from the world's most renowned institutions. The core of the meeting has been a series of invited talks from eminent experts (keynote speakers) as well as the presence of plenary research talks by students and junior speakers.
List of Conference photo and poster, Sponsors and funding acknowledgments, Committees and List of participants are available in this PDF.
We have built quasi-equilibrium models for uniformly rotating quark stars in general relativity. The conformal flatness approximation is employed and the Compact Object CALculator (cocal) code is extended to treat rotating stars with surface density discontinuity. In addition to the widely used MIT bag model, we have considered a strangeon star equation of state (EoS), suggested by Lai and Xu, that is based on quark clustering and results in a stiff EoS. We have investigated the maximum mass of uniformly rotating axisymmetric quark stars. We have also built triaxially deformed solutions for extremely fast rotating quark stars and studied the possible gravitational wave emission from such configurations.
Steep rise of parton densities in the limit of small parton momentum fraction x poses a challenge for describing the observed energy-dependence of the total and inelastic proton-proton cross sections σtot/inelpp : considering a realistic parton spatial distribution, one obtains a too-strong increase of σtot/inelpp in the limit of very high energies. We discuss various mechanisms which allow one to tame such a rise, paying special attention to the role of parton-parton correlations. In addition, we investigate a potential impact on model predictions for σtotpp, related to dynamical higher twist corrections to parton-production process.
The global energy system is undergoing a major transition, and in energy planning and decision-making across governments, industry and academia, models play a crucial role. Because of their policy relevance and contested nature, the transparency and open availability of energy models and data are of particular importance. Here we provide a practical how-to guide based on the collective experience of members of the Open Energy Modelling Initiative (Openmod). We discuss key steps to consider when opening code and data, including determining intellectual property ownership, choosing a licence and appropriate modelling languages, distributing code and data, and providing support and building communities. After illustrating these decisions with examples and lessons learned from the community, we conclude that even though individual researchers' choices are important, institutional changes are still also necessary for more openness and transparency in energy research.
In the last decades, energy modelling has supported energy planning by offering insights into the dynamics between energy access, resource use, and sustainable development. Especially in recent years, there has been an attempt to strengthen the science-policy interface and increase the involvement of society in energy planning processes. This has, both in the EU and worldwide, led to the development of open-source and transparent energy modelling practices.This paper describes the role of an open-source energy modelling tool in the energy planning process and highlights its importance for society. Specifically, it describes the existence and characteristics of the relationship between developing an open-source, freely available tool and its application, dissemination and use for policy making. Using the example of the Open Source energy Modelling System (OSeMOSYS), this work focuses on practices that were established within the community and that made the framework's development and application both relevant and scientifically grounded. Keywords: Energy system modelling tool, Open-source software, Model-based public policy, Software development practice, Outreach practice
Python for Power System Analysis (PyPSA) is a free software toolbox for simulating and optimising modern electrical power systems over multiple periods. PyPSA includes models for conventional generators with unit commitment, variable renewable generation, storage units, coupling to other energy sectors, and mixed alternating and direct current networks. It is designed to be easily extensible and to scale well with large networks and long time series. In this paper the basic functionality of PyPSA is described, including the formulation of the full power flow equations and the multi-period optimisation of operation and investment with linear power flow equations. PyPSA is positioned in the existing free software landscape as a bridge between traditional power flow analysis tools for steady-state analysis and full multi-period energy system models. The functionality is demonstrated on two open datasets of the transmission system in Germany (based on SciGRID) and Europe (based on GridKit).
In energy modelling, open data and open source code can help enhance traceability and reproducibility of model exercises which contribute to facilitate controversial debates and improve policy advice. While the availability of open power plant databases increased in recent years, they often differ considerably from each other and their data quality has not been systematically compared to proprietary sources yet. Here, we introduce the python-based ‘powerplantmatching’ (PPM), an open source toolset for cleaning, standardizing and combining multiple power plant databases. We apply it once only with open databases and once with an additional proprietary database in order to discuss and elaborate the issue of data quality, by analysing capacities, countries, fuel types, geographic coordinates and commissioning years for conventional power plants. We find that a derived dataset purely based on open data is not yet on a par with one in which a proprietary database has been added to the matching, even though the statistical values for capacity matched to a large degree with both datasets. When commissioning years are needed for modelling purposes in the final dataset, the proprietary database helps crucially to increase the quality of the derived dataset.
Use-dependent long-term changes of neuronal response properties must be gated to prevent irrelevant activity from inducing inappropriate modifications. Here we test the hypothesis that local network dynamics contribute to such gating. As synaptic modifications depend on temporal contiguity between presynaptic and postsynaptic activity, we examined the effect of synchronized gamma (ɣ) oscillations on stimulation-dependent modifications of orientation selectivity in adult cat visual cortex. Changes of orientation maps were induced by pairing visual stimulation with electrical activation of the mesencephalic reticular formation. Changes in orientation selectivity were assessed with optical recording of intrinsic signals and multiunit recordings. When conditioning stimuli were associated with strong ɣ-oscillations, orientation domains matching the orientation of the conditioning grating stimulus became more responsive and expanded, because neurons with preferences differing by less than 30° from the orientation of the conditioning grating shifted their orientation preference toward the conditioned orientation. When conditioning stimuli induced no or only weak ɣ-oscillations, responsiveness of neurons driven by the conditioning stimulus decreased. These differential effects depended on the power of oscillations in the low ɣ-band (20 Hz to 48 Hz) and not on differences in discharge rate of cortical neurons, because there was no correlation between the discharge rates during conditioning and the occurrence of changes in orientation preference. Thus, occurrence and polarity of use-dependent long-term changes of cortical response properties appear to depend on the occurrence of ɣ-oscillations during induction and hence on the degree of temporal coherence of the change-inducing network activity.
An incoming or outgoing hadron in a hard collision with large momentum transfer gets squeezed in the transverse direction to its momentum. In the case of nuclear targets, this leads to the reduced interaction of such hadrons with surrounding nucleons which is known as color transparency (CT). The identification of CT in exclusive processes on nuclear targets is of significant interest not only by itself but also due to the fact that CT is a necessary condition for the applicability of factorization for the description of the corresponding elementary process. In this paper we discuss the semiexclusive processes A(e,e′π+) , A(π−,l−l+) and A(γ,π−p) . Since CT is closely related to hadron formation mechanism, the reduced interaction of ’pre-hadrons’ with nucleons is a common feature of generic high-energy inclusive processes on nuclear targets, such as hadron attenuation in deep inelastic scattering (DIS). We will discuss the novel way to study hadron formation via slow neutron production induced by a hard photon interaction with a nucleus. Finally, the opportunity to study hadron formation effects in heavy-ion collisions in the NICA regime will be considered.
Surface color and predictability determine contextual modulation of V1 firing and gamma oscillations
(2019)
The integration of direct bottom-up inputs with contextual information is a core feature of neocortical circuits. In area V1, neurons may reduce their firing rates when their receptive field input can be predicted by spatial context. Gamma-synchronized (30–80 Hz) firing may provide a complementary signal to rates, reflecting stronger synchronization between neuronal populations receiving mutually predictable inputs. We show that large uniform surfaces, which have high spatial predictability, strongly suppressed firing yet induced prominent gamma synchronization in macaque V1, particularly when they were colored. Yet, chromatic mismatches between center and surround, breaking predictability, strongly reduced gamma synchronization while increasing firing rates. Differences between responses to different colors, including strong gamma-responses to red, arose from stimulus adaptation to a full-screen background, suggesting prominent differences in adaptation between M- and L-cone signaling pathways. Thus, synchrony signaled whether RF inputs were predicted from spatial context, while firing rates increased when stimuli were unpredicted from context.
PURPOSE: The purpose of this work is to analyze whether the Monte Carlo codes penh, fluka, and geant4/topas are suitable to calculate absorbed doses and fQ/fQ0 ratios in therapeutic high-energy photon and proton beams.
METHODS: We used penh, fluka, geant4/topas, and egsnrc to calculate the absorbed dose to water in a reference water cavity and the absorbed dose to air in two air cavities representative of a plane-parallel and a cylindrical ionization chamber in a 1.25 MeV photon beam and a 150 MeV proton beam - egsnrc was only used for the photon beam calculations. The physics and transport settings in each code were adjusted to simulate the particle transport as detailed as reasonably possible. From these absorbed doses, fQ0 factors, fQ factors, and fQ/fQ0 ratios (which are the basis of Monte Carlo calculated beam quality correction factors kQ,Q0 ) were calculated and compared between the codes. Additionally, we calculated the spectra of primary particles and secondary electrons in the reference water cavity, as well as the integrated depth-dose curve of 150 MeV protons in water.
RESULTS: The absorbed doses agreed within 1.4% or better between the individual codes for both the photon and proton simulations. The fQ0 and fQ factors agreed within 0.5% or better for the individual codes for both beam qualities. The resulting fQ/fQ0 ratios for 150 MeV protons agreed within 0.7% or better. For the 1.25 MeV photon beam, the spectra of photons and secondary electrons agreed almost perfectly. For the 150 MeV proton simulation, we observed differences in the spectra of secondary protons whereas the spectra of primary protons and low-energy delta electrons also agreed almost perfectly. The first 2 mm of the entrance channel of the 150 MeV proton Bragg curve agreed almost perfectly while for greater depths, the differences in the integrated dose were up to 1.5%.
CONCLUSION: penh, fluka, and geant4/topas are capable of calculating beam quality correction factors in proton beams. The differences in the fQ0 and fQ factors between the codes are 0.5% at maximum. The differences in the fQ/fQ0 ratios are 0.7% at maximum.
An overt pro-inflammatory immune response is a key factor contributing to lethal pneumococcal infection in an influenza pre-infected host and represents a potential target for therapeutic intervention. However, there is a paucity of knowledge about the level of contribution of individual cytokines. Based on the predictions of our previous mathematical modeling approach, the potential benefit of IFN-γ- and/or IL-6-specific antibody-mediated cytokine neutralization was explored in C57BL/6 mice infected with the influenza A/PR/8/34 strain, which were subsequently infected with the Streptococcus pneumoniae strain TIGR4 on day 7 post influenza. While single IL-6 neutralization had no effect on respiratory bacterial clearance, single IFN-γ neutralization enhanced local bacterial clearance in the lungs. Concomitant neutralization of IFN-γ and IL-6 significantly reduced the degree of pneumonia as well as bacteremia compared to the control group, indicating a positive effect for the host during secondary bacterial infection. The results of our model-driven experimental study reveal that the predicted therapeutic value of IFN-γ and IL-6 neutralization in secondary pneumococcal infection following influenza infection is tightly dependent on the experimental protocol while at the same time paving the way toward the development of effective immune therapies.
Classical Hodgkin lymphoma (cHL) is one of the most common malignant lymphomas in Western Europe. The nodular sclerosing subtype of cHL (NS cHL) is characterized by a proliferation of fibroblasts in the tumor microenvironment, leading to fibrotic bands surrounding the lymphoma infiltrate. Several studies have described a crosstalk between the tumour cells of cHL, the Hodgkin- and Reed-Sternberg (HRS) cells, and cancer-associated fibroblasts. However, to date a deep molecular characterization of these fibroblasts is lacking. Thus, the aim of the present study is a comprehensive characterization of these fibroblasts. Gene expression profiling and methylation profiles of fibroblasts isolated from primary lymph node suspensions revealed persistent differences between fibroblasts obtained from NS cHL and lymphadenitis. NS cHL derived fibroblasts exhibit a myofibroblastic phenotype characterized by myocardin (MYOCD) expression. Moreover, TIMP3, an inhibitor of matrix metalloproteinases, was strongly upregulated in NS cHL fibroblasts, likely contributing to the accumulation of collagen in sclerotic bands of NS cHL. As previously shown for other types of cancer-associated fibroblasts, treatment by luteolin could reverse this fibroblast phenotype and decrease TIMP3 secretion. NS cHL fibroblasts showed enhanced proliferation when they were exposed to soluble factors released from HRS cells. For HRS cells, soluble factors from fibroblasts were not sufficient to protect them from Brentuximab-Vedotin induced cell death. However, HRS cells adherent to fibroblasts were protected from Brentuximab-Vedotin induced injury. In summary, we confirm the importance of fibroblasts for HRS cell survival and identify TIMP3 which probably contributes as a major factor to the typical fibrosis observed in NS cHL.
Gravitational waves, electromagnetic radiation, and the emission of high energy particles probe the phase structure of the equation of state of dense matter produced at the crossroad of the closely related relativistic collisions of heavy ions and of binary neutron stars mergers. 3 + 1 dimensional special- and general relativistic hydrodynamic simulation studies reveal a unique window of opportunity to observe phase transitions in compressed baryon matter by laboratory based experiments and by astrophysical multimessenger observations. The astrophysical consequences of a hadron-quark phase transition in the interior of a compact star will be focused within this article. Especially with a future detection of the post-merger gravitational wave emission emanated from a binary neutron star merger event, it would be possible to explore the phase structure of quantum chromodynamics. The astrophysical observables of a hadron-quark phase transition in a single compact star system and binary hybrid star merger scenario will be summarized within this article. The FAIR facility at GSI Helmholtzzentrum allows one to study the universe in the laboratory, and several astrophysical signatures of the quark-gluon plasma have been found in relativistic collisions of heavy ions and will be explored in future experiments.
The graph theoretical analysis of structural magnetic resonance imaging (MRI) data has received a great deal of interest in recent years to characterize the organizational principles of brain networks and their alterations in psychiatric disorders, such as schizophrenia. However, the characterization of networks in clinical populations can be challenging, since the comparison of connectivity between groups is influenced by several factors, such as the overall number of connections and the structural abnormalities of the seed regions. To overcome these limitations, the current study employed the whole-brain analysis of connectional fingerprints in diffusion tensor imaging data obtained at 3 T of chronic schizophrenia patients (n = 16) and healthy, age-matched control participants (n = 17). Probabilistic tractography was performed to quantify the connectivity of 110 brain areas. The connectional fingerprint of a brain area represents the set of relative connection probabilities to all its target areas and is, hence, less affected by overall white and gray matter changes than absolute connectivity measures. After detecting brain regions with abnormal connectional fingerprints through similarity measures, we tested each of its relative connection probability between groups. We found altered connectional fingerprints in schizophrenia patients consistent with a dysconnectivity syndrome. While the medial frontal gyrus showed only reduced connectivity, the connectional fingerprints of the inferior frontal gyrus and the putamen mainly contained relatively increased connection probabilities to areas in the frontal, limbic, and subcortical areas. These findings are in line with previous studies that reported abnormalities in striatal–frontal circuits in the pathophysiology of schizophrenia, highlighting the potential utility of connectional fingerprints for the analysis of anatomical networks in the disorder.
Synesthesia is a phenomenon in which additional perceptual experiences are elicited by sensory stimuli or cognitive concepts. Synesthetes possess a unique type of phenomenal experiences not directly triggered by sensory stimulation. Therefore, for better understanding of consciousness it is relevant to identify the mental and physiological processes that subserve synesthetic experience. In the present work we suggest several reasons why synesthesia has merit for research on consciousness. We first review the research on the dynamic and rapidly growing field of the studies of synesthesia. We particularly draw attention to the role of semantics in synesthesia, which is important for establishing synesthetic associations in the brain. We then propose that the interplay between semantics and sensory input in synesthesia can be helpful for the study of the neural correlates of consciousness, especially when making use of ambiguous stimuli for inducing synesthesia. Finally, synesthesia-related alterations of brain networks and functional connectivity can be of merit for the study of consciousness.