Universitätspublikationen
Refine
Year of publication
- 2021 (27) (remove)
Document Type
- Doctoral Thesis (27) (remove)
Has Fulltext
- yes (27) (remove)
Is part of the Bibliography
- no (27)
Keywords
- 2D materials (1)
- ALICE upgrade (1)
- Astrophysics (1)
- Bohmian mechanics (1)
- CBM Experiment (1)
- Color superconductivity (1)
- Control System (1)
- EPICS (1)
- FEBID (1)
- Gravitational Waves (1)
Institute
- Physik (27) (remove)
Terahertz (THz) technology is an emerging field that considers the radiation between microwave and far-infrared regions where the electronic and photonic technologies merge. THz generation and THz sensing technologies should fill the gap between photonics and electronics which is defined as a region where THz generation power and THz sensing capabilities are at a low technology readiness level (TRL). As one of the options for THz detection technology, field-effect transistors with integrated antennae were suggested to be used as THz detectors in the 1990s by M. Dyakonov and M. Shur from where the development of field-effect transistor-based detector began. In this work, various FET technologies are presented, such as CMOS, AlGaN/GaN, and graphene-based material systems and their further sensitivity enhancement in order to reach the performance of well-developed Schottky diode-based THz sensing technology. Here presented FET-based detectors were explored in a wide frequency range from 0.1 THz up to 5 THz in narrowband and broadband configurations.
For proper implementation of THz detectors, the well-defined characterization is of high importance. Therefore, this work overviews the characterization methods, establishes various definitions of detector parameters, and summarizes the state-of-the-art THz detectors. The electrical, optical, and cryogenic characterization techniques are also presented here, as well as the best results obtained by the development of the characterization methods, namely graphene FET stabilization, low-power THz source characterization for detector calibration, and technology development for cryogenic detection.
Following the discussion about the detector characterization, a wide range of THz applications, which were tested during the last four years of Ph.D. and conducted under the ITN CELTA project from HORIZON2020 program, are presented in this work. The studies began with spectroscopy applications and imaging and later developed towards hyperspectral imaging and even passive imaging of human body THz radiation. As various options for THz applications, single-pixel detectors as well as multi-pixel arrays are also covered in this work.
The conducted research shows that FET-based detectors can be used for spectroscopy applications or be easily adapted for the relevant frequency range. State-of-the-art detectors considered in this work reach the resonant performance below 20 pW/√Hz at 0.3 THz and 0.5 THz, as well as 404 pW/√Hz cross-sectional NEP at 4.75 THz. The broadband detectors show NEP as low as 25 pW/√Hz at around 0.6 THz for the best AlGaN/GaN design and 25 pW/√Hz around 1 THz for the best CMOS design. As one of the most promising applications, metamaterial characterization was tested using the most sensitive devices. Furthermore, one of the single-pixel devices and a multi-pixel array were tested as an engineering solution for a radio astronomy system called GREAT in a stratosphere observatory named SOFIA. The exploration of the autocorrelation technique using FET-based devices shows the opportunity to employ such detectors for direct detection of THz pulses without an interferometric measurement setup.
This work also considers imaging applications, which include near-field and far-field visualization solutions. A considerable milestone for the theory of FET technology was achieved when scanning near-field microscopy led to the visualization of plasma (or carrier density) waves in a graphene FET channel. Whereas another important milestone for the THz technology was achieved when a 3D scan of a mobile phone was performed under the far-field imaging mode. Even though the imaging was done through the phone’s plastic cover, the image displayed high accuracy and good feature recognition of the smartphone, inching the FET-based detector technology ever so close to practical security applications. In parallel, the multi-pixel array testing was carried out on 6x7 pixel arrays that have been implemented in configurable-size aperture and imaging configurations. The configurable aperture size allowed the easier detector focusing procedure and a better fit for the beam size of the incident radiation. The imaging has been tested on various THz sources and compared to the TeraSense 16x16 pixel array. The experimental results show the big advantage of the developed multi-pixel array against the used commercial technology.
Furthermore, two ultra-low-power applications have been successfully tested. The application on hyper-frequency THz imaging tested in the specially developed dual frequency comb and our detector system for 300 GHz radiation with 9 spectral lines led to outstanding imaging results on various materials. The passive imaging of human body radiation was conducted using the most sensitive broadband CMOS detector with a log-spiral antenna working in the 0.1 – 1.5 THz range and reaching the optical NEP of 42 pW/√Hz. The NETD of this device reaches 2.1 K and overcomes the performance limit of passive room-temperature imaging of the human body radiation, which was less than 10 K above the room temperature. This experiment opened a completely new field that was explored before only by the multiplier chain-based or thermal detectors.
...
Classical light microscopy is one of the main tools for science to study small things. Microscopes and their technology and optics have been developed and improved over centuries, however their resolution is ultimately restricted physically by the diffraction of light based on its wave nature described by Maxwell’s equations. Hence, the nanoworld – often characterized by sub-100-nm structural sizes – is not accessible with classical far-field optics (apart from special x-ray laser concepts) since its lateral resolution scales with the wavelength.
It was not until the 20th century that various technologies emerged to circumvent the diffraction limit, including so-called near-field microscopy. Although conceptually based on Maxwell’s long known equations, it took a long time for the scientific community to recognize its powerful opportunities and the first embodiments of near-field microscopes were developed. One representative of them is the scattering-type Scanning Near-field Optical Microscope (s-SNOM). It is a Scanning Probe Microscope (SPM) that enables imaging and spectroscopy at visible light frequencies down to even radio waves with a sub-100-nm resolution regardless of the wavelength used. This work also reflects this wide spectral range as it contains applications from near-infrared light down to deep THz/GHz radiation.
This thesis is subdivided into two parts. First, new experimental capabilities for the s-SNOM are demonstrated and evaluated in a more technical manner. Second, among other things, these capabilities are used to study various transport phenomena in solids, as already indicated in the title.
On the technical side, preliminary studies on the suitability of the qPlus sensor – a novel scanning probe technology – for near-field microscopy are presented.
The scanning head incorporating the qPlus sensor–named TRIBUS – is originally intended and built for ultra-high vacuum, low temperature, and high resolution applications. These are desirable environments and properties for sensitive nearfield measurements as well. However, since its design was not planned for near-field measurements, several special technical and optical aspects have to be taken into account, among others the scanning tip design and a spring suspended measurement head.
In addition, in this thesis field-effect transistors are used as THz detectors in an s-SNOM for the first time. Although THz s-SNOM is already an emerging technology, it still suffers from the requirements of sophisticated and specialized infrastructure on both the detector and laser side. Field-effect transistors offer an alternative that is flexible, cost-efficient, room-temperature operating, and easy to handle. Here, their suitability for s-SNOM measurements, which in general require very sensitive and fast detectors, is evaluated.
In the scientific part of this thesis, electromagnetic surface waves on silver nanowires and the conductivity/charge carrier density in silicon are investigated. Both are completely different concepts of transport phenomena, but this already shows the general versatility of the s-SNOM as it can enter both fields. Silver nanowires are analysed by means of near-infrared radiation. Their plasmonic behaviour in this spectral region is studied complementing other simulations and studies in literature performed on them using for example far-field optics.
Furthermore, the surface wave imaging ability of the s-SNOM in the near-infrared regime is thoroughly investigated in this thesis. Mapping surface waves in the mid-infrared regime is widespread in the community, however for much smaller wavelengths there are several important aspects to be considered additionally, such as the smaller focal spot size.
After that, doped and photo-excited silicon substrates are investigated. As the characteristic frequencies of charge carriers in semiconductors – described by the plasma frequency and the Drude model – are within the THz range, the THz s-SNOM is very well suited to probe their behaviour and to reveal contrasts, which has already been shown qualitatively by numerous literature reports. Here, the photo-excitation enables to set and tune the charge carrier density continuously.
Furthermore, the analysis of all silicon samples focuses on a quantitative extraction of the charge carrier densities and doping levels ...
High-energy astrophysics plays an increasingly important role in the understanding of our universe. On one hand, this is due to ground-breaking observations, like the gravitational-wave detections of the LIGO and Virgo network or the black-hole shadow observations of the EHT collaboration. On the other hand, the field of numerical relativity has reached a level of sophistication that allows for realistic simulations that include all four fundamental forces of nature. A prime example of how observations and theory complement each other can be seen in the studies following GW170817, the first detection of gravitational waves from a binary neutron-star merger. The same detection is also the chronological starting point of this Thesis. The plethora of information and constraints on nuclear physics derived from GW170817 in conjunction with theoretical computations will be presented in the first part of this Thesis. The second part goes beyond this detection and prepares for future observations when also the high-frequency postmerger signal will become detectable. Specifically, signatures of a quark-hadron phase transition are discussed and the specific case of a delayed phase transition is analyzed in detail. Finally, the third part of this Thesis focuses on the inclusion of radiative transport in numerical astrophysics. In the context of binary neutron-star mergers, radiation in the form of neutrinos is crucial for realistic long-term simulations. Two methods are introduced for treating radiation: the approximate state-of-the-art two-moment method (M1) and the recently developed radiative Lattice-Boltzmann method. The latter promises
to be more accurate than M1 at a comparable computational cost. Given that most methods for radiative transport or either inaccurate or unfeasible, the derivation of this new method represents a novel and possibly paradigm-changing contribution to an accurate inclusion of radiation in numerical astrophysics.
Die vorliegende Arbeit präsentiert Forschungsarbeiten basierend auf nanoskopischen Oberflächenmessungen an plasmonischen Metaoberflächen und zweidimensionalen Materialien, insbesondere dem halbleitenden Übergangsmetal-Dichalcogenid (TMDC) WS_2. Die Thesis ist in sieben Kapitel untergegliedert. Die Einleitung vermittelt einen Überblick über die treibenden Kräfte hinter der Forschung im Bereich der Nanophotonik an zweidimensionalen Materialsystemen. Die Untersuchung der Licht-Materie-Wechselwirkung an dünnen Materialgrenzflächen zieht sich als roter Faden durch die gesamte Arbeit.
Das zweite Kapitel beschreibt den experimentellen Aufbau, der für die Durchführung der nanoskopischen Messungen in dieser Arbeit implementiert wurde. Es werden theoretische Grundlagen, das Messprinzip und die Implementierung des optischen Rasternahfeldmikroskops (s-SNOM) skizziert. Außerdem wird ein Strom-Spannungs-Rasterkraftmikroskop (c-AFM) im Kontaktmodus genutzt, um elektrische Ströme auf mikroskopischen zweidimensionalen TMDC-Terrassen zu messen. In den darauffolgenden vier Kapiteln werden die Beiträge dieser Arbeit zur Untersuchung der Licht-Materie-Wechselwirkung auf der Nanoskala aus verschiedenen Perspektiven vorgestellt. Jedes Kapitel enthält eine kurze Einleitung, einen Theorieteil, Messdaten oder Simulationsergebnisse sowie eine Analyse; vervollständigt durch einen Schlussteil.
Die zentrale Arbeit an einer metallischen Metaoberfläche aus elliptischen Goldscheiben wird in Kapitel 3 vorgestellt. Der zugehörige Theorieteil führt in das Konzept von Oberflächen-Plasmon-Polaritonen (SPP) ein, das für den Forschungsbereich der Plasmonik im Allgemeinen wesentlich ist. Verschiedene Methoden zur Berechnung der Dispersionsrelation dieser Oberflächenmoden an ein- und mehrschichtigen Grenzflächen werden auf die untersuchte Metaoberflächenprobe angewendet. Das Modell sagt drei verschiedene Moden voraus, die sich an der Grenzfläche ausbreiten. Eine teil-gebundene ins Substrat abstrahlende Oberflächenmode sowie zwei vergrabene stark gebundene anisotrope Moden. Eine auf der Probe platzierte Nanokugel aus Silizium wird als radiale Anregungsquelle verwendet.
Der Vergleich mit s-SNOM-Nahfeldbildern zeigt, dass nur die schwach gebundene geführte Modenresonanz ausreichend angeregt wurde, um durch s-SNOM-Bildgebung nachgewiesen werden zu können. Die schwache Oberflächenbindung erklärt die scheinbar isotrope Ausbreitung auf der anisotropen Oberfläche. Die Beobachtung der verbleibenden stark eingegrenzten anisotropen vergrabenen Moden würde eine verbesserte tiefenempfindliche Auflösung des Systems erfordern, die im Prinzip für Schichtdicken von 20 nm möglich sein sollte. Darüber hinaus wirft die Beobachtung die Frage auf, ob die durch Impuls- und Modenvolumenanpassung der Nanokugel gegebene Anregungseffizienz einen ausreichenden Anregungsquerschnitt erzeugt, um nachweisbare vergrabene SPP-Moden zu erzeugen.
In Kapitel 4 wird die Idee der Visualisierung vergrabener elektrischer Felder mit s-SNOM fortgesetzt. Hier wird es auf die Untersuchung von WS_2 angewendet, einem zweidimensionalen TMDC-Material, welches Photolumineszenz zeigt. Durch die Strukturierung des Galliumphosphid-Substrats unter der hängenden Monolage, die von einer dünnen Schicht aus hBN getragen wird, wird die Photolumineszenzausbeute um den Faktor 10 erhöht. Dies wird durch den Entwurf einer lateralen DBR-Mikrokavität mit zusätzlich optimierter vertikaler Tiefe erreicht, die in das Substrat geätzt wurde.
Die hochauflösende Abbildung der elektrischen Feldverteilung im Resonator wird durch den Einsatz von s-SNOM ermöglicht, um die Verbesserung der Einkopplung durch diese beiden Ansätze zu bewerten. Es konnte festgestellt werden, dass die laterale Struktur überwiegend zur verstärkten Photolumineszenzausbeute beiträgt, während für die Einkopplung keine offensichtliche Verstärkung auf die vertikale Strukturoptimierung zurückgeführt werden konnte.
Das zweidimensionale Material WS_2 wird in Kapitel 5 erneut mit Hilfe von c-AFM untersucht. Unterschiedlich dicke Multilagen auf Graphen und Gold dienen als Tunnelbarrieren für vertikale Ströme zwischen Substrat und leitender c-AFM-Messpitze. Die Daten können mit einem Fowler-Nordheim-Modell mit Parametern für die Tunnelbreite und Schottky-Barrierenhöhen der beiden Grenzflächen erklärt werden. Die Messungen zeigen jedoch eine schwache Reproduzierbarkeit, was eine detailliertere Zusammenfassung der relevanten Fehlerquellen erfordert. In der Schlussfolgerung des Kapitels werden mehrere Schlüsselaspekte vorgeschlagen, die bei künftigen Messungen berücksichtigt werden sollten. Entscheidend ist, dass c-AFM sehr empfindlich auf die Adsorption von Wasserfilmen an der Probenoberfläche reagiert, worunter WS_2-Oberflächen unter Umgebungsbedingungen leiden...
Die vorliegende Dissertation stellt die Strahldynamikdesigns zweier Hochfrequenzquadrupol-Linearbeschleuniger bzw. Radio Frequency Quadrupoles (RFQs) vor: das fur den RFQ des Protonen-Linearbeschleunigers (p-Linac) des FAIR2-Projekts an der GSI3 Darmstadt sowie einen ersten Designentwurf für einen kompakten RFQ, der u.a. zur Erzeugung von Radioisotopen für medizinische Zwecke genutzt werden könnte. Der Schwerpunkt liegt auf dem ersten Design.
The topic of this thesis is the theoretical description of the hadron gas stages in heavy-ion collisions. The overall addressed question hereby is: How does the hadronic medium evolve i.e. what are the relevant microscopic reaction mechanisms and the properties of the involved degrees of freedom? The main goal is to address this question specifically for hadronic multi-particle interactions. For this goal, the hadronic transport approach SMASH is extended with stochastic rates, which allow to include detailed balance fulfilling multi-particle reactions in the approach. Three types of reactions are newly-accounted for: 3-to-1, 3-to-2 and 5-to-2 reactions. After extensive verifications of the stochastic rates approach, they are used to study the effect of multi-particle interactions, particularly in afterburner calculations.
These studies follow complementary results for the dilepton and strangeness production with only binary reactions, which show that hadronic transport approaches are capable of describing observables when employed for the entire evolution of low-energy heavy-ion collisions. This is illustrated by the agreement of dilepton and strangeness production for smaller systems with SMASH calculations. It is, in particular, possible to match the measured strangeness production of phi and Xi hadrons via additional heavy nucleon resonance decay channels. For larger systems or higher energies, hadronic transport cascade calculations with vacuum resonance properties can point to medium effects. This is demonstrated extensively for the dilepton emission in comparisons to the full set of HADES dielectron data. The dilepton invariant mass spectra are sensitive to a medium modification of the vector meson spectral function for large collision systems already at low beam energies. The sensitivity to medium modifications is mapped out in detail by comparisons to a coarse-graining approach, which employs medium-modified spectral functions and is based on the same evolution.
The theoretical foundation of stochastic rates are collision probabilities derived from the Boltzmann equation's collision term with the assumption of a constant matrix element. This derivation is presented in a comprehensive and pedagogical fashion. The derived collision probabilities are employed for a stochastic collision criterion and various detailed-balance fulfilling multi-particle reactions: the mesonic Dalitz decay back-reaction (3-to-1), the deuteron catalysis (3-to-2) and the proton-antiproton annihilation back-reaction (5-to-2). The introduced stochastic rates approach is extensively verified by studies of the numerical stability and comparisons to previous results and analytic expectations. The stochastic rates results agree perfectly with the respective analytic results.
Physically, multi-particle reactions are demonstrated to be significant for different observables, most notably the yield of the partaking particles, even in the late dilute stage of heavy-ion reactions. They lead to a faster equilibration of the system than equivalent binary multi-step treatments. The difference in equilibration consequently influences the yield in afterburner calculations. Interestingly, the interpretation of results is not dependent on employing multi-particle or multi-step treatments, which a posteriori validates the latter.
As the first test case of multi-particle reactions in heavy-ion reactions, the mesonic 3-to-1 Dalitz decay is found to be dominated by the omega Dalitz decay back-reaction. While the effect on the medium is found to be negligible overall, the regeneration is found to be sizable: up to a quarter of Dalitz decays are regenerated.
Non-equilibrium rescattering effects are shown to be relevant for late collision stages for two particle species: deuteron and protons. In both cases, the relevant rescatterings involve multiple particles.
The deuteron pion and nucleon catalysis reactions equilibrate quickly in the afterburner stage at intermediate energies. The constant formation and destruction keeps the yield constant and microscopically explains the "snowballs in hell"-paradox. The yield is also generated with no d present at early times, which explains why coalescence models can also match the multiplicity.
New is the study of the 5-body back-reaction of proton-antiproton annihilations. This work marks the first realization of microscopic 5-body reactions in a transport approach to fulfill detailed balance for such reactions. A sizable regeneration due to the back-reaction of up to half of the proton-antiproton pairs lost due to annihilations is found. Consequently, both annihilation and regeneration in the late non-equilibrium stage are shown to have a significant effect on the p yield.
Ziel dieser Dissertation ist es, die Gleichgewichts- und Nichtgleichgewichts-Eigenschaften des stark wechselwirkenden QGP-Mediums nahe dem Phasenübergang unter extremen Bedingungen von hohen T und hohen Baryonendichten mit Hilfe der kinetischen Theorie im Rahmen von effektiven Modellen zu untersuchen. Wir werden zunächst die thermodynamischen und Transporteigenschaften des QGPs in der Nähe des Gleichgewichts auf der Basis des DQPM im Bereich moderater chemischer Baryonenpotentiale μB ≥ 0.5 GeV untersuchen. Insbesondere werden die EoS und die Schallgeschwindigkeit sowie die Transportkoeffizienten des QGP auf der Grundlage des DQPM bei endlichen T und μB berechnet. Transportkoeffizienten sind besonders interessant, da sie Informationen über die Wechselwirkungen im Medium erlauben, das im Gleichgewicht durch eine Temperatur T und ein chemisches Potential μB charakterisiert werden kann. Unter Berücksichtigung der Transportkoeffizienten und der EoS der QGP-Phase vergleichen wir unsere Ergebnisse mit verschiedenen Resultaten aus der Literatur, in denen Transportkoeffizienten des QGPs auf Basis von effektiven Modellen vorwiegend bei Null oder kleinem chemischen Potentialen untersucht wurden.
Darüber hinaus werden in Kapitel 3 die Gleichgewichtseigenschaften des QGPs und insbesondere die Auswirkungen der μB-Abhängigkeit der thermodynamischen und Transporteigenschaften des QGPs im Rahmen des erweiterten PHSD-Transportansatzes untersucht, der die vollständige Entwicklung des Systems einschließlich der partonischen Phase umfasst. Die Entwicklung des PHSD-Transportansatzes wird in der partonischen Phase erweitert, indem explizit die gesamt- und differentiellen partonischen Streuquerschnitte auf der Grundlage des DQPM berechnet und bei der tatsächlichen Temperatur T und dem baryonischen chemischen Potential μB in jeder einzelnen Raum-Zeit-Zelle, in der die partonische Streuung stattfindet, ausgewertet werden.
Um die Spuren der μB-Abhängigkeit des QGPs in den Observablen zu untersuchen, werden die Ergebnisse von PHSD5.0 (mit μB-Abhängigkeiten) mit den Ergebnissen von PHSD5.0 für μB = 0 sowie mit PHSD4.0, in dem die Massen/Breiten der Quarks und Gluonen sowie deren Wechselwirkungsquerschnitte nur von T abhängen, verglichen. Wir diskutieren die PHSD-Ergebnisse für verschiedene Observablen: (i) Rapiditäts- und pT -Verteilungen von identifizierten Hadronen für symmetrische Au+Au- und Pb+Pb- Kollisionen bei Energien von 30 AGeV (zukünftige NICA-Energie) sowie für die RHIC-Spitzenenergie von √sNN = 200 GeV; (ii) gerichteter Fluss v1 von identifizierten Hadronen für Au + Au bei invarianter Energie √sNN = 27 GeV und 200 GeV; (iii) elliptischer Fluss v2 der identifizierten Hadronen für Au+Au bei invarianten Energien √sNN = 27 und 200 GeV. Der Vergleich der "Bulk"-Observablen für Au+Au-Kollisionen innerhalb der drei PHSD-Einstellungen hat gezeigt, dass sie eine recht geringe Empfindlichkeit gegenüber den μB -Abhängigkeiten der Partoneigenschaften (Massen und Breiten) und ihrer Wechselwirkungsquerschnitte aufweisen, sodass die Ergebnisse von PHSD5.0 mit und ohne μB sehr nahe beieinander liegen. Nur im Fall von Kaonen, Antiprotonen ̄p und Antihyperonen ̄Λ + ̄Σ0 konnte ein kleiner Unterschied zwischen PHSD4.0 und PHSD5.0 bei den höchsten SPS- und RHIC-Energien festgestellt werden.
Wir finden nur geringe Unterschiede zwischen den Ergebnissen von PHSD4.0 und PHSD5.0 für die hier betrachteten hadronischen Observablen sowohl bei hohen als auch bei mittleren Energien. Dies hängt damit zusammen, dass bei hohen Energien, wo die Materie vom QGP dominiert wird, ein sehr kleines chemisches Baryonenpotential μB in zentralen Kollisionen bei mittlerer Rapidität gemessen wird, während mit abnehmender Energie und größerem μB der Anteil des QGPs rapide abnimmt, sodass die endgültigen Beobachtungswerte insgesamt von den Hadronen dominiert werden, die an der hadronischen Rückstreuung teilgenommen haben, und somit die Information über ihren QGP-Ursprung verwaschen oder verloren geht.
In Kapitel 4 betrachten wir die Transportkoeffizienten von QGP-Materie im erweiterten Polyakov-NJL-Modell entlang der Übergangslinie für moderate Werte des chemischen Baryonenpotenzials 0 ≤ μB ≤ 0.9 GeV sowie in der Nähe des kritischen Endpunkts(CEP) und bei großem chemischen Baryonenpotenzial μB = 1.2 GeV, wo ein Phasenübergang erster Ordnung stattfindet. Wir untersuchen, wie die Natur der Freiheitsgrade die Transporteigenschaften des QGPs beeinflusst. Darüber hinaus demonstrieren wir die Auswirkungen des Phasenübergangs erster Ordnung und des CEP auf die Transportkoeffizienten im dekonfinierten QCD-Medium.
Darüber hinaus wird in Kapitel 5 eine phänomenologische Erweiterung des DQPM auf große baryonchemische Potentiale μB einschließlich der Region mit einem möglichen CEP und späterem Phasenübergang erster Ordnung betrachtet. Eines der wichtigsten Merkmale des Modells ist das Auftreten einer ’kritischen‘ Skalierung in der Nähe des CEP. Das Hauptziel des vorgestellten Modells besteht darin, die mikroskopischen und makroskopischen Eigenschaften der partonischen Freiheitsgrade für den Bereich des Phasendiagramms bereitzustellen, der durch moderates T und moderates oder hohes μB gekennzeichnet ist.
...
The main subject of this thesis is the study of hadron and photon production in relativistic heavy-ion collisions by means of hydrodynamics+transport approaches. Two different kinds of such hybrid approaches are employed in this work, the SMASH-vHLLE-hybrid and a MUSIC+SMASH hybrid. While the former is capable of simulating heavy-ion collisions covering a wide range of collision energies down to √s = 4.3 GeV, reproducing the correct baryon stopping powers, the latter provides a framework to consistently model photon production in the hadronic stage of high-energy heavy-ion collisions.
The SMASH-vHLLE-hybrid is a novel state-of-the-art hybrid approach whose development constitutes a major contribution to this thesis. It couples the hadronic transport SMASH to the 3+1D viscous hydrodynamics approach vHLLE. Therein, SMASH is employed to provide the fluctuating 3D initial conditions and to model the late hadronic rescattering stage, and vHLLE for the fluid dynamical evolution of the hot and dense fireball. The initial conditions are provided on a hypersurface of constant proper time, and the macroscopic evolution of the fireball is carried out down to an energy density of ecrit = 0.5 GeV/fm3, where particlization occurs. Consistency at the interfaces is verified in view of global, on-average quantum number conservation and the SMASH-vHLLE-hybrid is validated by comparison to SMASH+CLVisc as well as UrQMD+vHLLE hybrid approaches. The establishment of the SMASH-vHLLE-hybrid to theoretically describe heavy-ion collisions at intermediate and high collision energies forms a basis for a range of extensions and future research projects. It is further made available to the heavy-ion community by virtue of being published on Github.
The SMASH-vHLLE-hybrid is applied to simulate Au+Au/Pb+Pb collisions between √s = 4.3 GeV and √s = 200.0 GeV. A good agreement with the experimentally measured rapidity and transverse mass spectra is obtained. In particular the baryon stopping dynamics are well reproduced at low, intermediate, and high collision energies. Excitation functions for the mid-rapidity yield and mean transverse momentum of pions, protons and kaons are demonstrated to agree well with their experimentally measured counterpart. These results further validate the approach and provide a solid baseline for potential future studies. The importance of annihilations and regenerations of protons and anti-protons is additionally investigated in Au+Au/Pb+Pb collisions between √s = 17.3 GeV and √s = 5.02 TeV with the SMASH-vHLLE-hybrid. It is found that, regarding the p + p ̄ ↔ 5 π reaction, 20-50% (depending on the rapidity range) of the (anti-)proton yield lost to annihila- tions in the hadronic rescattering stage is restored owing to the back reaction. The back reaction thus constitutes a non-negligible contribution to the final (anti-)proton yield and should not be neglected when modelling the late rescattering stage of heavy-ion collisions.
The MUSIC+SMASH hybrid is a hybrid approach ideally suited to model the production of photons in relativistic heavy-ion collisions. Therein, the macroscopic production of photons in the hadronic stage in MUSIC relies on the identical effective field theories as the photon cross sections implemented in SMASH for the microscopic production. The MUSIC+SMASH hybrid thus provides the first consistent framework to the end of hadronic photon production. It accounts for 2 → 2 scattering processes of the kind π + ρ → π + γ and pion bremsstrahlung processes π + π → π + π + γ. The MUSIC+SMASH hybrid is employed in an ideal 2D setup to systematically assess the importance of non-equliibrium dynamics in the hadronic rescattering stage on mid-rapidity transverse momentum spectra and elliptic flow of photons at RHIC/LHC energies. This is achieved by comparing the outcome of the MUSIC+SMASH hybrid, involving an out-of-equilibrium late rescattering stage, to macroscopically approximating late stage photon production by means of MUSIC, employed down to temperatures well below the switching temperature. It is found that non-equilibrium dynamics have only minor implications for photon transverse momentum spectra, but significantly enhance the photon elliptic flow. At RHIC energies, an enhancement of up to 70%, and at LHC of up to 65% is observed in the non-equilibrium afterburner as compared to its hydrodynamical counterpart. In combination with the large amount of photons produced above the particlization temperature, these differences are modest regarding the transverse momentum spectra, but a significant enhancement of the elliptic flow is observed at low transverse momenta. Below pT ≈ 1.4 GeV, the combined v2 is enhanced by up to 30% at RHIC, and up to 20% at the LHC within the non-equilibrium setup as compared to its approximation via hydrodynamics. Non-equilibrium dynamics in the hadronic rescattering stage are hence important, especially in view of momentum anisotropies at low transverse momenta. These findings thus contribute to the understanding of low-pT photons produced in heavy-ion collisions at RHIC/LHC energies and the MUSIC+SMASH hybrid employed for this study provides a baseline for additional studies regarding photon production in the future.
To summarize, the approaches and frameworks presented in this thesis provide a good baseline for further extensions and studies in order to improve the understanding of hadron and photon production in relativistic heavy-ion collisions across a wide range of collision energies. More broadly, such future studies of hadrons and photons may contribute to enhance the understandig of the properties of the fundamental building blocks of matter, of which everything that surrounds us is made of.
This dissertation describes the development of the beam dynamics design of a novel superconducting linear accelerator. At a main operating frequency of 216.816 MHz, ions with a mass-to-charge ratio of up to 6 can be accelerated at high duty cycles up to CW operation. Intended for construction at the GSI Helmholtz Centre for Heavy Ion Research in Darmstadt, the focus of the work is on the beam dynamic design of the accelerator section downstream of the high charge injector (HLI) at an injection energy of 1.39 MeV/u. An essential feature of this linear accelerator (Linac) is the use of the EQUUS (Equidistant Multigap Structure) beam dynamics concept for a variably adjustable output energy between 3.5 and 7.3 MeV/u (corresponding to about 12.4 % of the speed of light) with a required low energy spread of maximum 3 keV/u.
The GSI Helmholtz Centre for Heavy Ion Research is a large-scale research facility that uses its particle accelerators to perform basic research with ion beams. Research on super-heavy elements ("SHE") is a major focus. It is expected that their production and research will provide answers to a large number of scientific questions. The production and detection of elements with atomic numbers 107 to 112 (Bohrium, Hassium, Meitnerium, Darmstadtium, Röntgenium and Copernicium) was first achieved at GSI between 1981 and 1996.
Key to this remarkable progress in SHE research were continuous developments and technical innovations. On the one hand, in the field of experimental sensitivity and detection of the nuclear reaction products and, on the other hand, in the field of accelerator technology.
For the acceleration of the projectile beam, the UNILAC (Universal Linear Accelerator), which was put into operation in 1975, has been used at GSI so far. In the course of the reconstruction and expansion of the research infrastructure at GSI, a dedicated new particle accelerator, HELIAC (Helmholtz Linear Accelerator), is now under development to meet the special requirements of the beam parameters for the synthesis of new superheavy elements. Typically, the production rates of super-heavy elements with effective cross sections in the picobarn range are very low. Therefore, a high duty cycle (up to CW operation) is a key feature of HELIAC. Thus, the required beam time for the desired nuclear reactions can be significantly shortened.
Theoretical preliminary work by Minaev et al. and newly created knowledge about design, fabrication, and operation of superconducting drift tube cavities have laid the foundation for this work and thus the development of the HELIAC linear accelerator. It consists of a superconducting and a normal conducting part. Acceleration takes place in the superconducting part in four cryomodules, each about 5 m long. These contain three CH cavities, one buncher cavity, two solenoid magnets for transverse beam focusing, and two beam position monitors (BPMs).
The following 10 m long normal conducting part is primarily used for beam transport and ends with a buncher cavity. This is operated at a halved frequency of 108.408 MHz.
A key feature of this accelerator is the variability of the output energy from 3.5 to 7.3 MeV/u with a small energy uncertainty of ±3 keV/u maximum over the entire output energy range. For the development of HELIAC, the EQUUS beam dynamics concept used combined the advantages of conventional linac designs with the high acceleration gradients of superconducting CH-DTLs. By doubling the frequency (compared to the GSI high charge injector) to 216.816 MHz in the superconducting section and using CH cavities at an acceleration gradient of maximum 7.1 MV/m, an acceleration efficiency with superconducting drift tube structures that is unique in the world is made possible. At the same time, the compact lengths of the CH cavities ensure good handling for both production and operation. EQUUS leads to longitudinal beam stability in all energy ranges of the accelerator with the sliding motion of the synchronous phase within each CH cavity. The rms emittance growth is moderate in all levels. The modular design of the HELIAC with four cryomodules basically allows the Linac to be commissioned starting with the first cryomodule, the so-called Advanced Demonstrator. In the subsequent expansion stage with only the first two cryomodules of HELIAC, the lower limit of the energy range to be provided by HELIAC (3.5 MeV/u) can already be clearly exceeded, so that use in regular beam operation at GSI is already conceivable from here on.
By means of error tolerance studies, the stability of the HELIAC beam dynamics design against possible alignment errors of the magnetic focusing elements and accelerator cavities as well as errors of the electric field amplitudes and phases have been investigated, basically confirmed and critical parameters have been determined. An additional steering concept via dipole correction coils at the solenoid magnets allows transverse beam control as well as diagnostics by means of two BPMs per cryomodule.
With completion of this work in 2021, the CH1 and CH2 cavities have already been built and are in the final preparation and cold test phase. In parallel, the development of the CH cavities CH3-11 has also been started.
For finite baryon chemical potential, conventional lattice descriptions of quantum chromodynamics (QCD) have a sign problem which prevents straightforward simulations based on importance sampling.
In this thesis we investigate heavy dense QCD by representing lattice QCD with Wilson fermions at finite temperature and density in terms of Polyakov loops.
We discuss the derivation of $3$-dimensional effective Polyakov loop theories from lattice QCD based on a combined strong coupling and hopping parameter expansion, which is valid for heavy quarks.
The finite density sign problem is milder in these theories and they are also amenable to analytic evaluations.
The analytic evaluation of Polyakov loop theories via series expansion techniques is illustrated by using them to evaluate the $\SU{3}$ spin model.
We compute the free energy density to $14$th order in the nearest neighbor coupling and find that predictions for the equation of state agree with simulations to $\mathcal{O}(1\%)$ in the phase were the (approximate) $Z(3)$ center symmetry is intact.
The critical end point is also determined but with less accuracy and our results agree with numerical results to $\mathcal{O}(10\%)$.
While the accuracy for the endpoint is limited for the current length of the series, analytic tools provide valuable insight and are more flexible.
Furthermore they can be generalized to Polyakov-loop-theories with $n$-point interactions.
We also take a detailed look at the hopping expansion for the derivation of the effective theory.
The exponentiation of the action is discussed by using a polymer expansion and we also explain how to obtain logarithmic resummations for all contributions, which will be achieved by employing the finite cluster method know from condensed matter physics.
The finite cluster method can also be used to evaluate the effective theory and comparisons of the evaluation of the effective action and a direction evaluation of the partition function are made.
We observe that terms in the evaluation of the effective theory correspond to partial contractions in the application of Wick's theorem for the evaluation of Grassmann-valued integrals.
Potential problems arising from this fact are explored.
Next to next to leading order results from the hopping expansion are used to analyze and compare the onset transition both for baryon and isospin chemical potential.
Lattice QCD with an isospin chemical potential does not have a sign problem and can serve as a valuable cross-check.
Since we are restricted by the relatively short length of our series, we content ourselves with observing some qualitative phenomenological properties arising in the effective theory which are relevant for the onset transition.
Finally, we generalize our results to arbitrary number of colors $N_c$.
We investigate the transition from a hadron gas to baryon condensation and find that for any finite lattice spacing the transition becomes stronger when $N_c$ is increased and to be first order in the limit of infinite $N_c$.
Beyond the onset, the pressure is shown to scale as $p \sim N_c$ through all available orders in the hopping expansion, which is characteristic for a phase termed quarkyonic matter in the literature.
Some care has to be taken when approaching the continuum, as we find that the continuum limit has to be taken before the large $N_c$ limit.
Although we currently are unable to take the limits in this order, our results are stable in the controlled range of lattice spacings when the limits are approached in this order.
Neurons are cells with a highly complex morphology; their dendritic arbor spans up to thousands of micrometers. This extended arbor poses a challenge for the logistics of neuronal processes: mRNA, proteins, and organelles have to be transported to dendrites, hundreds of micrometers away from the soma. This thesis aims to calculate the minimum number of proteins needed to populate the dendritic trees for different scenarios.
In chapter 2, I analyzed the ability of different mechanisms to populate the dendritic arbor. I started from the solution of the diffusion equation in Sec. 2.1, then I included the contribution of active transport in Sec. 2.2 and showed how it could have either the effect of increasing the effective diffusion coefficient or of introducing a bias in the diffusion process. In Sec. 2.3 I studied the spatial distribution of locally synthesized protein, accordingly with actively and passively transported mRNA. In Sec. 2.5, I derived the boundary condition for branches showing a qualitatively different behavior of surface and cytoplasmic proteins induced by the medium’s dimensionality in which they diffuse.
In chapter 3, I introduced the concept of protein requirement, defined as the minimum number of proteins that the neuron needs to produce to provide at least one protein to each micrometer of the dendritic arbor. In Sec. 3.1, I derived the protein requirement for diffusive proteins for somatic translation and constant translation in the dendritic arbor. In Sec. 3.2, I analyzed numerically the protein requirement in the case of actively transported protein synthesized in the soma, and, in Sec. 3.3, in the case of actively transported proteins synthesized in the dendritic arbor. In Sec. 3.4, I analyzed the protein requirement of protein synthesized in the dendrite accordingly with the distribution of mRNA described in Sec. 3.3 and 3.2. In Sec. 3.5, I derived the protein requirement for a single branch and purely diffusive proteins.
In chapter 4, I analyzed the relation between the radii of the three afferent dendrites in a branch, their length, and the diffusion length of a protein. In Sec. 4.1 I derived the optimal ratio between the radii of the daughter dendrites that minimizes the protein requirement. In Sec. 4.3 I introduced the 3/2− Rall Rule and in Sec. 4.5 its generalization. Finally, I used those rules to estimate the fraction of proteins diffusing away from and toward the soma.
In chapter 5, I analyzed the radii distribution for three categories of neurons: cultured hippocampal neurons in Sec. 5.1, stomatogastric ganglia neuron in Sec. 5.2, and 3DEM reconstructed prefrontal pyramidal neurons in Sec. 5.3. For each of these three classes, I analyzed the distribution of radii, Rall exponents, and the probability ratio. For most of them, I found that the probability of a protein diffusing away from the soma is higher for surface proteins than for cytoplasmic ones. I quantified this with a parameter called surface bias.
In Chapter 6, I analyzed the fluorescent ratio imaged by our collaborators Anne-Sophie Hafner, for a surface protein, GFP::Nlg, and a soluble one, GFP, in cultured hippocampal neurons, and I compared the fluorescent ratio with the probability ratio obtained in 5.1, finding that they are in good agreement.
In chapter 7, I compared the real dendritic morphologies imaged by one of our collaborators Ali Karimi with the optimal branching rule obtained in Sec. 4.1 and I calculated the cost for not having optimal branching radii.
Finally, in Chapter 8, I used the knowledge of the branching statistics gathered in 5.3 to simulate the protein profile on three different classes of neurons: pyramidal neurons, granule neuron, and Purkinje neurons. I compared the protein profile for surface and cytoplasmic neurons for each morphology for two different values of the diffusion length: λ = 109µm and λ = 473µm, both for optimized radii and symmetrical radii. I showed how the radii optimization reduces the protein requirement of a factor 10 4 for pyramidal neurons.
The first part of this work addresses the automatic online tuning of transfer lines in particle accelerator facilities. In the second part the focus lies on the automatic construction and optimisation of such transport lines. It can be shown that genetic algorithms can be used very well for optimisation in both cases. Automatic online tuning can be performed very efficiently at accelerators under certain boundary conditions and is particularly well suited for initial beam commissioning with low intensity pilot beams. The construction of transfer lines can also be formulated and solved as an minimisation problem with an adopted parameterisation. Thereby, both the imaging properties of the beam transport and the robustness against error studies can be optimised at the same time.
The main focus of research in the field of high-energy heavy-ion physics is the study of the quark-gluon plasma (QGP). Topic of the present work is the measurement of electron-positron pairs (dielectrons), which grant direct access to some of the key properties of this state of matter, since after their formation they leave the hot and dense medium without significant interaction. In particular, the measurement of the initial QGP temperature is considered a "holy grail" of heavy-ion physics. Therefore, in addition to the analysis of existing data, a feasibility study has been conducted to determine to which extent this goal would be achievable by upgrading the ALICE experiment at CERN.
Dielectrons are produced during all stages of a heavy-ion collision, with their invariant mass reflecting the amount of energy available at the time of their formation. Dielectrons of highest mass are thus produced in the initial scatterings of the colliding nuclei by quark-antiquark annihilation. Correlated electron-positron pairs can also emerge from the decay chains of early-produced pairs of heavy-flavour (HF) particles. During the QGP stage and at the beginning of the hadronic phase, the system emits thermal radiation in the form of photons and dielectrons, which carry information about the medium temperature to the observer. In the final stage of the collision, decays of light-flavour (LF) hadrons produce additional contributions to the dielectron spectrum.
The present work is based on early data from the ALICE experiment recorded from lead-lead collisions at a center-of-mass energy of 2.76 TeV. Due to the limited amount of data, a focus is placed on achieving high efficiencies throughout the analysis. To this end, a special electron identification strategy is developed and a custom track selection applied, together resulting in a tenfold increase in pair efficiency. The dielectron spectrum is evaluated on a statistical basis, using a pair prefilter, which is optimized based on two signal quality criteria, to reduce the fraction of electrons and positrons from unwanted sources at minimum signal loss. In addition, an artifact of the track reconstruction is exploited to suppress pairs from photon conversions and to correct the dielectron yield for a contribution from different-conversion pairs. The main signal uncertainty is extracted from the deviation between results of 20 analysis settings and amounts to 20% in most of the studied kinematic range.
For comparison with the analysis results, a hadronic cocktail consisting of the LF and HF contributions is simulated, which can reasonably well describe the measured dielectron production, with a hint of an enhancement at low invariant mass. Two approaches to model the in-medium modification of the heavy-flavour are followed, resulting in up to 50% suppression, which creates some additional space for a thermal contribution at intermediate mass.
For a complete comparison between experimental data and theoretical expectation, two model calculations are consulted. The Thermal Fireball Model provides predictions for thermal dielectron radiation from the QGP and hadron gas. The data tends to be better described with these additional thermal contributions. For a comparison with a prediction by the UrQMD model, the HF component of the cocktail is subtracted from the data. This results in better agreement if the HF suppression by in-medium effects is taken into account.
The feasibility study in this work has served as a physical motivation for the ALICE upgrade for LHC Run 3. The precision with which the early temperature of the QGP can be determined via dielectrons is chosen as key observable. A multitude of individual contributions are merged into a fully modeled dielectron analysis. The resulting signal-to-background ratio represents some of the expected systematic uncertainties, while from the significance combined with the planned number of lead-lead collisions a realistic "measurement" with statistical fluctuations around the expected dielectron signal is generated using a Poisson sampling technique. Since the HF yield exceeds the QGP thermal radiation by about an order of magnitude, an additional analysis step exploiting the enhanced track reconstruction is introduced to reduce its contribution by up to a factor of five. The resulting reduction in pair efficiency is overcompensated by an up to hundred times higher collision rate. The entire cocktail is then subtracted from the sampled data to isolate the thermal excess yield. The final analysis of this spectrum shows that the inverse slope of the model prediction, which depends directly on the QGP temperature, can be reproduced within statistical and systematic uncertainties of about 10%.
The promising results of this study have contributed on the one hand to the realization of the ALICE upgrade and to a design decision for the new Inner Tracking System, and at the same time represent exciting predictions for upcoming measurements.
An investigation of photoelectron angular distributions and circular dichroism of chiral molecules
(2021)
The present work demonstrates the capability of several type of molecular frame photoelectron angular distributions (MFPADs) and their linked chiroptical phenomenon the photoelectron circular dichroism (PECD) to map in great detail the molecular geometry of polyatomic chiral molecules as a function of photoelectron energy. To investigate the influence of the molecular potential on the MFPADs, two chiral molecules were selected, namely 2-(methyl)oxirane (C3H6O, MOx, m = 58,08 uma) and 2-(trifluoromethyl)oxirane (C3H3F3O, TFMOx, m = 112,03 uma). The two molecules differs in one substitutional group and share an oxirane group where the O(1s) electron was directly photoionized with the use of synchrotron radiation in the soft X-ray regime. The direct photoionization of the K-shell electron is well localized in the molecule and it induces the ejection of two or more electrons; the excited system separates into several charged (and eventually neutral) fragments which undergo Coulomb explosion due to their charges. The electrons and the fragments were detected using the COLd Target Recoil Ion Momentum Spectroscopy (COLTRIMS) and the momentum vectors calculated for each fragment belonging from a single ionization. The former method gives the possibility to post-orient molecules in space, giving access to the molecular frame, thus the MFPAD and its related PECD for multiple light propagation direction.
Stereochemistry (from the Greek στερεο- stereo- meaning solid) refers to chemistry in three dimensions. Since most molecules show a three-dimensional structure (3D), stereochemistry pervades all fields of chemistry and biology, and it is an essential point of view for the understanding of chemical structure, molecular dynamics and molecular reactions. The understanding of the chemistry of life is tightly bounded with major discoveries in stereochemistry, which triggered tremendous technical advancements, making it a flourishing field of research since its revolutionary introduction in late 18th century. In chemistry, chirality is a brunch of stereochemistry which focuses on objects with the peculiar geometrical property of not being superimposable to their mirror-images. The word chirality is derived from the Greek χειρ for “hand”, and the first use of this term in chemistry is usually attributed to Lord Kelvin who called during a lecture at the Oxford University Junior Scientific Club in 1893 “any geometrical figure, or group of points, “chiral”, and say that it has chirality if its image in a plane mirror, ideally realized, cannot be brought to coincide with itself.”. Although the latter is usually considered as the birth of the word chirality, the concept underlying it was already present in several fields of science (above all mathematics), already proving the already multidisciplinary relevance of chirality across many field of science and beyond. Nature shows great examples of chiral symmetry on all scales. Empirically, it is possible to observe it at macroscopic scale (e.g. distribution of rotations of galaxies), down to the microscopic scale (e.g. structure of some plankton species), but it is at the molecular level where the number gets remarkable: most of the pharmaceutical drugs, food fragrances, pheromones, enzymes, amino acids and DNA molecules, in fact, are chiral. Moreover, the concept of chirality goes far beyond the mere spatial symmetry of objects being crucially entangled with the fundamental properties of physical forces in nature. The symmetry breaking, namely the different physical behaviour of a two chiral systems upon the same stimuli, is considered to be one of the best explanation for the long standing questions of homochirality in biological life, and ultimately to the chemical origin of life on Earth as we know it. Our organism shows high enantio-selectivity towards specific compounds ranging from drugs, to fragrances. Over 800 odour molecules commonly used in food and fragrance industries have been identified as chiral and their enantiomeric forms are perceived to have very different smells, as the well-know example of D- and L- limonene. Similarly, responses to pharmaceuticals drugs can be enantiomer specific, and in fact about 60 % the drugs currently on the market are chiral compounds, and nearly 90 % of them are sold as racemates. The same degree of enantio-selectivity is observed in the communications systems of plants and insects. Plants produce lipophilic liquids with high vapour pressure called plant volatiles (PVs) which are synthesized via different enzymes called tarpene synthases that are usually chiral. Chiral molecules and chiral effects have a strong impact on all the fields of science with exciting developments ranging from stereo-selective synthesis based on heterogeneous enantioselective catalysis, to optoelctronics, to photochemical asymmetric synthesis, and chiral surface science, just to cite a few.
Chiral molecules come in two forms called enantiomers. Their almost identical chemical and physical properties continue to pose technical challenges concerning the resolution of racemic mixtures, the determination of the enantiomeric excess, and the direct determination of the absolute configuration of an enantiomer. ...
This thesis deals with the phenomenology of QCD matter, its aspects in heavy ion collisions and in neutron stars. The first half of the work focuses on the hadronic phase of QCD matter. One focus is on how the hadronic phase shows itself in heavy ion collisions and how its dynamics can be simulated. The role of hadronic interactions is considered in the context of the lattice QCD data. The second part of this thesis presents a unified approach to QCD matter, the CMF model. The CMF model incorporates many aspects of QCD phenomenology which allows for a consistent description of the hadron-quark transition, making it applicable to the entire QCD phase diagram, i.e., to the cold nuclear matter and to the hot QCD matter. It is shown that a description of both the hot matter created in heavy ion collisions and the cold dense matter in neutron star interiors is possible within one single approach, the CMF model.
The putative effects of dark matter are most easily explained by a collisionless fluid on cosmological scales and by Modified Newtonian Dynamics (MOND) on galactic scales. Hybrid MOND dark matter models combine the successes of dark matter on cosmological scales and those of MOND on galactic scales. An example of such a model is superfluid dark matter (SFDM) which postulates that this differing behavior with scale is caused by a single underlying substance with two phases. In this thesis, I highlight successful observational tests of SFDM regarding strong lensing and the Milky Way rotation curve. I also discuss three problems due to the double role of the aforementioned single underlying substance and show how these may be avoided. Finally, I introduce a novel Cherenkov radiation constraint for hybrid MOND dark matter models. This constraint is different from standard modified gravity Cherenkov radiation constraints because such hybrid models allow even non-relativistic objects like stars to emit Cherenkov radiation.
The realization of a fast and robust closed orbit feedback (COFB) system for the on-ramp orbit correction at SIS18 synchrotron of FAIR project is reported in this thesis. SIS18 has some peculiar behaviors including on-ramp optics variation, very short lengths of the ramps (200 ms to 1 s) and a cycle-to-cycle variation of beam parameters. The realized fast COFB system being robust against above mentioned features of SIS18 is a first of its kind and the course to its realization led to some novel contributions in the field of closed orbit correction. A new method relying on the discrete Fourier transform (DFT)-based decomposition of the orbit response matrix (ORM) has been introduced, exploiting the symmetry in the arrangement of beam position monitors (BPMs) and the corrector magnets in the synchrotrons. A nearest-circulant approximation has also been introduced for synchrotrons having slight deviation from the symmetry, making the method applicable to a vast majority of synchrotrons. Moreover, the performance and the stability analysis of COFB systems in the presence of ORM mismatch between the synchrotron and the feedback controller is presented. The COFB systems are divided into slow and fast regimes and a new stability criterion consistent with measurements, is introduced. The practicality of the criterion is verified experimentally at COSY Jülich and is used for the analysis of various sources of ORM mismatch at SIS18. The commissioning of the SIS18 COFB system is also reported in detail which relies on Libera Hadron as the main hardware resource for the controller implementation. The on-ramp orbit correction is demonstrated for the horizontal plane of SIS18, for the disturbance rejection up to 600 Hz.
Topological phases set themselves apart from other phases since they cannot be understood in terms of the usual Landau theory of phase transitions. This fact, which is a consequence of the property that topological phase transitions can occur without breaking symmetries, is reflected in the complicated form of topological order parameters. While the mathematical classification of phases through homotopy theory is known, an intuition for the relation between phase transitions and changes to the physical system is largely inhibited by the general complexity.
In this thesis we aim to get back some of this intuition by studying the properties of the Chern number (a topological order parameter) in two scenarios. First, we investigate the effect of electronic correlations on topological phases in the Green's function formalism. By developing a statistical method that averages over all possible solutions of the manybody problem, we extract general statements about the shape of the phase diagram and investigate the stability of topological phases with respect to interactions. In addition, we find that in many topological models the local approximation, which is part of many standard methods for solving the manybody lattice model, is able to produce qualitatively correct phase transitions at low to intermediate correlations.
We then extend the statistical method to study the effect of the lattice, where we evaluate possible applications of standard machine learning techniques against our information theoretical approach. We define a measure for the information about particular topological phases encoded in individual lattice parameters, which allows us to construct a qualitative phase diagram that gives a more intuitive understanding of the topological phase.
Finally, we discuss possible applications of our method that could facilitate the discovery of new materials with topological properties.
Particle collisions provide insight into the structure of matter and the interaction of its constituents. Furthermore, they also allow a better understanding of the processes involved in the formation of the universe. To cover these diverse areas, it is necessary to study different observables and collision systems. A particular challenge is to find a suitable measurable observable for a theoretically meaningful variable and to develop a measurement process taking into account the experiment. The analyses of particle collisions in this thesis cover many of the challenges and objectives mentioned above. The focus of the work is the analysis of isolated photons at an energy of √s = 7 TeV. In addition, the work also includes measurements of the average transverse momentum in Pb-Pb collisions at an energy of √s = 2.76 TeV.
Apart from the collision system, the two analyses complement each other in other respects. The measurement of isolated photons represents the first measurement of this observable with ALICE and thus lays the foundation for further measurements at other collision systems and energies. The measurement of the mean transverse momentum, on the other hand, is based on an established measurement and thus allows the comparison of different collision systems. Likewise, the physical processes studied differ. With the measurement of isolated photons, hard scattering processes in the collisions can be investigated, while the average transverse momentum allows a description of the underlying event.
When measuring isolated photons, it should be noted that isolated photons are a measurable observable that cannot be assigned to an explicit physical process. The isolation criterion used in the analysis serves to increase the fraction of prompt photons from 2→2 processes. These photons can contribute to a better understanding of the parton density function (PDF) of gluons, as well as be used as a reference for perturbative QCD calculations.
Of particular importance for the analysis are the cluster shape and the energy within a certain radius around the potential photon. The combination of these two quantities allows determining the background using the ABCD method established by CDF and ATLAS. The result obtained in this way extends the previous measurements of the cross-section of isolated photons at the LHC to lower transverse momenta. Similarly, the previous measurements of the cross-section as a function of the scale variable xT are extended to lower values.
The main focus of the measurement of the average transverse momentum of charged particles ⟨pT⟩ is to compare the measurement for the pp, p-Pb, and Pb-Pb collision systems. To obtain a direct comparison between the different collision systems, ⟨pT ⟩ is measured against the true multiplicity nch. Since the multiplicity range of pp and p-Pb collisions is limited, the analysis in Pb-Pb collisions is restricted to nch = 100. This range corresponds to peripheral Pb-Pb collisions. A particular focus of the analysis is the determination and reduction of the electromagnetic background in peripheral Pb-Pb collisions and the determination of nch based on the measured multiplicity nacc . The different collision systems show similar behavior with increasing multiplicity. The steepest increase occurs at low multiplicities and changes for all collision systems at nch = 14. With higher multiplicities, the slope reduces further, with the effect being most pronounced in Pb-Pb collisions.
This thesis explores the phase diagrams of the Nambu--Jona-Lasinio (NJL) and quark-meson (QM) model in the mean-field approximation and beyond. The focus lies in the investigation of the interplay between inhomogeneous chiral condensates and two-flavor color superconductivity.
In the first part of this thesis, we study the NJL model with 2SC diquarks in the mean-field approximation and determine the dispersion relations for quasiparticle excitations for generic spatial modulations of the chiral condensate in the presence of a homogeneous 2SC-diquark condensate, provided that the dispersion relations in the absence of color superconductivity are known. We then compare two different Ansätze for the chiral order parameter, the chiral density wave (CDW) and the real-kink crystal (RKC). For both Ansätze we find for specific diquark couplings a so-called coexistence phase where both the inhomogeneous chiral condensate and the diquark condensate coexist. Increasing the diquark coupling disfavors the coexistence phase in favor of a pure diquark phase.
On the other hand, decreasing the diquark coupling favors the inhomogeneous phase over the coexistence phase.
In the second part of this thesis the functional renormalization group is employed to study the phase diagram of the quark-meson-diquark model. We observe that the region of the phase diagram found in previous studies, where the entropy density takes on unphysical negative values, vanishes when including diquark degrees of freedom. Furthermore, we perform a stability analysis of the homogeneous phase and compare the results with those of previous studies. We find that an increasing diquark coupling leads to a smaller region of instability as the 2SC phase extends to a smaller chemical potential. We also find a region where simultaneously an instability occurs and a non-vanishing diquark condensate forms, which is an indication of the existence of a coexistence phase in accordance with the results of the first part of this work.
The Compressed Baryonic Matter (CBM) Experiment will investigate heavy ion collisions and reactions at interaction rates of 100 kHz in a targeted energy range of up to 11 AGeV for systems such as gold-gold or lead-lead. It will be one of the major scientific experiments of the Facility for Antiproton and Ion Research in Europe (FAIR) currently under construction at the site of the GSI Helmholtzzentrum für Schwerionenforschung (GSI) in Darmstadt, Germany. CBM is going to be a fixed target experiment consisting of a superconducting magnet, multiple detectors of various types, and high-performance computing for online event reconstruction and selection. The detector closest to the interaction point of the experiment will be the Micro Vertex Detector (MVD). Consisting of four planar stations equipped with custom CMOS pixel sensors, it will allow to reconstruct the primary vertex with high precision and will help to reconstruct secondary vertices and identify particles originating from conversion in the detector material.
Due to the high interaction rates foreseen for CBM, understanding and minimizing systematic errors due to the detectors’ operating conditions will become all the more important to obtain significant measurement results, as statistical errors in the measurements of many observables are diminishing due to the enormous amount of data available.
Furthermore, the MVD will be the first detector based on CMOS pixel sensors used in a large physics experiment, that will be operated in vacuum. As a result, many aspects of the mechanical and electrical integration of the detector require careful testing and validation.
This thesis addresses both those challenges specifically for the Micro Vertex Detector with the development of a control system for the operation and validation of the MVD prototype “PRESTO” in vacuum. The prototype was selected as device under test as the final MVD is not yet built.
The developed control system helps a) to operate the prototype safely and keep it at the desired working point and b) to record important time-series data of the state of the detector prototype. Those two aspects allow the control system (which might later serve as a ‘blueprint’ for the final detector) to minimize the mentioned systematic errors as much as possible and to contribute to the understanding of remaining systematic errors using correlations with the time-series data. The controlled operation of the prototype in vacuum allowed to validate the integration concepts from a wide range of mechanical and electrical aspects in an endurance test for more than a year with 24/7 operation.
The prototype for this study itself was named “PRESTO” (standing for ‘PREcursor of the Second sTatiOn of the CBM-MVD’). It represents one quadrant of an MVD detector plane, equipped with a total of 15 MIMOSA-26 sensors on the front and back side of a carrier plate. Within this thesis, major parts of the prototype itself were designed. Custom ultra-thin flat flexible cables for data and power were designed and validated. Furthermore, the CNC-machined Aluminium heatsink to mount and cool the prototype design was refined to increase thermal performance. A custom vacuum feedthrough for a total of 21 flat ribbon cables was designed and fabricated. The read-out chain for MIMOSIS-26 was extended to cover a total of 8 sensors with a single and newer TRB-3 FPGA board and was set-up with the prototype. Vacuum equipment including chambers, hoses, pumps, valves and gauges were integrated to form a large vacuum testing system. A cooling circuit for the prototype was assembled comprising an external chiller, hoses, vacuum feedthroughs, as well as temperature, flow and pressure sensors.
The control system was developed to serve the needs of the prototype, while taking the requirements of the final MVD already into account. The main design goals of the control system are:
• compatibility with the other detectors and the overall CBM experiment,
• access to real-time measurements of all necessary parameters (‘process values’),
• reliable, fail-safe operation of the detector,
• recording of all time-series data (‘archiving’),
• cost efficiency and acceptance within the physics community,
• good usability for the users (‘operators’),
• long-term maintainability.
The recorded time-series data of the process variables (i.e. sensor readings) allow a post-measurement analysis of variations in the detector performance. The longterm archiving of all relevant system parameters is therefore of outstanding importance, which is why the software intended for this purpose – called “archiver” – was given special attention in this thesis.
For this reason in particular, it is necessary to implement a comprehensive control system that allows the detector to be operated safely under these conditions and cooled effectively. Before the start of this doctoral thesis, vigilant and extensively trained operators were always necessary for this. The control system that has been developed makes it possible that, after basic training, the detector can also be operated by a less specialised shift supervisor during measurement campaigns.
...
Die vorliegende Dissertation behandelt das Thema der Wechselstromleitfähigkeit nano-granularer Metalle, welche mit Hilfe der fokussierten elektronenstrahlinduzierten Direktabscheidung (FEBID) hergestellt wurden, sowie der dielektrischen Relaxation in metall-organischen Gerüstverbindungen (MOFs). Sie war eingebettet in das interdisziplinäre Projekt „Dielectric and Ferroelectric Surface-Mounted Metal-Organic Frameworks (SURMOFs) as Sensor Devices“ im Rahmen des DPG-Schwerpunktsprogramms „Coordination Networks: Building Blocks for Functional Systems“ (SPP 1928, COORNETs). Dabei verfolgt sie ein Sensorkonzept zur selektiven Detektion von Analytgasen. Der zentrale Erfolg der Arbeit besteht dabei in neuen Erkenntnissen über die Wechselstromleitfähigkeit nano-granularer Pt(C)-FEBID-Deponate. Die hierbei gewonnen Erkenntnisse können in Zukunft einen weiteren Baustein in der theoretischen Beschreibung dieses grundlegend interessanten und für sensorische Anwendungen wichtigen Teilgebiets der Festkörperphysik darstellen.
In this thesis we investigate the thermodynamic and dynamic properties of the D-dimensional quantum Heisenberg ferromagnet within the spin functional renormalization group (FRG); a
formalism describing the evolution of the system’s observables as the magnetic exchange inter-action is artificially deformed. Following an introduction providing a self contained summary of the conceptual and mathematical background, we present the spin FRG as developed by Krieg and Kopietz in references [1] and [2] in chapter two. Thereto, the generating functional of the imaginary time-spin correlation functions and its exact flow equation describing the deformation process of the exchange interaction are introduced. In addition, it is highlighted that - in contrast to conventional field-theoretic FRG approaches - the related Legendre trans-formed functional cannot be defined if the exchange interaction is initially switched off. Next, we show that this limitation can be circumvented within an alternativ hybrid approach, which treats transverse and longitudinal spin fluctuations differently. The relevant functionals are introduced and the relations of the corresponding functional Taylor coefficients with the spin correlation functions are discussed. Lastly, the associated flow equations are derived and the possibility of explicit or spontaneous symmetry breaking is taken into account.
In chapter three, we benchmark the hybrid formalism against a calculation of the thermo-dynamic properties of the one and two-dimensional Heisenberg model at low temperatures T and finite magnetic field H. For this purpose, we devise an anisotropic deformation scheme of the exchange interaction which allows for a controlled truncation of the infinite hierarchy of FRG flow equations. Thereby, contact with mean-field and spin-wave theory is made and the violation of the Mermin-Wagner theorem is discussed. To fulfill the latter, the truncation scheme is then complemented by a Ward identity relating the transverse self-energy and the magnetization. The resulting magnetization M (H, T ) and isothermal susceptibility χ(H, T ) are in quantitative agreement with the literature and the established behavior of the transverse correlation length and the zero-field susceptibility close to the critical point is qualitatively reproduced in the limit H → 0.
Finally, we investigate the longitudinal dynamics at low temperatures. To this end, the hierarchy of flow equations is solved within the same anisotropic deformation scheme complemented by an expansion in the inverse interaction range, and the resulting longitudinal dynamic structure factor is calculated within a low-momentum expansion. In D = 3, the large phase space accessible for the decay into transverse magnons yields only a broad hump centered at zero frequency whose width scales linearly in momentum. In contrast, at low temperatures and in a certain range of magnetic fields, a well-defined quasiparticle peak with linear dispersion emerges in D ≤ 2, which we identify as zero-magnon sound. Sound velocity and damping are discussed as a function of temperature and magnetic field, and the relevant momentum-frequency window is estimated and compared to the hydrodynamic
second-magnon regime.
The present research in high energy physics as well as in the nuclear physics requires the use of more powerful and complex particle accelerators to provide high luminosity, high intensity, and high brightness beams to experiments. With the increased technological complexity of accelerators, meeting the demand of experimenters necessitates a blend of accelerator physics with technology. The problem becomes severe when optimization of beam quality has to be provided in accelerator systems with thousands of free parameters including strengths of quadrupoles, sextupoles, RF voltages, etc. Machine learning methods and concepts of artificial intelligence are considered in various industry and scientific branches, and recently, these methods are used in high energy physics mainly for experiments data analysis.
In Accelerator Physics the machine learning approach has not found a wide application yet, and in general the use of these methods is carried out without a deep understanding on their effectiveness with respect to more traditional schemes or other alternative approaches. The purpose of this PhD research is to investigate the methods of machine learning applied to accelerator optimization, accelerator control and in particular on optics measurements and corrections. Optics correction, maximization of acceptance, and simultaneous control of various accelerator components such as focusing magnets is a typical accelerator scenario. The effectiven- ess of machine learning methods in a complex system such as the Large Hadron Collider, which beam dynamics exhibits nonlinear response to machine settings is the core of the study. This work presents successful application of several machine learning techniques such as clustering, decision trees, linear multivariate models and neural networks to beam optics measurements and corrections at the LHC, providing the guidelines for incorporation of machine learning techniques into accelerator operation and discussing future opportunities and potential work in this field.
Bohmian mechanics as formulated originally in 1952, has been useful in the implementation of numerical methods applied to quantum mechanics. The scientific community though has had ever since a critical thought about it. Therefore, there are still points to be clarified and rectified. The two main problems are basically: Bohmian mechanics gives a privilege role to the position representation. Secondly, the current interpretation of Bohmian trajectories has been recently proven wrong.
In this context, in Chapter 2, new complex Bohmian quantities are defined; so that they allow the capacity to formulate Bohmian mechanics in any arbitrary continuous representation, for instance, the momentum representation. This Chapter is fully based on two articles, regarding the proposed complex Bohmian formulation and its extension into momentum space.
Chapter 3 deals with a redefinition and reinterpretation of the Bohmian trajectories from the handling of the continuity equation, this is done without any need of additional postulates or interpretations. Also, it is proved that Bohmian mechanics is actually more than a projective aspect of the Wigner function.
As a third point, Chapter 4 presents a sytematic treatment of the hydrodynamic scheme of Bohmian mechanics. Then, a brief summary of the transport equations in Bohmian mechanics is done. Next, a unified hydrodynamic treatment is found for the Bohmian mechanics. This treatment is useful to sketch, a Bohmian treatment to efficiently find the steady value of the transmission integral.
In Chapter 5 conclusions of this thesis are drawn.
Das Feld der Hochenergie-Schwerionenforschung hat sich der Untersuchung des Quark-Gluon-Plasmas (QGP) gewidmet. Ein QGP ist ein sehr heißer und dichter Materiezustand, der kurz nach dem Urknall für einige Mikrosekunden das Universum füllte. Unter diesen extremen Bedingungen sind die fundamentalen Bausteine der Materie, die Quarks und Gluonen, quasi frei, also nicht in Hadronen eingeschlossen, wie es unter normalen Bedingungen der Fall ist. Hadronen sind Teilchen, die aus Quarks und Gluonen bestehen. Die bekanntesten Hadronen sind Protonen und Neutronen, die Bestandteile von Atomkernen, aus denen, zusammen mit Elektronen, die gesamte bekannte Materie aufgebaut ist.
Um ein QGP im Labor zu erzeugen, lässt man ultrarelativistische schwere Ionen, wie zum Beispiel Pb-208-Kerne, aufeinander prallen. Dies geschieht am CERN, dem größten Kernforschungszentrum der Welt. Der Teilchenbeschleuniger, welcher Protonen und Pb-Kerne beschleunigt und zur Kollision bringt, heißt Large Hadron Collider (LHC) und ist mit 27 km Umfang der größte der Welt. Bei einer einzigen Pb-Pb Kollision am LHC werden mehrere Tausend Teilchen und Antiteilchen erzeugt. Das dedizierte Experiment zur Untersuchung von Schwerionenkollisionen am LHC ist ALICE. ALICE ist mit mehreren Teilchendetektoren ausgerüstet, die es ermöglichen, tausende Teilchen gleichzeitig zu messen und zu identifizieren.
Unter den produzierten Teilchen befinden sich auch leichte Atomkerne, wenngleich diese nur sehr selten erzeugt werden. Die Anzahl der produzierten Teilchen pro Teilchensorte hängt nämlich von deren Masse ab. In Pb-Pb Kollisionen am LHC sinkt die Anzahl der produzierten (Anti)kerne exponentiell um einen Faktor 1/330 bei Hinzufügen jedes weiteren Nukleons. Die Menge an produzierten Teilchen pro Spezies stellt Informationen über den Produktionsmechanismus beim Übergang vom QGP zum Hadrongas zur Verfügung. Hierbei sind leichte (Anti)kerne von besonderem Interesse, da sie vergleichsweise groß sind und ihre Bindungsenergie bis zu zwei Größenordnungen kleiner ist als die Temperaturen, die bei der Erzeugung der Hadronen vorherrschen. Es ist bis heute noch nicht verstanden, wie leichte (Anti)kerne bei diesen Bedingungen erzeugt werden und überleben können.
Für diese Arbeit wurden ca. 270 Millionen Pb-Pb Kollisionen bei einer Schwerpunktsenergie von 5,02 TeV, die von ALICE im November 2018 aufgezeichnet wurden, analysiert. Es wurde die Produktion von (Anti)triton und (Anti)alpha untersucht. Wegen ihrer großen Masse werden beide Kerne sehr selten produziert, bei weitem nicht bei jeder Kollision. Antialpha ist der schwerste Antikern, der jemals gemessen wurde. Aufgrund dieser Seltenheit ist die Größe des zur Verfügung stehenden Datensatzes entscheidend. Es war möglich, das erste jemals gemessene Antialpha-Transversalimpulsspektrum zu extrahieren. Auch für (Anti)triton und Alpha wurden Transversalimpulsspektren bestimmt.
Die Ergebnisse wurden mit theoretischen Modellen und anderen ALICE Messungen verglichen.
Am Ende wird in einem Ausblick auf das kürzlich durchgeführte Upgrade der ALICE Spurendriftkammer (TPC) eingegangen. In der nächsten, bald startenden Datennahmeperiode wird der LHC seine Kollisionsrate erheblich erhöhen, was es ermöglichen wird, mehr als 100 mal so viele Daten wie bisher aufzuzeichnen. Hiervon werden die in dieser Arbeit beschriebenen (Anti)triton- und (Anti)alpha-Analysen beachtlich profitieren. Um mit den erheblich höheren Kollisionsraten zurecht zu kommen, mussten einige Detektoren, unter anderem die TPC, maßgeblich erneuert werden. In den ersten beiden Datennahmeperioden wurde die TPC mit Vieldrahtproportionalkammern betrieben. Diese sind allerdings viel zu langsam für die geplanten Kollisionsraten. Deshalb wurden sie im Jahr 2019, während einer langen Betriebspause des LHC, durch Quadrupel-GEM (Gas Electron Multiplier) Folien basierte Auslesekammern ersetzt, welche eine kontinuierliche Auslese der TPC ermöglichen. Da es sich um die erste jemals gebaute GEM TPC im Großformat handelt, war ein umfangreiches Forschungs- und Entwicklungs- (F&E) Programm notwendig, um die GEM Auslesekammern zu charakterisieren und zu testen. Im Rahmen dieses F&E Programms wurden am Anfang dieser Promotion systematische Messungen an einer kleinen Test TPC mit Quadrupel-GEM Auslese, die extra zu diesem Zweck gebaut worden war, durchgeführt. Hierbei wurde der Rückfluss der bei der Gasverstärkung erzeugten Ionen in das Driftvolumen der TPC und die Energieauflösung mit verschiedenen GEM Folien Typen und unterschiedlicher Anordnung gemessen. Das Ziel war, möglichst kleine Ionenrückflüsse bei möglichst guter Energieauflösung zu erreichen. Hierbei musste ein Kompromiss gefunden werden, da die beiden Größen sich gegenläufig verhalten. Es war jedoch möglich, mit mehreren GEM Konfigurationen Spannungseinstellungen zu identifizieren, bei denen beide Größen den gewünschten Anforderungen entsprachen.
Next-generation DIRC detectors, like the PANDA Barrel DIRC, with improved optical designs and better spatial and timing resolution, require correspondingly advanced reconstruction and PID methods. The investigation of the PID performance of two DIRC counters and the evaluation of the reconstruction and PID algorithms form the core of this thesis. Several reconstruction and PID approaches were developed, optimized, and tested using hadronic beam particles, experimental physics events, and Geant simulations. The near-final design of the PANDA Barrel DIRC was evaluated with a prototype in the T9 beamline at CERN in 2018. The analysis finds excellent agreement between the experimental data and the Geant simulations for all reconstruction algorithms. The best PID performance of up to $5.2 \pm 0.2$ s.d. $\pi$/K separation at 3.5 GeV/c, was obtained with a time imaging PID method. The PANDA Barrel DIRC simulation, as well as the reconstruction and PID algorithms, were evaluated using experimental data from the GlueX DIRC as part of the FAIR Phase-0 program. The performance validation was carried out using physics events of the GlueX experiment and simulations. The initial analysis results of the commissioning dataset show a $\pi$/K separation power of up to 3 s.d. at a momentum of 3.0-3.5 GeV/c, obtained using a geometric reconstruction algorithm.