Refine
Year of publication
Document Type
- Doctoral Thesis (596) (remove)
Has Fulltext
- yes (596)
Is part of the Bibliography
- no (596) (remove)
Keywords
- Quark-Gluon-Plasma (8)
- Schwerionenphysik (8)
- CERN (5)
- Heavy Ion Collisions (5)
- Ionenstrahl (5)
- LHC (5)
- Monte-Carlo-Simulation (5)
- Quantenchromodynamik (5)
- Schwerionenstoß (5)
- Teilchenbeschleuniger (5)
Institute
- Physik (596) (remove)
The Standard Model is one of the greatest successes of modern theoretical physics. Itl describes the physics of elementary particles by means of three forces, the electro-magnetisc, the weak and the strong interactions. The electro-magnetic and the weak interaction are rather well understood in comparison to the strong interaction.
The latest is as fundamental as the others, it is responsible for the formation of all hadrons which are classified into mesons and baryons. Well-known examples of the former is the pion and of the latter is the proton and the neutron, which form the nucleus of every atom. This fundamental force is believed to be described by the Quantum Chromodynamics (QCD) theory. According to this theory, hadrons are not elementary particles but are composed of quarks and gluons. The latter are the vector particles of the force and so are bosons of spin 1 and the former constitute the matter and are fermions with spin 1/2. To describe the interaction a new quantum number had to be introduced: the color charge which exists in three different types (blue, green and red). The name has not been chosen arbitrary as elements created from three quarks of different colors are colorless in the same way that mixing the three primary colors leads to white. However, experimentally no colored structure has ever been observed. The quarks and the gluons seem to be confined in colorless hadrons. This property of QCD is called confinement and results from a large coupling constant at low energy (or large distance). For high energy (or small distance), the perturbative analysis of QCD permits to establish the coupling constant to be small and quarks and gluons are almost free. This property is called asymptotic freedom. The possibility for QCD to describe both behaviors is one of its amazing characteristics. However, both phenomena are not well understood and one needs a method to study both the pertubative and the confining regime.
The only known method which fulfills the above criteria is Lattice QCD and more generally Lattice Quantum Field Theory (LQFT). It consists of a discretization of the spacetime and a formulation of QCD on a four-dimensional Euclidean spacetime grid of spacing a. In this way, the theory is naturally regularized and mathematically well-defined. On the other hand, the path integral formalism allows the theory to be treated as a Statistical Mechanics system which can be evaluated via a Markov chain Monte-Carlo algorithm. This method was first suggested by Wilson in 1974 [1] and shortly after Creutz performed the first numerical simulations of Yang-Mills theory [2] using a heath-bath Monte-Carlo algorithm. It appears that this method is extremely demanding in computational power. In its early days the method was criticized as the only feasible simulations involved non-physical values such as extremely large quark masses, large lattice spacing a and no dynamical quarks. With the progress of the computers and the appearance of the super-computer, the studies have come close to the physical point. But one still needs to deal with discrete space time and finite volume. Several techniques have been developed to estimate the infinite volume limit and the continuum limit. The smaller the lattice spacing and the larger the volume, the better the extrapolation to continuum and infinite volume limits is. The simulations are still very expensive and for the moment a typical length of the box is L ≈ 4fm and a ≈ 0.08fm. However, it has been realized simulating pure Yang-Mills theory and other lower dimensional models that the topology is freezing at small a [3]. It was also observed recently on full QCD simulations [4,5].
The typical lattice spacing for which this problem appears in QCD is a ≈ 0.05fm but this value depends on the quark mass used and on the algorithm. The freezing of topology leads to results which differ from physical results. Solving this issue is important for the future of LQCD [6]. Recently several methods to overcome the problem have been suggested, one of the most popular is the used of open boundary conditions [7] but this promising method has still its own issues, mainly the breaking of translation invariance.
Bei den Projekten wie der Europäischen und der Amerikanischen Spallationsneutronenquelle aber auch den geplanten aktuellen Großprojekten wie dem Upgrade von CERN oder ISIS werden negative Ionen benötigt. Bei solchen Anlagen werden am Ende des üblichen linearen Beschleunigers Speicherringe eingesetzt, die den Teilchenstrom akkumulieren und danach longitudinal komprimieren. Durch die Verwendung eines Strahls aus negativen Ionen kann die Injektion in den Speicherring wesentlich vereinfacht werden. In der vorliegenden Dissertation wurde die Extraktion und der Transport von negativen Wasserstoffionen für den ersten Abschnitt eines Linearbeschleunigers, bestehend aus Quelle, Extraktion und niederenergetischem Strahltransport (LEBT), sowohl experimentell als auch theoretisch untersucht. In dieser Sektion wird der grundlegende Strahlstrom und die Strahlqualität eines Linearbeschleunigers definiert. Eine komplette Untersuchung dieses Abschnitts lag bis dato für negative Ionen nicht vor. Um die Unterschiede aufzudecken und die einflußnehmenden Größen zu bewerten, mußten alle Experimente sowohl mit positiven als auch mit negativen Ionen durchgeführt werden. In allen Sektionen führen verschiedene Faktoren zu Strahlstromverlusten und Qualitätsverschlechterung, sprich Emittanzvergrößerung. Im Zuge dieser Arbeit wurde eine Quelle für negative Ionen entwickelt und gebaut und eine neue Methode zur Produktionssteigerung von negativen Ionen entwickelt. Die Innenwand der Plasmakammer der Ionenquelle wurde mit dem Edelmetallkatalysator Platin beschichtet. Die Plasmazusammensetzung innerhalb der Quelle verlagerte sich dadurch auf 80–90% H3 , 5-10% H2 und nur noch ein geringer Anteil an Protonen. Dieser hohe molekulare Anteil war über eine große Spanne aller Plasmaparameter stabil und führt zu einer drastischen Produktionssteigerung von angeregtem H2 und H- . Zur Formierung des Ionenstrahls wurde von mir ein sogenannter stromtoleranten Extraktor entwickelt. Trotz einer Veränderung des extrahierten Stroms um den Faktor 5 kommt es mit diesem Extraktor zu keinem nennenswerten Emittanzwachstum. Dieser eignet sich allgemein für die Extraktion gepulster Ionenstrahlen, im Besonderen aber für die Extraktion von negativen Ionen, da hierbei gleichzeitig Elektronen mit extrahiert werden. Dieser meist hohe Strahlanteil aus hochenergetischen Elektronen muß vor dem Einschuß der negativen Ionen in den RFQ durch ein geeignetes System aus dem Strahl ausgelenkt und abgeführt werden. Grundlagen, Entwicklung und Einflüsse dieser sogenannten Dumpingsysteme werden in Kap. 5 beschrieben. Für die Realisierung einer Niederenergietransportstrecke für negative Ionen stehen die beiden Möglichkeiten des magnetischen LEBT (Kap. 6) und des elektrostatischen LEBT (Kap. 7) zu Verfügung. Mit verschiedenen Meßaufbauten werden im anschließenden Kap. 8 die in den vorigen Kapiteln aufgeführten relevanten Größen der Erzeugung, der Extraktion und des Transport experimentell untersucht. Zusätzlich zu den bekannten klassischen Analyseverfahren kommen im Rahmen dieser Arbeit entwickelte optische Meßmethoden zum Einsatz, mit deren Hilfe man Plasmatemperatur und Plasmaverteilung innerhalb der Ionenquelle bestimmen kann. Mit Hilfe der Untersuchungen gelang es, die Unterschiede zwischen der Extraktion von negativen Ionen und von positiven Ionen aufzuzeigen und mit Hilfe der experimentellen Beobachtungen ein neues Modell für die Extraktion von negativen Ionen zu entwickeln. Mit der vorliegenden Arbeit wurde zudem gezeigt: - Der extrahierbare negative Strom ist hauptsächlich abhängig vom Diffusionsprozeß der Teilchen durch einen positiven Potentialwall innerhalb der Ionenquelle. - Durch Kompensation der magnetischen Felder in der Extraktionsregion wird die Emittanz reduziert und der Strom gesteigert. - Der beobachtete planare Plasmameniskus wird maßgeblich durch die rückfließenden Restgasionen beeinflußt. - Der Transport der negativen Ionen mit einer magnetischen LEBT stellt kein wesentliches Problem dar, da eine hinreichende Anzahl an positiven Restgasionen für den raumladungs-kompensierten Transport vorliegt.
High resolution gamma spectroscopy with sophisticated detector arrays significantly contributes to nuclear structure physics. The Advanced Gamma Tracking Array (AGATA) combines gamma tracking and pulse shape analysis to achieve an efficiency and quality of the spectra that could not be reached with spectrometers of the previous generation. Tracking of the photons interacting in the detector requires a precise knowledge of the individual interaction positions. The task of the pulse shape analysis is to provide a position resolution of better than $5mm$ FWHM, a value that could not be achieved by segmentation of the detector alone. As the signals induced on the electrodes of the detectors depends on the position of interaction, the charge pulses can be used to infer the interaction position. To be able to handle high rates, algorithms that are used have to be optimized to be able to process the data in real-time. Pulse shape analysis is the most involved part of the real-time processing and requires further improvement. This work is dealing with optimizations and improvements of pulse shape analysis algorithms. The Grid Search algorithm localizes the interaction position by comparing the measured pulse shape with precomputed shapes in a database to find the best fit. Two linear filters based on orthogonal transformations have been compared and it could be concluded that the one based on a singular value decomposition of the pulse shapes works best. It speeds up the pulse shape analysis by a factor of roughly $2-3$ (depending on how it is combined with the other modifications). Further, a new method to exclude most signals from the database as best fit has been developed based on the principle of lateration. Most interaction positions can be excluded by means of a fast check and for single interactions on average only $34.8\%$ of all signals from the database have to be compared to the measured one. The overhead introduced by the method is negligible and the reduced number of comparisons almost direclty translates into increased efficiency of the algorithm. A similar method could also be applied for double interactions. Two or more interactions taking place in the same segment require special treatment as the measured signals cannot be directly compared to signals from the database. A new method to calculate the figure of merit that quantifies the fit in case of a double interaction has been introduced. Compared to the unmodified algorithm the new method finds the best fit for double interactions roughly two orders of magnitude faster. Actually, the time required to localize double interactions is almost the same as for single interactions. Apart from optimizing the algorithm, also the achievable position resolution was investigated. It strongly varies inside the volume of the detector and it crucially depends on the shape of all signals in the database and the amplitude of the noise present in the measured signals. As a first step towards a precise analytic expression for the position resolution, an estimate for the probability to find the correct position has been derived.
Heavy-ion collisions at ultra-relativistic energies allow access to the Quark-Gluon Plasma, the deconfined phase of the strong interaction, a state which is believed to have existed fractions of seconds after the big bang. Two-particle correlations at small relative momenta, and particularly their dependence on pair transverse mass, are distinctly sensitive to the reaction dynamics of the fireball created in heavy-ion collisions.
Being the heaviest system to extract a size of the particle emitting source, proton-lambda correlations extend the studied range in pair transverse mass and are therefore well suited to explore the dynamical behavior of the matter created in Pb-Pb collisions at the Large Hadron Collider. The centrality dependence of the extracted source radii affirms the expectations of a larger source for more central collisions. Source radii were attained over a span of more than 0.9 GeV/c2 in mean pair transverse mass with a source radius extracted for a mean transverse mass as high as 2.18 GeV/c2. The source radii decrease with increasing pair transverse mass, as expected in a hydrodynamical picture. The comparison with radii obtained from other particle species exhibits the clear breaking of an elsewhere proposed scaling behavior of source radii with mean pair transverse mass for all particle species.
Gamma-gamma correlations possibly allow to look past the barrier of kinetic freeze-out. Additionally, they bear the potential to solve the puzzling observation in heavy-ion collisions of an excess of photons with a large temperature-like inverse slope parameter on the one hand and an elliptic flow coefficient of photons comparable to the one of hadrons on the other hand. A striking signal in the two-photon correlation function is observed; however it seems likely to not be of quantum statistical origin. A path for further studies is laid out.
In this thesis, we presented the theoretical description of the magnetic properties of various frustrated spin systems. Especially in search of exotic states, such as quantum spin liquids, magnetically frustrated systems have been subject of intense research within the last four decades. Relating experimental observations in real materials with theoretical models that capture those exotic magnetic phenomena has been one of the great challenges within the field of magnetism in condensed matter.
In order to build such a bridge between experimental observations and theoretical models, we followed two complementary strategies in this thesis. One strategy was based on first principles methods that enable the theoretical prediction of electronic properties of real materials without further experimental input than the crystal structure. Based on these predictions, low-energy models that describe magnetic interactions can be extracted and, through further theoretical modelling, can be compared to experimental observations. The second strategy was to establish low-energy models through comparison of data from experiments, such as inelastic neutron scattering intensities, with calculated predictions based on a variety of plausible magnetic models guided by microscopic insights. Both approaches allow to relate theoretical magnetic models with real materials and may provide guidance for the design of new frustrated materials or the investigation of promising models related to exotic magnetic states.
By combining two unique facilities at the Gesellschaft fuer Schwerionenforschung (GSI), the Fragment Separator (FRS) and the Experimental Storage Ring (ESR), the first direct measurement of a proton capture reaction of stored radioactive isotopes was accomplished. The combination of well-defined ion energy, an ultra-thin internal gas target, and the ability to adjust the beam energy in the storage ring enables precise, energy-differentiated measurements of the (p,gamma) cross sections. The new setup provides a sensitive method for measuring (p,gamma) reactions relevant for nucleosynthesis processes in supernovae, which are among the most violent explosions in the universe and are not yet well understood. The cross sections of the 118Te(p,gamma) and 124Xe(p,gamma) reactions were measured
at energies of astrophysical interest. The heavy ions were stored with energies of 6 MeV/nucleon and 7 MeV/nucleon and interacted with a hydrogen gas-jet target.
The produced proton-capture products were detected with a double-sided silicon strip detector. The radiative recombination process of the fully stripped ions and electrons from the hydrogen target was used as a luminosity monitor.
Additionally, post-processing nucleosynthesis simulations within the NuGrid [1] research platform have been performed. The impact of the new experimental results on the p-process nucleosynthesis around 124Xe and 118Te in a core-collapse supernova was investigated. The successful measurement of the proton capture cross sections of radioactive isotopes rises the motivation to proceed with experiments in lower energy regions.
[1] M. Pignatari and F. Herwig, “The nugrid research platform: A comprehensive simulation approach for nuclear astrophysics,” Nuclear Physics News, vol. 22, no. 4, pp. 18–23, 2012.
Das Ziel der Untersuchung von ultra-relativistischen Schwerionenkollisionen ist die Suche nach dem Quark Gluon Plasma (QGP), einem Zustand hochdichter stark wechselwirkender Materie in dem der Einschluss von Quarks und Gluonen in Hadronen aufgehoben ist. Die bisher gewonnenen experimentellen Hinweise deuten daraufhin,daß in Schwerionenkollisionen bei den derzeit höchsten zur Verfügung stehenden Energien von 158 GeV/Nukleon in Pb+Pb Reaktionen am CERN-SPS die Rahmenbedingungen für einen Phasenübergang von hadronischer Materie zu einer partonischen Phaseerfüllt sind. Die exakte Phasenstruktur stark wechselwirkender Materie hingegen ist derzeit noch nicht vollständig verstanden. Da inklusive hadronische Observablen und "penetrierende Proben" nicht direkt sensitiv auf die Existenz und Natur des Phasenübergangs sind, wurde die Analyse von Einzelereignis-"event-by-event"-Fluktuationenvorgeschlagen. Das Fluktuationsverhalten von Einzelereignis-Observablen sollte direkt sensitiv auf die Natur des zu beobachtenden Phasenübergangssein. In dieser Arbeit wurden Fluktuationen in der "chemischen" Zusammensetzung der Teilchenquelle untersucht und erste Ergebnisse werden präsentiert.
The current thesis is devoted to a systematic study of fluctuations and correlations in heavy-ion collisions, which might be considered as probes for the phase transition and the critical point in the phase diagram, within the Hadron-String- Dynamics (HSD) microscopic transport approach. This is a powerful tool to study nucleus-nucleus collisions and allows to completely simulate experimental collisions on an event-by-event basis. Thus, the transport model has been used to study fluctuations and correlations including the influence of experimental acceptance as well as centrality, system size and collision energy. The comparison to experimental data can separate the effects induced by a phase transition since there is no phase transition in the HSD version used here. Firstly the centrality dependence of multiplicity fluctuations has been studied. Different centrality selections have been performed in the analysis in correspondence to the experimental situation. For the fixed target experiment NA49 events with fixed numbers of the projectile participants have been studied while in the collider experiment PHENIX centrality classes of events have been defined by the multiplicity in certain phase space region. A decrease of participant number fluctuations (and thus volume fluctuations) in more central collisions for both experiments has been obtained. Another area of this work addresses to transport model calculations of multiplicity fluctuations in nucleus-nucleus collisions as a function of colliding energy and system size. This study is in full correspondence to the experimental program of the NA61 Collaboration at the SPS. Central C+C, S+S, In+In, and Pb+Pb nuclear collisions at Elab = 10, 20, 30, 40, 80, 158 AGeV have been investigated. The expected enhanced fluctuations - attributed to the critical point and phase transition - can be observed experimentally on top of a monotonic and smooth ‘hadronic background’. These findings should be helpful for the optimal choice of collision systems and collision energies for the experimental search of the QCD critical point. Other observables are fluctuations of ratios of hadrons (e.g. pions, kaons, protons, etc.) which are not so much affected by volume fluctuations. In particular HSD results for the kaon-to-pion ratio fluctuations, which has been regarded as promising observable for a long time, are presented from low SPS energies up to high energies at RHIC. In addition to the HSD calculations statistical model is also used in terms of microcanonical, canonical and grand canonical ensembles. Further a study of the system size event-by-event fluctuations causing rapidity forward-backward correlations in relativistic heavy-ion collisions is presented. The HSD simulations reveal strong forward-backward correlations and reproduce the main qualitative features of the STAR data in A+A collisions at RHIC energies. It has been shown that strong forward-backward correlations arise due to an averaging over many different events that belong to one centrality bin. An optimization of the experimental selection of centrality classes is presented, which is relevant for the program of the NA61 collaboration at CERN, the low-energy program at RHIC, as well as future experiments at FAIR.
Quantum chromodynamics predicts the existence of a phase transition from hadronic to quark-gluon matter when temperature and pressure are sufficiently high. Colliding heavy nuclei at ultra-relativistic speeds allows to deposit large amounts of energy in a small volume of space, and is the only available experimental mean to produce the extreme conditions necessary to obtain the deconfined state. Numerous models and ideas were developed in the last decades to study heavy ion physics and understand the properties of extremely heated and compressed nuclear matter. With the ever increasing energy available in the center of mass frame (and thus number of particles produced) and the development of large acceptance detectors, it has become possible to study the fluctuations of physical quantities on an event-by-event basis, and access thermodynamical properties not present in particle spectra. The characteristics of the highly excited matter produced, e.g. thermalization, effect of resonance decay. . . can be investigated by fluctuation analyses. In fact, fluctuations are good indicators for a phase transition and a plethora of fluctuation probes have been proposed to pin down the existence and the properties of the QGP. We study various fluctuation quantities within the Ultra-relativistic Quantum Molecular Dynamics UrQMD and the quantum Molecular Dynamics qMD models. UrQMD is based on hadron and string degrees of freedom and allows to disentangle purely hadronic effects. In contrast, the qMD model includes an explicit transition from quark to hadronic matter and can serve to test adequate probes of the initial QGP state. We show that the qMD model can reasonably reproduce various experimental particles rapidity distributions and transverse mass spectra in wide energy range. Within the frame of the dynamical recombination procedure used in qMD, we study the enhancement of protons over pions (p/π) ratio in the intermediate pt range (1.5 < pt < 2.5). We show that qMD can reproduce the large p/π ≈ 1 observed experimentally at RHIC energies at hadronization. However, the subsequent decay of resonances makes the ratio fall to values incompatible with experimental data. We thus conclude that resonance decay might have a drastic influence on this observable in the quark recombination picture. Charged particles multiplicity fluctuations measured at SPS by the NA49 collaboration are enhanced in midperipheral events for Pb+Pb collisions at Elab = 160 AGeV. This feature is not reproduce by hadron-string transport approaches, which show a flat centrality dependence, within the proper experimental acceptance and with the proper centrality selection procedure. However, we show that the behavior of multiplicity fluctuations in transport codes is similar to the experimental result in full 4π acceptance. We identify the centrality selection procedure as the reason for the enhanced particle multiplicity fluctuations in midperipheral reactions and argue that it can be used to distinguish between different scenarios of particle productions. We show that experimental data might indicate a strong mixing of projectile and target related production sources. Strangeness over entropy K/π and baryon number over entropy p/π ratio fluctuations have been measured by the NA49 experiment in the SPS energy range, from Elab = 20 AGeV up to Elab = 160 AGeV. We investigate the sensitivity of this observable to kinematical cuts and discuss the influence of resonance decay. We find the dynamical p/π ratio fluctuations to increase with beam energy, in agreement with the measured data points. On the contrary, the dynamical K/π ratio fluctuations are essential flat as a function of centrality and depend only weakly on the kinematical cuts applied. Our results are in line with the simulations performed earlier by the NA49 collaboration in their detector acceptance filter. Finally, we focus on the correlations and fluctuations of conserved charges. It was proposed that these fluctuations are sensitive to the fractional charge carried by the quarks in the initial QGP stage and survive the whole course of heavy ion reactions. A crucial point is the influence of hadronization that may relax the initial QGP fluctuation/correlation signals to their hadronic values. We use the quark Molecular Dynamics qMD model to disentangle the effect of recombination-hadronization on charged particles ratio fluctuations, charge transfer fluctuations, baryon number-strangeness correlation coefficient and various ratios of susceptibilities (i.e. correlations over fluctuations). We find that the dynamical recombination procedure implemented in the qMD model destroys all studied initial QGP fluctuations and correlations and might ex- plain why no signal of a phase transition based on event-by-event fluctuations was found in the experimental data until now.