Informatik
Refine
Year of publication
Document Type
- Preprint (748)
- Article (401)
- Working Paper (119)
- Doctoral Thesis (93)
- Diploma Thesis (46)
- Conference Proceeding (41)
- Book (37)
- Bachelor Thesis (36)
- diplomthesis (29)
- Report (25)
Has Fulltext
- yes (1607)
Is part of the Bibliography
- no (1607)
Keywords
Institute
- Informatik (1607)
- Frankfurt Institute for Advanced Studies (FIAS) (1002)
- Physik (982)
- Mathematik (55)
- Präsidium (41)
- Medizin (25)
- Biowissenschaften (21)
- Exzellenzcluster Makromolekulare Komplexe (8)
- Psychologie (8)
- Deutsches Institut für Internationale Pädagogische Forschung (DIPF) (5)
- Geowissenschaften (5)
- Senckenbergische Naturforschende Gesellschaft (5)
- Geographie (4)
- Hochschulrechenzentrum (4)
- Pharmazie (4)
- Biochemie und Chemie (3)
- Goethe-Zentrum für Wissenschaftliches Rechnen (G-CSC) (3)
- Universitätsbibliothek (3)
- Biodiversität und Klima Forschungszentrum (BiK-F) (2)
- Center for Membrane Proteomics (CMP) (2)
- Institut für Ökologie, Evolution und Diversität (2)
- Sportwissenschaften (2)
- Wirtschaftswissenschaften (2)
- Center for Scientific Computing (CSC) (1)
- E-Finance Lab e.V. (1)
- Erziehungswissenschaften (1)
- Geschichtswissenschaften (1)
- Gesellschaftswissenschaften (1)
- Informatik und Mathematik (1)
- Institut für Bienenkunde (1)
- Kulturwissenschaften (1)
- Neuere Philologien (1)
- Zentrum für Arzneimittelforschung, Entwicklung und Sicherheit (ZAFES) (1)
- Zentrum für Weiterbildung (1)
A Large Ion Collider Experiment (ALICE) is a high-energy physics experiment, designed to study heavy ion collisions at the European Organization for Nuclear Research (CERN)Large Hadron Collider (LHC). ALICE is built to study the fundamental properties of matter as it existed shortly after the big bang. This requires reading out millions of sensors with high frequency, enabling high statistics for physics analysis, resulting in a considerable computing demand concerning network throughput and processing power. With the ALICE Run 3 upgrade [14], requirements for a High Throughput Computing
(HTC) online processing cluster increased significantly, due to more than an order of magnitude more data than in Run 2, resulting in a processing input rate of up to 900 GB/s. Online (real-time) event reconstruction allows for the compression of the data stream to 130 GB/s, which is stored on disk for physics analysis.
This thesis presents the implementation of the ALICE Event Processing Node (EPN) compute farm, to cope with the Run 3 online computing challenges. Building a Data Centre tailored to ALICE requirements for the Run 3 and Run 4 EPN farm. Providing the operational conditions for a dynamic compute environment of a High Performance Computing (HPC) cluster, with significant load changes in a short time span, when starting or stopping a data-taking run. EPN servers provide the required computing resources for online reconstruction and data compression. The farm includes network connectivity towards First Level Processors (FLPs), requiring reliable throughput of 900 GB/s between FLPs and EPNs and connectivity from the internal InfiniBand network to the CERN Exabyte Object Storage (EOS) Ethernet network, with more than 100 GB/s.
The results of operating the EPN computing infrastructure during the first year of Run 3 LHC collisions are described in the context of the ALICE experiment. The EPN farm was delivering the expected performance for ALICE data-taking. Data Centre environmental conditions remained stable during the last more than two years, in particular during starting and stopping runs, which include significant changes in IT load. Several unforeseen external circumstances lead to increasing demands for the Online Offline System (O2). Higher data rates than anticipated required network performance to exceed the initial design specifications, for the throughput between FLPs and EPNs. In particular, the high throughput from an internal EPN InfiniBand network towards the storage Ethernet network was one of the challenges to overcome.
Natural Language Processing (NLP) for big data requires an efficient and sophisticated infrastructure to complete tasks both fast and correctly. Providing an intuitive and lightweight interaction with a framework that abstracts and simplifies complex tasks assists in reaching this goal. This bachelor thesis extends the NLP framework Docker Unified UIMA Interface (DUUI) by an API and a web-based graphical user interface to control and manage pipelines for automated analysis of large quantities of natural language. The extension aims to reduce the entry barrier into the field as well as to accelerate the creation and management of pipelines according to UIMA standards. Pipelines can be executed in the browser or using the web API directly and then monitored on a document level. The evaluation in usability and user experience indicates that the implementation benefits the framework by making its usage more user friendly, lightweight, and intuitive while also making the management of pipelines more efficient.
Graph4Med: a web application and a graph database for visualizing and analyzing medical databases
(2022)
Background: Medical databases normally contain large amounts of data in a variety of forms. Although they grant significant insights into diagnosis and treatment, implementing data exploration into current medical databases is challenging since these are often based on a relational schema and cannot be used to easily extract information for cohort analysis and visualization. As a consequence, valuable information regarding cohort distribution or patient similarity may be missed. With the rapid advancement of biomedical technologies, new forms of data from methods such as Next Generation Sequencing (NGS) or chromosome microarray (array CGH) are constantly being generated; hence it can be expected that the amount and complexity of medical data will rise and bring relational database systems to a limit.
Description: We present Graph4Med, a web application that relies on a graph database obtained by transforming a relational database. Graph4Med provides a straightforward visualization and analysis of a selected patient cohort. Our use case is a database of pediatric Acute Lymphoblastic Leukemia (ALL). Along routine patients’ health records it also contains results of latest technologies such as NGS data. We developed a suitable graph data schema to convert the relational data into a graph data structure and store it in Neo4j. We used NeoDash to build a dashboard for querying and displaying patients’ cohort analysis. This way our tool (1) quickly displays the overview of patients’ cohort information such as distributions of gender, age, mutations (fusions), diagnosis; (2) provides mutation (fusion) based similarity search and display in a maneuverable graph; (3) generates an interactive graph of any selected patient and facilitates the identification of interesting patterns among patients.
Conclusion: We demonstrate the feasibility and advantages of a graph database for storing and querying medical databases. Our dashboard allows a fast and interactive analysis and visualization of complex medical data. It is especially useful for patients similarity search based on mutations (fusions), of which vast amounts of data have been generated by NGS in recent years. It can discover relationships and patterns in patients cohorts that are normally hard to grasp. Expanding Graph4Med to more medical databases will bring novel insights into diagnostic and research.
Interacting with the environment to process sensory information, generate perceptions, and shape behavior engages neural networks in brain areas with highly varied representations, ranging from unimodal sensory cortices to higher-order association areas. Recent work suggests a much greater degree of commonality across areas, with distributed and modular networks present in both sensory and non-sensory areas during early development. However, it is currently unknown whether this initially common modular structure undergoes an equally common developmental trajectory, or whether such a modular functional organization persists in some areas—such as primary visual cortex—but not others. Here we examine the development of network organization across diverse cortical regions in ferrets of both sexes using in vivo widefield calcium imaging of spontaneous activity. We find that all regions examined, including both primary sensory cortices (visual, auditory, and somatosensory—V1, A1, and S1, respectively) and higher order association areas (prefrontal and posterior parietal cortices) exhibit a largely similar pattern of changes over an approximately 3 week developmental period spanning eye opening and the transition to predominantly externally-driven sensory activity. We find that both a modular functional organization and millimeter-scale correlated networks remain present across all cortical areas examined. These networks weakened over development in most cortical areas, but strengthened in V1. Overall, the conserved maintenance of modular organization across different cortical areas suggests a common pathway of network refinement, and suggests that a modular organization—known to encode functional representations in visual areas—may be similarly engaged in highly diverse brain areas.
Significance Different areas of the mature brain encode vastly different representations of the world. This study shows that a modular functional organization where nearby neurons participate in similar functional networks is shared across different brain areas not only during early development, but also as the brain matures where it remains a shared feature that shapes neural activity. The largely conserved trajectory of developmental changes across brain areas suggests that similar circuit mechanisms may drive this maturation. This implies that the large literature on developing cortical circuits, which is largely focused on sensory areas, may also apply more broadly, and that perturbations during development that impinge on any such shared mechanisms may produce deficits that extend across multiple brain systems.
The production cross section of inclusive isolated photons has been measured by the ALICE experiment at the CERN LHC in pp collisions at centre-of-momentum energy of s√=13 TeV collected during the LHC Run 2 data-taking period. The measurement is performed by combining the measurements of the electromagnetic calorimeter EMCal and the central tracking detectors ITS and TPC, covering a pseudorapidity range of |ηγ|<0.67 and a transverse momentum range of 7<pγT<200 GeV/c. The result extends to lower pγT and xγT=2pγT/s√ ranges, the lowest xγT of any isolated photon measurements to date, extending significantly those measured by the ATLAS and CMS experiments towards lower pγT at the same collision energy with a small overlap between the measurements. The measurement is compared with next-to-leading order perturbative QCD calculations and the results from the ATLAS and CMS experiments as well as with measurements at other collision energies. The measurement and theory prediction are in agreement with each other within the experimental and theoretical uncertainties.
Assessing communicative accommodation in the context of large language models : a semiotic approach
(2023)
Recently, significant strides have been made in the ability of transformer-based chatbots to hold natural conversations. However, despite a growing societal and scientific relevancy, there are few frameworks systematically deriving what it means for a chatbot conversation to be natural. The present work approaches this question through the phenomenon of communicative accommodation/interactive alignment. While there is existing research suggesting that humans adapt communicatively to technologies, the aim of this work is to explore the accommodation of AI-chatbots to an interlocutor. Its research interest is twofold: Firstly, the structural ability of the transformer-architecture to support accommodative behavior is assessed using a frame constructed in accordance with existing accommodationtheories.
This results in hypotheses to be tested empirically. Secondly, since effective accommodation produces the same outcomes, regardless of technical implementation, a behavioral experiment is proposed. Existing quantifications of accommodation are reconciled,
extended, and modified to apply them to nonhuman-interlocutors. Thus, a measurement scheme is suggested which evaluates textual data from text-only, double-blind interactions between chatbots and humans, chatbots and chatbots and humans and humans. Using the generated human-to-human convergence data as a reference, the degree of artificial accommodation can be evaluated. Accommodation as a central facet of artificial interactivity can thus be evaluated directly against its theoretical paradigm, i.e. human interaction. In case that subsequent examinations show that chatbots effectively do not accommodate, there may be a new form of algorithmic bias, emerging from the aggregate accommodation towards chatbots but not towards humans. Thus, existing, hegemonic semantics could be cemented through chatbot-learning. Meanwhile, the ability to effectively accommodate would render chatbots vastly more susceptible to misuse.
The first measurement of the impact-parameter dependent angular anisotropy in the decay of coherently photoproduced ρ0 mesons is presented. The ρ0 mesons are reconstructed through their decay into a pion pair. The measured anisotropy corresponds to the amplitude of the cos(2ϕ) modulation, where ϕ is the angle between the two vectors formed by the sum and the difference of the transverse momenta of the pions, respectively. The measurement was performed by the ALICE Collaboration at the LHC using data from ultraperipheral Pb−Pb collisions at a center-of-mass energy of sNN−−−√ = 5.02 TeV per nucleon pair. Different impact-parameter regions are selected by classifying the events in nuclear-breakup classes. The amplitude of the cos(2ϕ) modulation is found to increase by about one order of magnitude from large to small impact parameters. Theoretical calculations, which describe the measurement, explain the cos(2ϕ) anisotropy as the result of a quantum interference effect at the femtometer scale that arises from the ambiguity as to which of the nuclei is the source of the photon in the interaction.
Investigating strangeness enhancement with multiplicity in pp collisions using angular correlations
(2024)
A study of strange hadron production associated with hard scattering processes and with the underlying event is conducted to investigate the origin of the enhanced production of strange hadrons in small collision systems characterised by large charged-particle multiplicities. For this purpose, the production of the single-strange meson K0S and the double-strange baryon Ξ± is measured, in each event, in the azimuthal direction of the highest-pT particle (``trigger" particle), related to hard scattering processes, and in the direction transverse to it in azimuth, associated with the underlying event, in pp collisions at s√=5.02 TeV and s√=13 TeV using the ALICE detector at the LHC. The per-trigger yields of K0S and Ξ± are dominated by the transverse-to-leading production (i.e., in the direction transverse to the trigger particle), whose contribution relative to the toward-leading production is observed to increase with the event charged-particle multiplicity. The transverse-to-leading and the toward-leading Ξ±/K0S yield ratios increase with the multiplicity of charged particles, suggesting that strangeness enhancement with multiplicity is associated with both hard scattering processes and the underlying event. The relative production of Ξ± with respect to K0S is higher in transverse-to-leading processes over the whole multiplicity interval covered by the measurement. The K0S and Ξ± per-trigger yields and yield ratios are compared with predictions of three different phenomenological models, namely PYTHIA 8.2 with the Monash tune, PYTHIA 8.2 with ropes and EPOS LHC. The comparison shows that none of them can quantitatively describe either the transverse-to-leading or the toward-leading yields of K0S and Ξ±.
The transverse momentum (pT) differential production cross section of the promptly-produced charm-strange baryon Ξ0c (and its charge conjugate Ξ0c¯¯¯¯¯¯) is measured at midrapidity via its hadronic decay into π+Ξ− in p−Pb collisions at a centre-of-mass energy per nucleon−nucleon collision sNN−−−√ = 5.02 TeV with the ALICE detector at the LHC. The Ξ0c nuclear modification factor (RpPb), calculated from the cross sections in pp and p−Pb collisions, is presented and compared with the RpPb of Λ+c baryons. The ratios between the pT-differential production cross section of Ξ0c baryons and those of D0 mesons and Λ+c baryons are also reported and compared with results at forward and backward rapidity from the LHCb Collaboration. The measurements of the production cross section of prompt Ξ0c baryons are compared with a model based on perturbative QCD calculations of charm-quark production cross sections, which includes only cold nuclear matter effects in p−Pb collisions, and underestimates the measurement by a factor of about 50. This discrepancy is reduced when the data is compared with a model in which hadronisation is implemented via quark coalescence. The pT-integrated cross section of prompt Ξ0c-baryon production at midrapidity extrapolated down to pT = 0 is also reported. These measurements offer insights and constraints for theoretical calculations of the hadronisation process. Additionally, they provide inputs for the calculation of the charm production cross section in p−Pb collisions at midrapidity.
Measurements of (anti)deuteron and (anti)3He production in the rapidity range |y|< 0.5 as a function of the transverse momentum and event multiplicity in Xe−Xe collisions at a center-of-mass energy per nucleon−nucleon pair of sNN−−−√ = 5.44 TeV are presented. The coalescence parameters B2 and B3 are measured as a function of the transverse momentum per nucleon. The ratios between (anti)deuteron and (anti)3He yields and those of (anti)protons and pions are reported as a function of the mean charged-particle multiplicity density, and compared with two implementations of the statistical hadronization model (SHM) and with coalescence predictions. The elliptic flow of (anti)deuterons is measured for the first time in Xe−Xe collisions and shows features similar to those already observed in Pb−Pb collisions, i.e., the mass ordering at low transverse momentum and the meson−baryon grouping at intermediate transverse momentum. The production of nuclei is particularly sensitive to the chemical freeze-out temperature of the system created in the collision, which is extracted from a grand-canonical-ensemble-based thermal fit, performed for the first time including light nuclei along with light-flavor hadrons in Xe−Xe collisions. The extracted chemical freeze-out temperature Tchem = (154.2 ± 1.1) MeV in Xe−Xe collisions is similar to that observed in Pb−Pb collisions and close to the crossover temperature predicted by lattice QCD calculations.