Refine
Year of publication
Document Type
- Preprint (759)
- Article (402)
- Working Paper (119)
- Doctoral Thesis (93)
- Diploma Thesis (47)
- Conference Proceeding (41)
- Book (37)
- Bachelor Thesis (36)
- diplomthesis (28)
- Report (25)
Has Fulltext
- yes (1619)
Is part of the Bibliography
- no (1619)
Keywords
Institute
- Informatik (1619) (remove)
The number of multilingual texts in the World Wide Web (WWW) is increasing dramatically and a multilingual economic zone like the European Union (EU) requires the availability of multilingual Natural Language Processing (NLP) tools. Due to a rapid development of NLP tools, many lexical, syntactic, semantic and other linguistic features have been used in different NLP applications. However, there are some situations where these features can not be used due the application type or unavailability of NLP resources for some of the languages. That is why an application that is intended to handle multilingual texts must have features that are not dependent on a particular language and specific linguistic tools. In this thesis, we will focus on two such applications: text readability and source and translation classification.
In this thesis, we provide 18 features that are not only suitable for both applications, but are also language and linguistic tools independent. In order to build a readability classifier, we use texts from three different languages: English, German and Bangla. Our proposed features achieve a classification accuracy that is comparable with a classifier using 40 linguistic features. The readability classifier achieves a classification F-score of 74.21% on the English Wikipedia corpus, an F-score of 75.47% on the English textbook corpus, an F-score of 86.46% on the Bangla textbook corpus and an F-score of 86.26% on the German GEO/GEOLino corpus.
We used more than two million sentence pairs from 21 European languages in order to build the source and translation classifier. The classifier using the same eighteen features achieves a classification accuracy of 86.63%. We also used the same features to build a classifier that classifies translated texts based on their origin. The classifier achieves classification accuracy of 75% for texts from 10 European languages. In this thesis, we also provide four different corpora, three for text readability analysis and one for corpus based translation studies.
Multiharmonic correlations of different flow amplitudes in Pb-Pb collisions at √sNN = 2.76 TeV
(2021)
The event-by-event correlations between three flow amplitudes are measured for the first time in Pb--Pb collisions, using higher-order Symmetric Cumulants. We find that different three-harmonic correlations develop during the collective evolution of the medium, when compared with correlations that exist in the initial state. These new results cannot be interpreted in terms of previous lower-order flow measurements, since contributions from two-harmonic correlations are explicitly removed in the new observables. Comparison with Monte Carlo simulations provides new and independent constraints for the initial conditions and system properties of nuclear matter created in heavy-ion collisions.
Poster presentation: Introduction The brain is a highly interconnected network of constantly interacting units. Understanding the collective behavior of these units requires a multi-dimensional approach. The results of such analyses are hard to visualize and interpret. Hence tools capable of dealing with such tasks become imperative. ....
The multi-strange baryon yields in PbPb collisions have been shown to exhibit an enhancement relative to pp reactions. In this work, Ξ and Ω production rates have been measured with the ALICE experiment as a function of transverse momentum, pT, in pPb collisions at a centre-of-mass energy of sNN=5.02 TeV. The results cover the kinematic ranges 0.6 GeV/c<pT<7.2 GeV/c and 0.8 GeV/c<pT<5 GeV/c, for Ξ and Ω respectively, in the common rapidity interval −0.5<yCMS<0. Multi-strange baryons have been identified by reconstructing their weak decays into charged particles. The pT spectra are analysed as a function of event charged-particle multiplicity, which in pPb collisions ranges over one order of magnitude and lies between those observed in pp and PbPb collisions. The measured pT distributions are compared to the expectations from a Blast-Wave model. The parameters which describe the production of lighter hadron species also describe the hyperon spectra in high multiplicity pPb collisions. The yield of hyperons relative to charged pions is studied and compared with results from pp and PbPb collisions. A continuous increase in the yield ratios as a function of multiplicity is observed in pPb data, the values of which range from those measured in minimum bias pp to the ones in PbPb collisions. A statistical model qualitatively describes this multiplicity dependence using a canonical suppression mechanism, in which the small volume causes a relative reduction of hadron production dependent on the strangeness content of the hyperon.
The multi-strange baryon yields in Pb--Pb collisions have been shown to exhibit an enhancement relative to pp reactions. In this work, Ξ and Ω production rates have been measured with the ALICE experiment as a function of transverse momentum, pT, in p-Pb collisions at a centre-of-mass energy of sNN−−−√ = 5.02 TeV. The results cover the kinematic ranges 0.6 GeV/c<pT<7.2 GeV/c and 0.8 GeV/c<pT< 5 GeV/c, for Ξ and Ω respectively, in the common rapidity interval -0.5 <yCMS< 0. Multi-strange baryons have been identified by reconstructing their weak decays into charged particles. The pT spectra are analysed as a function of event charged-particle multiplicity, which in p-Pb collisions ranges over one order of magnitude and lies between those observed in pp and Pb-Pb collisions. The measured pT distributions are compared to the expectations from a Blast-Wave model. The parameters which describe the production of lighter hadron species also describe the hyperon spectra in high multiplicity p-Pb. The yield of hyperons relative to charged pions is studied and compared with results from pp and Pb-Pb collisions. A statistical model is employed, which describes the change in the ratios with volume using a canonical suppression mechanism, in which the small volume causes a species-dependent relative reduction of hadron production. The calculations, in which the magnitude of the effect depends on the strangeness content, show good qualitative agreement with the data.
The multi-strange baryon yields in Pb--Pb collisions have been shown to exhibit an enhancement relative to pp reactions. In this work, Ξ and Ω production rates have been measured with the ALICE experiment as a function of transverse momentum, pT, in p-Pb collisions at a centre-of-mass energy of sNN−−−√ = 5.02 TeV. The results cover the kinematic ranges 0.6 GeV/c<pT<7.2 GeV/c and 0.8 GeV/c<pT< 5 GeV/c, for Ξ and Ω respectively, in the common rapidity interval -0.5 <yCMS< 0. Multi-strange baryons have been identified by reconstructing their weak decays into charged particles. The pT spectra are analysed as a function of event charged-particle multiplicity, which in p-Pb collisions ranges over one order of magnitude and lies between those observed in pp and Pb-Pb collisions. The measured pT distributions are compared to the expectations from a Blast-Wave model. The parameters which describe the production of lighter hadron species also describe the hyperon spectra in high multiplicity p-Pb. The yield of hyperons relative to charged pions is studied and compared with results from pp and Pb-Pb collisions. A statistical model is employed, which describes the change in the ratios with volume using a canonical suppression mechanism, in which the small volume causes a species-dependent relative reduction of hadron production. The calculations, in which the magnitude of the effect depends on the strangeness content, show good qualitative agreement with the data.
A measurement of the multi-strange Ξ− and Ω− baryons and their antiparticles by the ALICE experiment at the CERN Large Hadron Collider (LHC) is presented for inelastic proton–proton collisions at a centre-of-mass energy of 7 TeV. The transverse momentum (pT) distributions were studied at mid-rapidity (|y|<0.5) in the range of 0.6<pT<8.5 GeV/c for Ξ− and Ξ¯+ baryons, and in the range of 0.8<pT<5 GeV/c for Ω− and Ω¯+. Baryons and antibaryons were measured as separate particles and we find that the baryon to antibaryon ratio of both particle species is consistent with unity over the entire range of the measurement. The statistical precision of the current data has allowed us to measure a difference between the mean pT of Ξ− (Ξ¯+) and Ω− (Ω¯+). Particle yields, mean pT, and the spectra in the intermediate pT range are not well described by the PYTHIA Perugia 2011 tune Monte Carlo event generator, which has been tuned to reproduce the early LHC data. The discrepancy is largest for Ω− (Ω¯+). This PYTHIA tune approaches the pT spectra of Ξ− and Ξ¯+ baryons below pT<0.85 GeV/c and describes the Ξ− and Ξ¯+ spectra above pT>6.0 GeV/c. We also illustrate the difference between the experimental data and model by comparing the corresponding ratios of (Ω−+Ω¯+)/(Ξ−+Ξ¯+) as a function of transverse mass.
Multi-harmonic correlations of different flow amplitudes in Pb–Pb collisions at √sNN = 2.76 TeV
(2021)
The event-by-event correlations between three flow amplitudes are measured for the first time in Pb--Pb collisions, using higher-order Symmetric Cumulants. We find that different three-harmonic correlations develop during the collective evolution of the medium, when compared with correlations that exist in the initial state. These new results cannot be interpreted in terms of previous lower-order flow measurements, since contributions from two-harmonic correlations are explicitly removed in the new observables. Comparison with Monte Carlo simulations provides new and independent constraints for the initial conditions and system properties of nuclear matter created in heavy-ion collisions.
Multi-view microscopy techniques are used to increase the resolution along the optical axis for 3D imaging. Without this, the resolution is insufficient to resolve subcellular events. In addition, parts of the images of opaque specimens are often highly degraded or masked. Both problems motivate scientists to record the same specimen from multiple directions. The images, then have to be digitally fused into a single high-quality image. Selective-plane illumination microscopy has proven to be a powerful imaging technique due to its unsurpassed acquisition speed and gentle optical sectioning. However, even in the case of multi view imaging techniques that illuminate and image the sample from multiple directions, light scattering inside tissues often severely impairs image contrast.
Here we show that for c-elegans embryos multi view registration can be achieved based on segmented nuclei. However, segmentation of nuclei in high density distribution like c-elegans embryo is challenging. We propose a method which uses 3D Mexican hat filter for preprocessing and 3D Gaussian curvature for the post-processing step to separate nuclei. We used this method successfully on 3 data sets of c-elegans embryos in 3 different views. The result of segmentation outperforms previous methods. Moreover, we provide a simple GUI for manual correction and adjusting the parameters for different data.
We then proposed a method that combines point and voxel registration for an accurate multi view reg- istration of c-elegans embryo, which does not need any special experimental preparation. We demonstrate the performance of our approach on data acquired from fixed embryos of c-elegans worms. This multi step approach is successfully evaluated by comparison to different methods and also by using synthetic data. The proposed method could overcome the typically low resolution along the optical axis and enable stitching to- gether the different parts of the embryo available through the different views. A tool for running the code and analyzing the results is developed.
We present a measurement of e+e− pair production in central PbAu collisions at 158A GeV/c. As reported earlier, a significant excess of the e+e− pair yield over the expectation from hadron decays is observed. The improved mass resolution of the present data set, recorded with the upgraded CERES experiment at the CERN-SPS, allows for a comparison of the data with different theoretical approaches. The data clearly favor a substantial in-medium broadening of the ρ spectral function over a density-dependent shift of the ρ pole mass. The in-medium broadening model implies that baryon induced interactions are the key mechanism to the observed modifications of the ρ meson at SPS energy.
Modification of charged-particle jets in event-shape engineered Pb–Pb collisions at √sNN = 5.02 TeV
(2023)
Charged-particle jet yields have been measured in semicentral Pb−Pb collisions at center-of-mass energy per nucleon-nucleon collision sNN−−−√=5 TeV with the ALICE detector at the LHC. These yields are reported as a function of the jet transverse momentum, and further classified by their angle with respect to the event plane and the event shape, characterized by ellipticity, in an effort to study the path-length dependence of jet quenching. Jets were reconstructed at midrapidity from charged-particle tracks using the anti-kT algorithm with resolution parameters R= 0.2 and 0.4, with event-plane angle and event-shape values determined using information from forward scintillating detectors. The results presented in this letter show that, in semicentral Pb−Pb collisions, there is no significant difference between jet yields in predominantly isotropic and elliptical events. However, out-of-plane jets are observed to be more suppressed than in-plane jets. Further, this relative suppression is greater for low transverse momentum (< 50 GeV/c) R= 0.2 jets produced in elliptical events, with out-of-plane to in-plane jet-yield ratios varying up to 5.2σ between different event-shape classes. These results agree with previous studies indicating that jets experience azimuthally anisotropic suppression when traversing the QGP medium, and can provide additional constraints on the path-length dependence of jet energy loss.
In der Computergraphik werden immer wieder verschiedenste Objekte des realen Lebens modelliert. Dabei werden oft die Regeln ihres Aufbaus ausgenutzt, um diese Modelle automatisch zu erzeugen.
Gotische Architektur bietet daf¨ur gute Voraussetzungen. Auf Grund ihres hohen Grades an selbst¨ahnlichen Strukturen besteht die Möglichkeit, solche Regeln aus ihrem Aufbau abzuleiten. Wie bei vielen gotischen Elementen lassen sich auch bei den Pfeilern, die die Basis jedes gotischen Gewölbes bilden, solche Strukturen in ihrem teilweise komplexen Grundriss finden.
Die vorliegende Diplomarbeit stellt eine Methode vor, mit der die Grundrisse verschiedener gotischer Pfeiler beschrieben werden können. Die in dieser Arbeit entwickelte Querschnittsbeschreibung, wird die Darstellung der Grundrisse möglichst vieler verschiedener Pfeiler erlauben und automatisch erzeugbar sein. Der Aufbau der Beschreibung erm¨oglicht die Generierung eines 3D-Modells.
Um dies zu erreichen, wird zunächst eine Analyse der Querschnitte verschiedener gotischer Pfeiler vorgenommen. Mit den in der Analyse gewonnenen Informationen wird formal eine Querschnittsbeschreibung entwickelt, die die oben beschriebenen Anforderungen erf¨ullt. Die automatische Erzeugung erfolgt über ein parametrisches L-System. Aus der Beschreibung des Querschnitts wird schließlich das 3D-Modell erzeugt.
Die Implementierung erfolgt komplett in C++. Für die Erzeugung des 3D-Modells wird der Open Source Szenengraph Ogre3D verwendet, der die notwendige 3D-Grafik-Funktionalit¨at zur Verfügung stellt.
Mit der realisierten Anwendung ist es m¨oglich, mit wenigen Eingaben ein Modell eines komplexen gotischen Pfeilers zu erstellen.
Motivation: Arabidopsis thaliana is a well-established model system for the analysis of the basic physiological and metabolic pathways of plants. Nevertheless, the system is not yet fully understood, although many mechanisms are described, and information for many processes exists. However, the combination and interpretation of the large amount of biological data remain a big challenge, not only because data sets for metabolic paths are still incomplete. Moreover, they are often inconsistent, because they are coming from different experiments of various scales, regarding, for example, accuracy and/or significance. Here, theoretical modeling is powerful to formulate hypotheses for pathways and the dynamics of the metabolism, even if the biological data are incomplete. To develop reliable mathematical models they have to be proven for consistency. This is still a challenging task because many verification techniques fail already for middle-sized models. Consequently, new methods, like decomposition methods or reduction approaches, are developed to circumvent this problem.
Methods: We present a new semi-quantitative mathematical model of the metabolism of Arabidopsis thaliana. We used the Petri net formalism to express the complex reaction system in a mathematically unique manner. To verify the model for correctness and consistency we applied concepts of network decomposition and network reduction such as transition invariants, common transition pairs, and invariant transition pairs.
Results: We formulated the core metabolism of Arabidopsis thaliana based on recent knowledge from literature, including the Calvin cycle, glycolysis and citric acid cycle, glyoxylate cycle, urea cycle, sucrose synthesis, and the starch metabolism. By applying network decomposition and reduction techniques at steady-state conditions, we suggest a straightforward mathematical modeling process. We demonstrate that potential steady-state pathways exist, which provide the fixed carbon to nearly all parts of the network, especially to the citric acid cycle. There is a close cooperation of important metabolic pathways, e.g., the de novo synthesis of uridine-5-monophosphate, the γ-aminobutyric acid shunt, and the urea cycle. The presented approach extends the established methods for a feasible interpretation of biological network models, in particular of large and complex models.
In bioinformatics, biochemical signal pathways can be modeled by many differential equations. It is still an open problem how to fit the huge amount of parameters of the equations to the available data. Here, the approach of systematically obtaining the most appropriate model and learning its parameters is extremely interesting. One of the most often used approaches for model selection is to choose the least complex model which “fits the needs”. For noisy measurements, the model which has the smallest mean squared error of the observed data results in a model which fits too accurately to the data – it is overfitting. Such a model will perform good on the training data, but worse on unknown data. This paper propose as model selection criterion the least complex description of the observed data by the model, the minimum description length. For the small, but important example of inflammation modeling the performance of the approach is evaluated. Keywords: biochemical pathways, differential equations, septic shock, parameter estimation, overfitting, minimum description length.
It is known that deterministic finite automata (DFAs) can be algorithmically minimized, i.e., a DFA M can be converted to an equivalent DFA M' which has a minimal number of states. The minimization can be done efficiently [6]. On the other hand, it is known that unambiguous finite automata (UFAs) and nondeterministic finite automata (NFAs) can be algorithmically minimized too, but their minimization problems turn out to be NP-complete and PSPACE-complete [8]. In this paper, the time complexity of the minimization problem for two restricted types of finite automata is investigated. These automata are nearly deterministic, since they only allow a small amount of non determinism to be used. On the one hand, NFAs with a fixed finite branching are studied, i.e., the number of nondeterministic moves within every accepting computation is bounded by a fixed finite number. On the other hand, finite automata are investigated which are essentially deterministic except that there is a fixed number of different initial states which can be chosen nondeterministically. The main result is that the minimization problems for these models are computationally hard, namely NP-complete. Hence, even the slightest extension of the deterministic model towards a nondeterministic one, e.g., allowing at most one nondeterministic move in every accepting computation or allowing two initial states instead of one, results in computationally intractable minimization problems.
Anaplastic large cell lymphoma (ALCL) and classical Hodgkin lymphoma (cHL) are lymphomas that contain CD30-expressing tumor cells and have numerous pathological similarities. Whereas ALCL is usually diagnosed at an advanced stage, cHL more frequently presents with localized disease. The aim of the present study was to elucidate the mechanisms underlying the different clinical presentation of ALCL and cHL. Chemokine and chemokine receptor expression were similar in primary ALCL and cHL cases apart from the known overexpression of the chemokines CCL17 and CCL22 in the Hodgkin and Reed-Sternberg (HRS) cells of cHL. Consistent with the overexpression of these chemokines, primary cHL cases encountered a significantly denser T cell microenvironment than ALCL. Additionally to differences in the interaction with their microenvironment, cHL cell lines presented a lower and less efficient intrinsic cell motility than ALCL cell lines, as assessed by time-lapse microscopy in a collagen gel and transwell migration assays. We thus propose that the combination of impaired basal cell motility and differences in the interaction with the microenvironment hamper the dissemination of HRS cells in cHL when compared with the tumor cells of ALCL.
Um in verschiedenen Anwendungsumgebungen eingesetzt werden zu können, lässt die CORBA-Spezifikation einen weiten Spielraum für Implementierungen. Sollte CORBA in einem speziellen Umfeld eingesetzt werden, so war bisher eine Neu-Implementierung notwendig, da herkämmliche CORBA-Implementierungen nicht oder nur sehr eingeschränkt an spezielle Anwendungsumgebungen anpassbar sind. In dieser Arbeit wurde ein Ansatz für eine erweiterbare CORBA-Implementierung vorgestellt und implementiert.