Refine
Year of publication
- 2006 (49) (remove)
Document Type
- Preprint (49) (remove)
Has Fulltext
- yes (49)
Is part of the Bibliography
- no (49) (remove)
Keywords
- Chatten <Kommunikation> (3)
- Deutsch (3)
- Dialektologie (3)
- Lexicalized Tree Adjoining Grammar (2)
- RHIC (2)
- Schweizerdeutsch (2)
- Syntax (2)
- syntax (2)
- ADD (1)
- African Diaspora (1)
Institute
- Physik (24)
- Frankfurt Institute for Advanced Studies (FIAS) (16)
- Extern (10)
- Informatik (1)
- Medizin (1)
- Sportwissenschaften (1)
- Universitätsbibliothek (1)
- Zentrum für Weiterbildung (1)
Die Driften der Wörter in öffentlichen Räumen sind vielfältig. Neue Wortentwicklungen belegen unterschiedliche Interessen, "chillen" und "dissen" andere als das in der konservativen Züricher Zeitung zuerst erschienene "share-holder-value". Im Folgenden soll eine sinnbezoge Verallgemeinerung unternommen werden, die die Handlungen der Akteure mit der strukturellen Ebene verbindet. Die Veränderungen in den Verwendungen sollen zu strukturellen sozialen und sprachlichen Rahmenbedingungen in Bezug gesetzt werden. Wie werden Neuerungen und Änderungen der Anwendungsbedingungen von Wörtern vor dem Hintergrund des Wissens um die traditionelle Standardsprache und deren soziale Funktion wahrgenommen? Welche Funktionen haben Neologismen in Abgrenzung zu diesem Standard?
In this paper, we investigate the usefulness of a wide range of features for their usefulness in the resolution of nominal coreference, both as hard constraints (i.e. completely removing elements from the list of possible candidates) as well as soft constraints (where a cumulation of violations of soft constraints will make it less likely that a candidate is chosen as the antecedent). We present a state of the art system based on such constraints and weights estimated with a maximum entropy model, using lexical information to resolve cases of coreferent bridging.
The work presented here addresses the question of how to determine whether a grammar formalism is powerful enough to describe natural languages. The expressive power of a formalism can be characterized in terms of i) the string languages it generates (weak generative capacity (WGC)) or ii) the tree languages it generates (strong generative capacity (SGC)). The notion of WGC is not enough to determine whether a formalism is adequate for natural languages. We argue that even SGC is problematic since the sets of trees a grammar formalism for natural languages should be able to generate is difficult to determine. The concrete syntactic structures assumed for natural languages depend very much on theoretical stipulations and empirical evidence for syntactic structures is rather hard to obtain. Therefore, for lexicalized formalisms, we propose to consider the ability to generate certain strings together with specific predicate argument dependencies as a criterion for adequacy for natural languages.
The causative/anticausative alternation has been the topic of much typological and theoretical discussion in the linguistic literature. This alternation is characterized by verbs with transitive and intransitive uses, such that the transitive use of a verb V means roughly "cause to Vintransitive" (see Levin 1993). The discussion revolves around two issues: the first one concerns the similarities and differences between the anticausative and the passive, and the second one concerns the derivational relationship, if any, between the transitive and intransitive variant. With respect to the second issue, a number of approaches have been developed. Judging the approach conceptually unsatisfactory, according to which each variant is assigned an independent lexical entry, it was concluded that the two variants have to be derivationally related. The question then is which one of the two is basic and where this derivation takes place in the grammar. Our contribution to this discussion is to argue against derivational approaches to the causative / anticausative alternation. We focus on the distribution of PPs related to external arguments (agent, causer, instrument, causing event) in passives and anticausatives of English, German and Greek and the set of verbs undergoing the causative/anticausative alternation in these languages. We argue that the crosslinguistic differences in these two domains provide evidence against both causativization and detransitivization analyses of the causative / anticausative alternation. We offer an approach to this alternation which builds on a syntactic decomposition of change of state verbs into a Voice and a CAUS component. Crosslinguistic variation in passives and anticausatives depends on properties of Voice and its combinations with CAUS and various types of roots.
Von der welt louff vnd gestallt (3b) [Anm. 1] - vom Lauf der Welt und ihrem Zustand - handelt ein Werk, das im Zentrum der nachfolgenden Überlegungen stehen soll: die Reimchronik zum Schwaben- bzw. Schweizerkrieg des Hans Lenz vom Jahr 1499. In Form eines fiktiven Gesprächs, einer disputatz (62b) zwischen dem Autor und einem Waldbruder, werden die historische Zeitgeschichte und die damalige politisch-gesellschaftliche Situation gesichtet, geordnet, diskutiert, gedeutet und in größere, insbesondere heilsgeschichtliche Zusammenhänge gebracht. Text und Kontext stehen in diesem Beispiel (wie in der Historiographie ganz allgemein) in besonders offensichtlicher Beziehung zueinander - es leuchtet unmittelbar ein, daß ein solcher Text ohne den geschichtlichen Hintergrund nicht angemessen beurteilt werden kann. Dabei darf allerdings nicht allein danach gefragt werden, wie der Historiograph mit den geschichtlichen Fakten (soweit diese überhaupt objektiv rekonstruiert werden können!) umgeht, es muß auch dem Umfeld des Verfassers selbst und seiner Rezipienten sowie dem Zweck und der Funktion seiner Dichtung Rechnung getragen werden, den literarischen und außerliterarischen Einflüssen und Vorbildern, den Denk- und Argumentationsmustern, kurz: die "Lebenswelt" [Anm. 2] des Textes sollte zu seinem Verständnis im gesellschaftlich-kulturellen Kontext soweit als möglich erschlossen werden.
Eine Reihe von nicht in Kodifikationen des Standards aufgenommenen sprachlichen Mustern wird im Blick auf ihre Karrieren in verschiedenen mündlichen und schriftlichen Texten in einer Flut von Veröffentlichungen thematisiert, meist in der Hoffnung hier grammatische Entwicklungen und die Basis für eine Orientierung der Grammatikschreibung an der Pragmatik zu entdecken. Im Folgenden soll Sprache nicht „konzeptuell schriftlich“ gedacht und „sozusagen literal idealisiert“ werden. Es soll argumentiert werden für eine einheitliche, mit Sprachgeschichte, ontogenetischem Spracherwerb und Variantenbildung verträgliche Erklärung nicht-standardisierter sprachlicher Muster im Rahmen einer Grammatikalisierungstheorie.
This paper presents an approach to the question whether it is possible to construct a parser based on ideas from case-based reasoning. Such a parser would employ a partial analysis of the input sentence to select a (nearly) complete syntax tree and then adapt this tree to the input sentence. The experiments performed on German data from the Tüba-D/Z treebank and the KaRoPars partial parser show that a wide range of levels of generality can be reached, depending on which types of information are used to determine the similarity between input sentence and training sentences. The results are such that it is possible to construct a case-based parser. The optimal setting out of those presented here need to be determined empirically.
In recent years, research in parsing has extended in several new directions. One of these directions is concerned with parsing languages other than English. Treebanks have become available for many European languages, but also for Arabic, Chinese, or Japanese. However, it was shown that parsing results on these treebanks depend on the types of treebank annotations used. Another direction in parsing research is the development of dependency parsers. Dependency parsing profits from the non-hierarchical nature of dependency relations, thus lexical information can be included in the parsing process in a much more natural way. Especially machine learning based approaches are very successful (cf. e.g.). The results achieved by these dependency parsers are very competitive although comparisons are difficult because of the differences in annotation. For English, the Penn Treebank has been converted to dependencies. For this version, Nivre et al. report an accuracy rate of 86.3%, as compared to an F-score of 92.1 for Charniaks parser. The Penn Chinese Treebank is also available in a constituent and a dependency representations. The best results reported for parsing experiments with this treebank give an F-score of 81.8 for the constituent version and 79.8% accuracy for the dependency version. The general trend in comparisons between constituent and dependency parsers is that the dependency parser performs slightly worse than the constituent parser. The only exception occurs for German, where F-scores for constituent plus grammatical function parses range between 51.4 and 75.3, depending on the treebank, NEGRA or TüBa-D/Z. The dependency parser based on a converted version of Tüba-D/Z, in contrast, reached an accuracy of 83.4%, i.e. 12 percent points better than the best constituent analysis including grammatical functions.
This paper presents a comparative study of probabilistic treebank parsing of German, using the Negra and TüBa-D/Z treebanks. Experiments with the Stanford parser, which uses a factored PCFG and dependency model, show that, contrary to previous claims for other parsers, lexicalization of PCFG models boosts parsing performance for both treebanks. The experiments also show that there is a big difference in parsing performance, when trained on the Negra and on the TüBa-D/Z treebanks. Parser performance for the models trained on TüBa-D/Z are comparable to parsing results for English with the Stanford parser, when trained on the Penn treebank. This comparison at least suggests that German is not harder to parse than its West-Germanic neighbor language English.
Using a qualitative analysis of disagreements from a referentially annotated newspaper corpus, we show that, in coreference annotation, vague referents are prone to greater disagreement. We show how potentially problematic cases can be dealt with in a way that is practical even for larger-scale annotation, considering a real-world example from newspaper text.