Linguistik
Refine
Year of publication
- 2006 (16) (remove)
Document Type
- Preprint (16) (remove)
Has Fulltext
- yes (16)
Is part of the Bibliography
- no (16)
Keywords
- Dialektologie (3)
- Chatten <Kommunikation> (2)
- Deutsch (2)
- Lexicalized Tree Adjoining Grammar (2)
- Schweizerdeutsch (2)
- syntax (2)
- Alemannisch (1)
- Englisch (1)
- Grammatiktheorie (1)
- Griechisch (1)
Institute
- Extern (9)
In recent years, research in parsing has extended in several new directions. One of these directions is concerned with parsing languages other than English. Treebanks have become available for many European languages, but also for Arabic, Chinese, or Japanese. However, it was shown that parsing results on these treebanks depend on the types of treebank annotations used. Another direction in parsing research is the development of dependency parsers. Dependency parsing profits from the non-hierarchical nature of dependency relations, thus lexical information can be included in the parsing process in a much more natural way. Especially machine learning based approaches are very successful (cf. e.g.). The results achieved by these dependency parsers are very competitive although comparisons are difficult because of the differences in annotation. For English, the Penn Treebank has been converted to dependencies. For this version, Nivre et al. report an accuracy rate of 86.3%, as compared to an F-score of 92.1 for Charniaks parser. The Penn Chinese Treebank is also available in a constituent and a dependency representations. The best results reported for parsing experiments with this treebank give an F-score of 81.8 for the constituent version and 79.8% accuracy for the dependency version. The general trend in comparisons between constituent and dependency parsers is that the dependency parser performs slightly worse than the constituent parser. The only exception occurs for German, where F-scores for constituent plus grammatical function parses range between 51.4 and 75.3, depending on the treebank, NEGRA or TüBa-D/Z. The dependency parser based on a converted version of Tüba-D/Z, in contrast, reached an accuracy of 83.4%, i.e. 12 percent points better than the best constituent analysis including grammatical functions.
This paper presents an approach to the question whether it is possible to construct a parser based on ideas from case-based reasoning. Such a parser would employ a partial analysis of the input sentence to select a (nearly) complete syntax tree and then adapt this tree to the input sentence. The experiments performed on German data from the Tüba-D/Z treebank and the KaRoPars partial parser show that a wide range of levels of generality can be reached, depending on which types of information are used to determine the similarity between input sentence and training sentences. The results are such that it is possible to construct a case-based parser. The optimal setting out of those presented here need to be determined empirically.
The causative/anticausative alternation has been the topic of much typological and theoretical discussion in the linguistic literature. This alternation is characterized by verbs with transitive and intransitive uses, such that the transitive use of a verb V means roughly "cause to Vintransitive" (see Levin 1993). The discussion revolves around two issues: the first one concerns the similarities and differences between the anticausative and the passive, and the second one concerns the derivational relationship, if any, between the transitive and intransitive variant. With respect to the second issue, a number of approaches have been developed. Judging the approach conceptually unsatisfactory, according to which each variant is assigned an independent lexical entry, it was concluded that the two variants have to be derivationally related. The question then is which one of the two is basic and where this derivation takes place in the grammar. Our contribution to this discussion is to argue against derivational approaches to the causative / anticausative alternation. We focus on the distribution of PPs related to external arguments (agent, causer, instrument, causing event) in passives and anticausatives of English, German and Greek and the set of verbs undergoing the causative/anticausative alternation in these languages. We argue that the crosslinguistic differences in these two domains provide evidence against both causativization and detransitivization analyses of the causative / anticausative alternation. We offer an approach to this alternation which builds on a syntactic decomposition of change of state verbs into a Voice and a CAUS component. Crosslinguistic variation in passives and anticausatives depends on properties of Voice and its combinations with CAUS and various types of roots.
Bern, bis ins 18. Jh. Zentrum der regionalen Großmacht, heute mit nicht ganz 130.000 Einwohnern die viertgrößte Stadt der Schweiz und seit 1848 die Hauptstadt der Schweiz. Auf Grund dieser Ausgangslage würde man erwarten, dass Bern wie andere Städte eine sprachliche Strahlungskraft in die unmittelbare Umgebung aufweist. Entgegen der allgemeinen Vorstellung zeigt sich jedoch in den Karten des Sprachatlas der deutschen Schweiz (SDS) kaum eine der für die Umgebung von Städten typischen sprachgeographische Verbreitungsbilder. So finden sich viele Isoglossen in unmittelbarer Nähe der Stadt Bern: trichter-, keil- oder gar kreisförmige Bündelungen von Isoglossen, die auf eine sprachliche Wirkung der Stadt hindeuten würde, lassen sich kaum nachweisen.
Wenn Deutsch geschrieben wird, wird im Allgemeinen die standardsprachliche Form gewählt. King: nei nei nöd eso Häx ..... verschtasch mi wieder falsch :-( *sniff (bluewin.ch, #flirt60plus, 1.10.2004) Elle: HeinEr: öhm jez versteh ich gar nix mehr (Antenne Bayern #flirten40, 16.9.2005) Big: Mu auch niemand verstehen (IRCnet, #mannheim,9.2.2003) Tezo: verstehe (IRCnet, #linux, 7.1.2003) In Büchern, Zeitungen, Zeitschriften und auch im Internet ist die deutsche Standardsprache Standard. Sie ist die für die Schriftlichkeit normierte Varietät, die überregional verständlich sein soll. Diese Standardsprache ist zwar überall ähnlich, aber nicht gleich. So zeigen sich Besonderheiten im Lexikon, in der Wortbildung und vereinzelt in der Grammatik, welche in einer Region üblich sind, in der anderen jedoch nicht, oder die da eine andere Bedeutung tragen. Diese Besonderheiten sind aber nicht an einem Ort einfach falsch, sondern sie stellen regionale Ausprägungen des Standards dar (vgl. dazu das Variantenwörterbuch; Ammon u. a. 2004). ...
Die 15. Arbeitstagung zur alemannischen Dialektologie, zu der die Vorarlberger Landesregierung vom 18.–21.Septemer auf Schloss Hofen im vorarlbergischen Lochau bei Bregenz eingeladen hatte, hat wiederum einen Einblick in unterschiedliche aktuelle Forschungsaspekte der Dialektologie im Südwesten des deutschen Sprachgebiets geboten. Im dicht gedrängten Programm mit einer öffentlichen Präsentation des VALTS und des Tonarchivs der Mundarten Vorarlbergs von EUGEN GABRIEL und ARNO RUOF folgen sich an drei Tagen insgesamt 18 Vorträge, 10 Kurzvorträge, 14 Arbeitsberichte und ein dialektologischer Ausflug in den Bregenzer Wald. Im Folgenden werden die Vorträge in der präsentierten Reihenfolge referiert, während die zahlreichen Arbeitsberichte von Sprachatlanten, Wörter- und Namenbüchern sowie Tondokumentationsstellen nur pauschal erwähnt werden, welche einerseits einen Blick in die laufenden großen und größeren Projekte der alemannischen Dialektologie öffnen und andererseits die Finanzknappheit als übergreifendes Element der Dialektologie offenbaren.
Relative quantifier scope in German depends, in contrast to English, very much on word order. The scope possibilities of a quantifier are determined by its surface position, its base position and the type of the quantifier. In this paper we propose a multicomponent analysis for German quantifiers computing the scope of the quantifier, in particular its minimal nuclear scope, depending on the syntactic configuration it occurs in.
Our paper aims at capturing the distribution of negative polarity items (NPIs) within lexicalized Tree Adjoining Grammar (LTAG). The condition under which an NPI can occur in a sentence is for it to be in the scope of a negation with no quantifiers scopally intervening. We model this restriction within a recent framework for LTAG semantics based on semantic unification. The proposed analysis provides features that signal the presence of a negation in the semantics and that specify its scope. We extend our analysis to modelling the interaction of NPI licensing and neg raising constructions.
This paper presents a comparative study of probabilistic treebank parsing of German, using the Negra and TüBa-D/Z treebanks. Experiments with the Stanford parser, which uses a factored PCFG and dependency model, show that, contrary to previous claims for other parsers, lexicalization of PCFG models boosts parsing performance for both treebanks. The experiments also show that there is a big difference in parsing performance, when trained on the Negra and on the TüBa-D/Z treebanks. Parser performance for the models trained on TüBa-D/Z are comparable to parsing results for English with the Stanford parser, when trained on the Penn treebank. This comparison at least suggests that German is not harder to parse than its West-Germanic neighbor language English.
Der folgende Text betrachtet die Varietätenverwendung von Schweizer ChatterInnen und rückt dabei altersspezifische Fragen in den Vordergrund. Im Gegensatz zu vielen Versuchen, an die Sprache Jugendlicher heranzugehen, kommt hier ein quantitativer Ansatz zur Anwendung, der die Sprache der jugendlichen ChatterInnen mit der Sprache von ChatterInnen anderer Generationen vergleicht.