Linguistik
Refine
Year of publication
- 2006 (23) (remove)
Document Type
- Preprint (9)
- Article (4)
- Part of a Book (4)
- Book (2)
- Conference Proceeding (2)
- Working Paper (2)
Language
- English (19)
- German (3)
- Portuguese (1)
Has Fulltext
- yes (23)
Is part of the Bibliography
- no (23)
Keywords
- Computerlinguistik (3)
- Deutsch (3)
- Luxemburgisch (2)
- Morphologie (2)
- Suchmaschine (2)
- Ableitung <Linguistik> (1)
- Allomorph (1)
- Benutzeroberfläche (1)
- Englisch (1)
- Everyday language (1)
Institute
- Extern (23) (remove)
Eine unübersehbare Menge neuer Anglizismen findet über Fach- und Gruppensprachen Eingang in die deutsche Alltagssprache, in der ein Teil von ihnen inzwischen seinen festen Platz hat. […] Insbesondere in den Bereichen der Lautung und der Schreibung bleibt bei den neueren Entlehnungen oberflächlich eine große Nähe zu gebersprachlichen Strukturen erhalten. Diese Entwicklung wird von einigen Fachleuten und Politikern […] als Indiz für eine schleichende ‚Kolonialisierung’ der deutschen Sprache durch das Englische herangezogen. [...] Dieser Einschätzung widersprechen zahlreiche Organe […] und Autoren […] ausdrücklich. […] Im Kontext dieser Auseinandersetzung ist die vorliegende Arbeit verortet. Ihr Ziel ist es zu zeigen, daß die Sprecher des Deutschen Anglizismen sehr wohl phonologisch, graphematisch und morphologisch in die deutsche Sprache integrieren. Untersuchungsgegenstand sind mehrgliedrige Verben, die aus dem Englischen entlehnt wurden und überwiegend in Fach- und Gruppensprachen und/oder in informellem, vorwiegend mündlichem Text auftreten. Für das Problemfeld der verbalen Wortbildung wird dargelegt, daß morphologische Integration nicht unsystematisch erfolgt, sondern sich an den Flexionsmustern deutscher komplexer Verben orientiert. Der Integrationsgrad der einzelnen Lexeme ist dynamisch und sprecherabhängig.
In this paper, we investigate the usefulness of a wide range of features for their usefulness in the resolution of nominal coreference, both as hard constraints (i.e. completely removing elements from the list of possible candidates) as well as soft constraints (where a cumulation of violations of soft constraints will make it less likely that a candidate is chosen as the antecedent). We present a state of the art system based on such constraints and weights estimated with a maximum entropy model, using lexical information to resolve cases of coreferent bridging.
Prefácio (...) O Objectivo A finalidade deste livrinho é de publicar dados de uma língua minoritária para contribuir ao património cultural da nação moçambicana, de que Imarenje faz parte. O próximo passo será a implem-entação de mais correcções e modificações necessárias, seja nos detalhes ortográficos, seja na escolha de exemplos e frases. Neste sentido faço um apelo a todos que se interessam pelo desenvolvi-mento das línguas nacionais, em particular aos falantes de Imarenje: Façam comentários, contribuam para que futuras edições deste livrinho possam ser mais ricas! Oliver Kröger Editor da série Monografias Linguísticas Moçambicanas Nampula, Outubro de 2006
This report explores the question of compatibility between annotation projects including translating annotation formalisms to each other or to common forms. Compatibility issues are crucial for systems that use the results of multiple annotation projects. We hope that this report will begin a concerted effort in the field to track the compatibility of annotation schemes for part of speech tagging, time annotation, treebanking, role labeling and other phenomena.
Die synchrone wie diachrone Untersuchung von vier Passivauxiliaren in der deutschen Standardsprache und in deutschen Dialekten, im Schwedischen und im Luxemburgischen liefert deutliche Evidenz dafür, dass Vollverben nicht direkt zu Passivauxiliaren grammatikalisieren, sondern dass dieser Pfad über die Inchoativkopula verläuft. Inchoativkopulas sind soweit grammatikalisiert (und damit reduziert), dass sie über den Weg einer Reanalyse zu Vorgangspassivauxiliaren mutieren können: Erst verbinden sie sich mit (prädikativen) Substantiven, dann mit Adjektiven und schließlich partizipialen Verben. Bereits im Kopulastadium haben sie sich (sofern vorhanden gewesen) ihres Dativ- und Akkusativobjekts entledigt (Intransitivierung). Das Subjekt ist nach seiner Entkoppelung mit dem Agens eine neue Koppelung mit dem Patiens eingegangen. Damit hat die einstige Handlungsperspektive eine Umkehr zur Geschehensperspektive erfahren. Diese Schritte dokumentiert die folgende Figur: .... Als weniger problematisch hat sich, bedingt durch die Ausgangssemantik, der Grammatikalisierungspfad bei nhd. werden, bair.lalem. kommen und schwed. bli erwiesen im Gegensatz zu lux. ginn 'geben', das in jeder Hinsicht die stärksten Reduktionen erfahren hat und einen besonders langen, verschlungenen und "steinigen" Weg absolviert hat. Mit Sicherheit kann geben nicht als Idealkandidat für Passivgrarnmatikalisierungen gelten. Nur so lässt sich erklären, weshalb diese Grarnmatikalisierung in anderen Sprachen der Welt bisher nicht beobachtet wurde.
The retreat of BE as perfect auxiliary in the history of English is examined. Corpus data are presented showing that the initial advance of HAVE was most closely connected to a restriction against BE in past counterfactuals. Other factors which have been reported to favor the spread of HAVE are either dependent on the counterfactual effect, or significantly weaker in comparison. It is argued that the effect can be traced to the semantics of the BE perfect, which denoted resultativity rather than anteriority proper. Related data from other older Germanic and Romance languages are presented, and finally implications for existing theories of auxiliary selection stemming from the findings presented are discussed.
The work presented here addresses the question of how to determine whether a grammar formalism is powerful enough to describe natural languages. The expressive power of a formalism can be characterized in terms of i) the string languages it generates (weak generative capacity (WGC)) or ii) the tree languages it generates (strong generative capacity (SGC)). The notion of WGC is not enough to determine whether a formalism is adequate for natural languages. We argue that even SGC is problematic since the sets of trees a grammar formalism for natural languages should be able to generate is difficult to determine. The concrete syntactic structures assumed for natural languages depend very much on theoretical stipulations and empirical evidence for syntactic structures is rather hard to obtain. Therefore, for lexicalized formalisms, we propose to consider the ability to generate certain strings together with specific predicate argument dependencies as a criterion for adequacy for natural languages.
This paper compares two approaches to computational semantics, namely semantic unification in Lexicalized Tree Adjoining Grammars (LTAG) and Lexical Resource Semantics (LRS) in HPSG. There are striking similarities between the frameworks that make them comparable in many respects. We will exemplify the differences and similarities by looking at several phenomena. We will show, first of all, that many intuitions about the mechanisms of semantic computations can be implemented in similar ways in both frameworks. Secondly, we will identify some aspects in which the frameworks intrinsically differ due to more general differences between the approaches to formal grammar adopted by LTAG and HPSG.
Using a qualitative analysis of disagreements from a referentially annotated newspaper corpus, we show that, in coreference annotation, vague referents are prone to greater disagreement. We show how potentially problematic cases can be dealt with in a way that is practical even for larger-scale annotation, considering a real-world example from newspaper text.
In the past, a divide could be seen between ’deep’ parsers on the one hand, which construct a semantic representation out of their input, but usually have significant coverage problems, and more robust parsers on the other hand, which are usually based on a (statistical) model derived from a treebank and have larger coverage, but leave the problem of semantic interpretation to the user. More recently, approaches have emerged that combine the robustness of datadriven (statistical) models with more detailed linguistic interpretation such that the output could be used for deeper semantic analysis. Cahill et al. (2002) use a PCFG-based parsing model in combination with a set of principles and heuristics to derive functional (f-)structures of Lexical-Functional Grammar (LFG). They show that the derived functional structures have a better quality than those generated by a parser based on a state-of-the-art hand-crafted LFG grammar. Advocates of Dependency Grammar usually point out that dependencies already are a semantically meaningful representation (cf. Menzel, 2003). However, parsers based on dependency grammar normally create underspecified representations with respect to certain phenomena such as coordination, apposition and control structures. In these areas they are too "shallow" to be directly used for semantic interpretation. In this paper, we adopt a similar approach to Cahill et al. (2002) using a dependency-based analysis to derive functional structure, and demonstrate the feasibility of this approach using German data. A major focus of our discussion is on the treatment of coordination and other potentially underspecified structures of the dependency data input. F-structure is one of the two core levels of syntactic representation in LFG (Bresnan, 2001). Independently of surface order, it encodes abstract syntactic functions that constitute predicate argument structure and other dependency relations such as subject, predicate, adjunct, but also further semantic information such as the semantic type of an adjunct (e.g. directional). Normally f-structure is captured as a recursive attribute value matrix, which is isomorphic to a directed graph representation. Figure 5 depicts an example target f-structure. As mentioned earlier, these deeper-level dependency relations can be used to construct logical forms as in the approaches of van Genabith and Crouch (1996), who construct underspecified discourse representations (UDRSs), and Spreyer and Frank (2005), who have robust minimal recursion semantics (RMRS) as their target representation. We therefore think that f-structures are a suitable target representation for automatic syntactic analysis in a larger pipeline of mapping text to interpretation. In this paper, we report on the conversion from dependency structures to fstructure. Firstly, we evaluate the f-structure conversion in isolation, starting from hand-corrected dependencies based on the TüBa-D/Z treebank and Versley (2005)´s conversion. Secondly, we start from tokenized text to evaluate the combined process of automatic parsing (using Foth and Menzel (2006)´s parser) and f-structure conversion. As a test set, we randomly selected 100 sentences from TüBa-D/Z which we annotated using a scheme very close to that of the TiGer Dependency Bank (Forst et al., 2004). In the next section, we sketch dependency analysis, the underlying theory of our input representations, and introduce four different representations of coordination. We also describe Weighted Constraint Dependency Grammar (WCDG), the dependency parsing formalism that we use in our experiments. Section 3 characterises the conversion of dependencies to f-structures. Our evaluation is presented in section 4, and finally, section 5 summarises our results and gives an overview of problems remaining to be solved.