Refine
Year of publication
Document Type
- Conference Proceeding (746) (remove)
Language
- English (746) (remove)
Keywords
- Computerlinguistik (20)
- Informationsstruktur (16)
- Phonetik (12)
- Japanisch (9)
- Democracy (8)
- Englisch (7)
- Grammatik (7)
- Law (6)
- Maschinelle Übersetzung (6)
- Nungisch (6)
Institute
- Physik (241)
- Rechtswissenschaft (101)
- Medizin (79)
- Universitätsbibliothek (68)
- Informatik (37)
- Extern (27)
- Frankfurt Institute for Advanced Studies (FIAS) (27)
- Wirtschaftswissenschaften (14)
- Geowissenschaften (13)
- Biochemie und Chemie (11)
1. There are two classes of theories of Universal Grammar: (1) Formalist theories, such as the widespread varieties of generative grammar. These theories start from the assumption that certain strings of linguistic forms are grammatical while other strings are ungrammatical. A grammar of this type produces grammatical strings and does not produce ungrammatical ones. All theories of this class fail in the same respect: they do not account for the meaning of the strings. (2) Semiotactic theories, which describe the meaning of a string in terms of the meanings of its constituent forms and their interrelations. The only elaborate formalized theory of this class presently available is the one advanced by C.L. Ebeling (Syntax and Semantics, Leiden: Brill, 1978). I shall discuss some of its mathematical properties here.
To reach high luminosities in future linear colliders short range wakes havea to be controlled in the range of X-band frequencies or higher. Rectangular irises can be used to introduce strong focusing quadrupole-like rf-fields. Even circular irises in iris-loaded accelarator structures have the capability of focusing if the particle velocity differs from phase velocity. Theoretical investigations concerning the focusing strength to be expected are presented. Their applicability for linear colliders is discussed.
A new method of measuring quality factors in cavities is presented. This method is well suited to measure quality factors in undamped cavities as well as in heavily damped cavities, and in addition this method provides a possibility of separating modes and measuring quality factors especially in cases of overlapping modes. Measurements have been carried out on HOM-damped cavities for the DESY/THD linear collider project. Results are presented.
Damping cells for the higher order modes are necessary for the S-band linear collider to minimize BBU (Beam-Break-Up). The construction of the damper cells has to take into account the different field geometries of the higher order modes. So two different types of dampers have been designed: a wall slotted an an iris slotted cell. In order to optimize the two types of damping cells with respect to damping strength, impedance matching between coupling system and waveguide dampers and between damping cell and undamped cells and the tuning system, damping cells of both types have been built and examinated.
The lemmings theory of case
(1995)
In this paper I show that Clitic Climbing (CC) in Spanish and Long Scrambling (LS) in German (and Polish) are (im-)possible out of the same environments. For an explanation of this fact I propose a feature-oriented analysis of incorporation phenomena. The idea is that restructuring is a phenomenon of syntactic incorporation. In German and Polish, Agro incorporates covertly into the matrix clause and licenses LS out of the infinitival into the matrix clause. Similarily the clitic in Spanish, which is analysed as an Agro-head, incorporates into the matrix clause. I argue that this movement is necessary for reasons of feature-checking, i. e. for checking of an [+R]- or Restructuring-feature. In section 2 I discuss several differences between CC and LS. For example, the proposed analysis correctly predicts that clitics in contrast to scrambled phrases are subject to several serialization restrictions. Throughout the paper I use the term restructuring only in a descriptive sense, in order to describe the phenomenon in question.
Guess how?
(1996)
Given x small epsilon, Greek Rn an integer relation for x is a non-trivial vector m small epsilon, Greek Zn with inner product <m,x> = 0. In this paper we prove the following: Unless every NP language is recognizable in deterministic quasi-polynomial time, i.e., in time O(npoly(log n)), the ℓinfinity-shortest integer relation for a given vector x small epsilon, Greek Qn cannot be approximated in polynomial time within a factor of 2log0.5 − small gamma, Greekn, where small gamma, Greek is an arbitrarily small positive constant. This result is quasi-complementary to positive results derived from lattice basis reduction. A variant of the well-known L3-algorithm approximates for a vector x small epsilon, Greek Qn the ℓ2-shortest integer relation within a factor of 2n/2 in polynomial time. Our proof relies on recent advances in the theory of probabilistically checkable proofs, in particular on a reduction from 2-prover 1-round interactive proof-systems. The same inapproximability result is valid for finding the ℓinfinity-shortest integer solution for a homogeneous linear system of equations over Q.
Preferences and defaults for definiteness and number in japanese to german machine translation
(1996)
A significant problem when translating Japanese dialogues into German is the missing information on number and definiteness in the Japanese analysis output. The integration of the search for such information into the transfer process provides an efficient solution. General transfer includes conditions to make it possible to consider external knowledge. Thereby, grammatical and lexical knowledge of the source language, knowledge of lexical restrictions on the target language, domain knowledge and discourse knowledge are accessible.
The behavior of hadronic matter at high baryon densities is studied within Ultrarelativistic Quantum Molecular Dynamics (URQMD). Baryonic stopping is observed for Au+Au collisions from SIS up to SPS energies. The excitation function of flow shows strong sensitivities to the underlying equation of state (EOS), allowing for systematic studies of the EOS. Dilepton spectra are calculated with and without shifting the rho pole. Except for S+Au collisions our calculations reproduce the CERES data.
An anaphor resolution algorithm is presented which relies on a combination of strategies for narrowing down and selecting from antecedent sets for re exive pronouns, nonre exive pronouns, and common nouns. The work focuses on syntactic restrictions which are derived from Chomsky's Binding Theory. It is discussed how these constraints can be incorporated adequately in an anaphor resolution algorithm. Moreover, by showing that pragmatic inferences may be necessary, the limits of syntactic restrictions are elucidated.
In this short note on my talk I want to point out the mathematical difficulties that arise in the study of the relation of Wightman and Euclidean quantum field theory, i.e., the relation between the hierarchies of Wightman and Schwinger functions. The two extreme cases where the reconstructed Wightman functions are either tempered distributions - the well-known Osterwalder-Schrader reconstruction - or modified Fourier hyperfunctions are discussed in some detail. Finally, some perpectives towards a classification of Euclidean reconstruction theorems are outlined and preliminary steps in that direction are presented.
Due to the additional need of very short bunches for the FEL operation with the TESLA-machine strong wakefield effects are expected. One third of the total wakefield energy per bunch is radiated into the frequency region above the energy gap of Cooper pairs in superconducting niobium. The energy of the cooper pairs in superconducting niobium at 2 K corresponds to a frequency of 700 GHz. An analytical and experimental estimation for the overall energy loss of the FEL bunch above energy gap is presented. The analytical method is based on a study from R. B. Palmer [1]. The results of the wakefield estimations are used to calculate possible quality factor reduction of the TESLA cavities during FEL operation. Results are presented.
Syntactic coindexing restrictions are by now known to be of central importance to practical anaphor resolution approaches. Since, in particular due to structural ambiguity, the assumption of the availability of a unique syntactic reading proves to be unrealistic, robust anaphor resolution relies on techniques to overcome this deficiency. In this paper, two approaches are presented which generalize the verification of coindexing constraints to de cient descriptions. At first, a partly heuristic method is described, which has been implemented. Secondly, a provable complete method is specified. It provides the means to exploit the results of anaphor resolution for a further structural disambiguation. By rendering possible a parallel processing model, this method exhibits, in a general sense, a higher degree of robustness. As a practically optimal solution, a combination of the two approaches is suggested.
This paper proposes a new approach for the encoding of images by only a few important components. Classically, this is done by the Principal Component Analysis (PCA). Recently, the Independent Component Analysis (ICA) has found strong interest in the neural network community. Applied to images, we aim for the most important source patterns with the highest occurrence probability or highest information called principal independent components (PIC). For the example of a synthetic image composed by characters this idea selects the salient ones. For natural images it does not lead to an acceptable reproduction error since no a-priori probabilities can be computed. Combining the traditional principal component criteria of PCA with the independence property of ICA we obtain a better encoding. It turns out that this definition of PIC implements the classical demand of Shannon’s rate distortion theory.
Thermodynamical variables and their time evolution are studied for central relativistic heavy ion collisions from 10.7 to 160 AGeV in the microscopic Ultrarelativistic Quantum Molecular Dynamics model (UrQMD). The UrQMD model exhibits drastic deviations from equilibrium during the early high density phase of the collision. Local thermal and chemical equilibration of the hadronic matter seems to be established only at later stages of the quasi-isentropic expansion in the central reaction cell with volume 125 fm 3. Baryon energy spectra in this cell are reproduced by Boltzmann distributions at all collision energies for t > 10 fm/c with a unique rapidly dropping temperature. At these times the equation of state has a simple form: P = (0.12 - 0.15) Epsilon. At SPS energies the strong deviation from chemical equilibrium is found for mesons, especially for pions, even at the late stage of the reaction. The final enhancement of pions is supported by experimental data.
We analyze the reaction dynamics of central Pb+Pb collisions at 160 GeV/nucleon. First we estimate the energy density pile-up at mid-rapidity and calculate its excitation function: The energy density is decomposed into hadronic and partonic contributions. A detailed analysis of the collision dynamics in the framework of a microscopic transport model shows the importance of partonic degrees of freedom and rescattering of leading (di)quarks in the early phase of the reaction for E >= 30 GeV/nucleon. The energy density reaches up to 4 GeV/fm 3, 95% of which are contained in partonic degrees of freedom. It is shown that cells of hadronic matter, after the early reaction phase, can be viewed as nearly chemically equilibrated. This matter never exceeds energy densities of 0.4 GeV/fm 3, i.e. a density above which the notion of separated hadrons loses its meaning. The final reaction stage is analyzed in terms of hadron ratios, freeze-out distributions and a source analysis for final state pions.
High perveance negative ion beams with low emittance are essential for several next generation particle accelerators (i. g. spallation sources like ESS [1] and SNS [2]). The extraction and transport of these beams have intrinsic difficulties different from positive ion beams. Limitation of beam current and emittance growth have to be avoided. To fulfill the requirements of those projects a detailed knowledge of the physics of beam formation the interaction of the H- with the residual gas and transport is substantial. A compact cesium free H- volume source delivering a low energy high perveance beam (6.5 keV, 2.3 mA, perveance K= 0.0034) has been built to study the fundamental physics of beam transport and will be integrated into the existing LEBT section in the near future. First measurements of the interaction between the ion beam and the residual gas will be presented together with the experimental set up and preliminary results.
The knowledge of the build up time of space charge compensation (SCC) and the investigation of the compensation process is of main interest for low energy beam transport of pulsed high perveance ion beams under space charge compensated conditions. To investigate experimentally the rise of compensation an LEBT system consisting of a pulsed ion source, two solenoids and a drift tube as diagnostic section has been set up. The beam potential has been measured time resolved by a residual gas ion energy analyser (RGA). A numerical simulation for the calculation of self-consistent equilibrium states of the beam plasma has been developed to determine plasma parameters which are difficult measure directly. The results of the simulation has been compared with the measured data to investigate the behavior of the compensation electrons as a function of time. The acquired data shows that the theoretical rise time of space charge compensation is by a factor of two shorter than the build up time determined experimentally. In view of description the process of SCC an interpretation of the gained results is given.
This paper presents the design and implementation of a group oriented, decentralised, virtual learning settings where students meet in groups of 3 –5 people at different locations all over the world and communicate via the internet. After presenting the objective of such a didactical design the paper gives an insight into the technical implementation. It presents the advantages and disadvantages of several internet services in such a virtual setting and a way of combining these internet applications according to their special characteristics. While the role of teachers change to those of coordinators the communication process within and between the groups becomes more important – as discussed in the following chapter. The paper concludes with the presentation of two practical applications as offered by the Institute for Didactics and Economics at the Johann Wolfgang Goethe-University Frankfurt/Main (Germany) and some evaluating remarks.
Coherence in hypertext
(1999)
At first sight hypertext does not look !ike a good subject for research on coherence. Hypertext is non-linear text, and coherence is typically defined for linear text. So coherence does not seem to be involved in hypertext at all. But on closer inspection it emerges that some of the basic structural problems with hypertexts are classical problems of coherence.
Particles fullfill several distinct central roles in the Japanese language. They can mark arguments as well as adjuncts, can be functional or have semantic functions. There is, however, no straightforward matching from particles to functions, as, e.g., 'ga' can mark the subject, the object or the adjunct of a sentence. Particles can cooccur. Verbal arguments that could be identified by particles can be eliminated in the Japanese sentence. And finally, in spoken language particles are often omitted. A proper treatment of particles is thus necessary to make an analysis of Japanese sentences possible. Our treatment is based on an empirical investigation of 800 dialogues. We set up a type hierarchy of particles motivated by their subcategorizational and modificational behaviour. This type hierarchy is part of the Japanese syntax in VERBMOBIL.
This paper presents a new timing driven approach for cell replication tailored to the practical needs of standard cell layout design. Cell replication methods have been studied extensively in the context of generic partitioning problems. However, until now it has remained unclear what practical benefit can be obtained from this concept in a realistic environment for timing driven layout synthesis. Therefore, this paper presents a timing driven cell replication procedure, demonstrates its incorporation into a standard cell placement and routing tool and examines its benefit on the final circuit performance in comparison with conventional gate or transistor sizing techniques. Furthermore, we demonstrate that cell replication can deteriorate the stuck-at fault testability of circuits and show that stuck-at redundancy elimination must be integrated into the placement procedure. Experimental results demonstrate the usefulness of the proposed methodology and suggest that cell replication should be an integral part of the physical design flow complementing traditional gate sizing techniques.
Low energy beam transport (LEBT) for a future heavy ion driven inertial fusion (HIDIF [1]) facility is a crucial point using a Bi+ beam of 40 mA at 156 keV. High space charge forces (generalised perveance K=3.6*10-3) restrict the use of electrostatic focussing systems. On the other hand magnetic lenses using space charge compensation suffer from the low particle velocity. Additionally the emittance requirements are very high in order to avoid particle losses in the linac and at ring injection [2]. urthermore source noise and rise time of space charge compensation [3] might enhance particle losses and emittance. Gabor lenses [4] using a continuous space charge cloud for focussing could be a serious alternative to conventional LEBT systems. They combine strong cylinder symmetric focussing with partly space charge compensation and low emittance growth due to lower non linear fields. A high tolerance against source noise and current fluctuations and reduced investment costs are other possible advantages. The proof of principle has already been shown [5, 6]. To broaden the experiences an experimental program was started. Therefrom the first experimental results using a double Gabor lens (DGPL, see fig. 1 ) LEBT system for transporting an high perveance Xe+ beam will be presented and the results of numerical simulations will be shown.
The determination of the beam emittance using conventional destructive methods suffers from two main disadvantages. The interaction between the ion beam and the measurement device produces a high amount of secondary particles. Those particles interact with the beam and can change the transport properties of the accelerator. Particularly in the low energy section of high current accelerators like proposed for IFMIF, heavy ion inertial fusion devices (HIDIF) and spallation sources (ESS, SNS) the power deposited on the emittance measurement device can lead to extensive heat on the detector itself and can destruct or at least dejust the device (slit or grit for example). CCD camera measurements of the incident light emitted from interaction of beam ions with residual gas are commonly used for determination of the beam emittance. Fast data acquisition and high time resolution are additional features of such a method. Therefore a matrix formalism is used to derive the emittance from the measured profile of the beam [1,2] which does not take space charge effects and emittance growth into account. A new method to derive the phase space distribution of the beam from a single CCD camera image using statistical numerical methods will be presented together with measurements. The results will be compared with measurements gained from a conventional Allison type (slit-slit) emittance measurement device.
In the current globalization debate the law appears to be entangled in economic and political developments which move into a new dimension of depoliticization, de-centralization and de-individualization. For all the correct observations in detail, though, this debate is bringing about a drastic (polit)economic reduction of the role of law in the globalization process that I wish to challenge in this paper. Here one has to take on Wallerstein’s misconception of “worldwide economies” according to which the formation of the global society is seen as a basically economic process. Autonomous globalization processes in other social spheres running parallel to economic globalization need to be taken seriously. In protest against such (polit)economic reductionism several strands of the debate, among them the neo-institutionalist theory of “global culture”, post-modern concepts of global legal pluralism, systems theory studies of differentiated global society and various versions of “global civil society” have shaped a concept of a polycentric globalization. From these angles the remarkable multiplicity of the world society, in which tendencies to re-politicization, re-regionalization and re-individualization are becoming visible at the same time, becomes evident. I shall contrast two current theses on the globalization of law with two less current counter-theses: First thesis: globalization is relevant for law because the emergence of global markets undermines the control potential of national policy, and therefore also the chances of legal regulation. First counter-thesis: globalization produces a set of problems intrinsic to law itself, consisting in a change to the dominant lawmaking processes. Second thesis: globalization means that the law institutionalizes the worldwide shift in power from governmental actors to economic actors. Second counter-thesis: globalization means that the law has a chance of contributing to a dual constitution of autonomous sectors of world society.
We present a solution for the representation of Japanese honorifical information in the HPSG framework. Basically, there are three dimensions of honorification. We show that a treatment is necessary that involves both the syntactic and the contextual level of information. The japanese grammar is part of a machine translation system.
Expletives as features
(2000)
Expletives have always been a central topic of theoretical debate and subject to different analyses within the different stages of the Principles and Parameter theory (see Chomsky 1981, 1986, 1995; Lasnik 1992, 1995; Frampton and Gutman 1997; among others). However, most analyses center on the question how to explain the behavior of expletives in A-chains (such as there in English or Þad in Icelandic). No account relates wh-expletives (as one finds them in so-called partial wh-movement constructions in languages such as Hungarian, Romani, and German) to expletives in Achains. In this paper, I argue that the framework of the Minimalist Program opens up the possibility of accounting for expletive-associate relations in A-/A'-chains in a unified manner. The main idea of the unitary analysis is that an expletive is an overtly realized feature bundle that is (sub)extracted from its associate DP. There in an expletive-associate chain is a moved D-feature which orginates inside the associate DP. Similarily, in A'-chains, the whexpletive originates as a focus-/wh-feature in the wh-phrase with which it is associated. This analysis provides evidence for the feature-checking theory in Chomsky (1995). The paper is organized as follows. Section 2 contains the discussion of expletive there. In section 3 I suggest an analysis for whexpletives, and I also explore whether this analysis can be extended to relations between X°-categories such as auxiliary and participle complexes.
This paper describes a first version of the GPS flight recorder for homing pigeons. The GPS recorder consists of a hybrid GPS board, a patch antenna 19*19 mm, a 3 V Lithium battery as power supply, a DCDC converter, a logging facility and an additional microprocessor. It has a weight of 33g. Prototypes were tested and worked reliably with a sampling rate of 1/sec and with an operation time of about 3 h. In first tests on homing pigeons 9 flight paths were recorded, showing details like loops flown immediately after the release, complete routes over 30 km including detours, rest periods and speed.
What role does language play in the development of numerical cognition? In the present paper I argue that the evolution of symbolic thinking (as a basis for language) laid the grounds for the emergence of a systematic concept of number. This concept is grounded in the notion of an infinite sequence and encompasses number assignments that can focus on cardinal aspects ("three pencils"), ordinal aspects ("the third runner"), and even nominal aspects ("bus #3"). I show that these number assignments are based on a specific association of relational structures, and that it is the human language faculty that provides a cognitive paradigm for such an association, suggesting that language played a pivotal role in the evolution of systematic numerical cognition.
I discuss the status of WH-words for interrogative interpretations, and show that the derivation of constituent questions evolves from a specific interplay of syntactic and semantic representations with pragmatics. I argue that WH-pronouns are not ‘interrogative’. Rather, they are underspecified elements; due to this underspecification, WH-words can form a constitutive part not only of interrogative, but also of exclamative and declarative clauses. WH-words introduce a variable of a particular conceptual domain into the semantic representation. Accordingly, they have to be specified for interpretation. Different WH-contexts give rise to different interpretations. In a cross-linguistic overview, I discuss the characteristic elements contributing to the derivation of interrogatives. I argue that specific particles or their phonologically empty counterparts in the head of CP contribute the interrogative aspect. The speech act of ‘asking’ is then carried out via an intonational contour that identifies a question. By default, this intonational contour operates on interrogative sentences; however, other sentence formats – in particular, those of declarative sentences – are possible as well. The distinction of (a) grammatical (syntactic, semantic and phonological) sentence formats for interrogative and declarative sentences, and (b) intonational contours serving the discrimination of speech acts like questions and assertions, can be related to psychological and neurological evidence.
In its first part, this contribution reviews shortly the application of neural network methods to medical problems and characterizes its advantages and problems in the context of the medical background. Successful application examples show that human diagnostic capabilities are significantly worse than the neural diagnostic systems. Then, paradigm of neural networks is shortly introduced and the main problems of medical data base and the basic approaches for training and testing a network by medical data are described. Additionally, the problem of interfacing the network and its result is given and the neuro-fuzzy approach is presented. Finally, as case study of neural rule based diagnosis septic shock diagnosis is described, on one hand by a growing neural network and on the other hand by a rule based system. Keywords: Statistical Classification, Adaptive Prediction, Neural Networks, Neurofuzzy, Medical Systems
In linguistics and the philosophy of language, the mass/count distinction has traditionally been regarded as a bi-partition on the nominal domain, where typical instances are nouns like "beef" (mass) vs."cow" (count). In the present paper, we argue that this partition reveals a system that is based on both syntactic features and conceptual features, and present experimental evidence suggesting that the discrimination of the two kinds of features has a psychological reality.
The experiment NA49 at the CERN SPS is a large acceptance detector for charmed hadrons. The identification of neutral strange hadrons Lambda and AntiLambda is based on the measurement of their charged decay particles and the reconstruciton of the decay vertex. The charged particles were measured with the 4 time projection chambers (TPC), two of them are situated inside 2 large dipole magnets, the two others are downstream of the magnet. Lambda and AntiLambda baryons have been measured in central Pb+Pb collisions at 40, 80 and 160 GeV/nucleon over a wide range in rapidity (1 - 5) and transverse momentum (0 - 3 GeV/c). Particle yields and spectra will be shown for the different energies. The results will be put into the existing systematics of Lambda-production as a function of beam energy.
A model is proposed that interprets a variety of connected speech processes as resulting from prosodic modulations at different tiers of functional speech motor control along the hypo-hyper dimension [10]. The general background of the model is given by the trichotomy of A-, B- and C-prosodic phenomena [15] that together constitute the acoustic makeup of any speech utterance (with regard to their respective time domains at the uttarance/phrase level, the syllabic level and the segmental level).
Evaluating the quality of credit portfolio risk models is an important issue for both banks and regulators. Lopez and Saidenberg (2000) suggest cross-sectional resampling techniques in order to make efficient use of available data. We show that their proposal disregards cross-sectional dependence in resampled portfolios, which renders standard statistical inference invalid. We proceed by suggesting the Berkowitz (1999) procedure, which relies on standard likelihood ratio tests performed on transformed default data. We simulate the power of this approach in various settings including one in which the test is extended to incorporate cross-sectional information. To compare the predictive ability of alternative models, we propose to use either Bonferroni bounds or the likelihood-ratio of the two models. Monte Carlo simulations show that a default history of ten years can be sufficient to resolve uncertainties currently present in credit risk modeling.
We introduce a new method for representing and solving a general class of non-preemptive resource-constrained project scheduling problems. The new approach is to represent scheduling problems as de- scriptions (activity terms) in a language called RSV, which allows nested expressions using pll, seq, and xor. The activity-terms of RSV are similar to concepts in a description logic. The language RSV generalizes previous approaches to scheduling with variants insofar as it permits xor's not only of atomic activities but also of arbitrary activity terms. A specific semantics that assigns their set of active schedules to activity terms shows correctness of a calculus normalizing activity terms RSV similar to propositional DNF-computation. Based on RSV, this paper describes a diagram-based algorithm for the RSV-problem which uses a scan-line principle. The scan-line principle is used for determining and resolving the occurring resource conflicts and leads to a nonredundant generation of all active schedules and thus to a computation of the optimal schedule.
In this paper we show an approach to the customization of GermaNet to the German HPSG grammar lexicon developed in the Verbmobil project. GermaNet has a broad coverage of the German base vocabulary and fine-grained semantic classification; while the HPSG grammar lexicon is comparatively small und has a coarse-grained semantic classification. In our approach, we have developed a mapping algorithm to relate the synsets in GermaNet with the semantic sorts in HPSG. The evaluation result shows that this approach is useful for the lexical extension of our deep grammar development to cope with real-world text understanding.
A contrast to a trace
(2001)
For movement, such as quantifier raising, the three different structures illustrated in (1) are discussed in the recent literature.
(1) A girl danced with every boy
a. [every boy]x a girl danced with x (copy + replace)
b. [every boy]x a girl danced with [every boy] (copy)
c. [every boy]x a girl danced with [thex boy] (copy + modify)
In this paper, I'll call the proposal illustrated by (1a) the copy+replace theory since the movement is analyzed as first copying the moving phrase followed by replacing the moving phrase with a trace in the base position of movement. Chomsky (1993) and Fox (1999) argue against the copy+replace theory (1a) on the basis of Condition C data that show that moved material can behave as if it occupied the base position of movement. This behavior would, for example, be expected on the copy theory of movement illustrated by (1b), which also seems conceptually simpler than the copy+replace theory since it involves only copying without replacement. This conceptual advantage, however, is probably only apparent since a theory of the interpretation of structures like (1b) would probably be more complicated than for (1a). Standard assumptions about interpretation, at least, don't predict the right meaning when applied to (1b). For this reason, Chomsky and Fox propose what I'll call the copy+modify-theory illustrated in (1c). This proposes that copying is followed by a trace modification operation that replaces the determiner of the moved DP with something else. I assume that this is an indexed definite determiner, the interpretation of which is to be clarified below.
We present new concepts to integrate logic synthesis and physical design. Our methodology uses general Boolean transformations as known from technology-independent synthesis, and a recursive bi-partitioning placement algorithm. In each partitioning step, the precision of the layout data increases. This allows effective guidance of the logic synthesis operations for cycle time optimization. An additional advantage of our approach is that no complicated layout corrections are needed when the netlist is changed.
Retiming is a widely investigated technique for performance optimization. It performs powerful modifications on a circuit netlist. However, often it is not clear, whether the predicted performance improvement will still be valid after placement has been performed. This paper presents a new retiming algorithm using a highly accurate timing model taking into account the effect of retiming on capacitive loads of single wires as well as fanout systems. We propose the integration of retiming into a timing-driven standard cell placement environment based on simulated annealing. Retiming is used as an optimization technique throughout the whole placement process. The experimental results show the benefit of the proposed approach. In comparison with the conventional design flow based on standard FEAS our approach achieved an improvement in cycle time of up to 34% and 17% on the average.
Retiming is a widely investigated technique for performance optimization. In general, it performs extensive modifications on a circuit netlist, leaving it unclear, whether the achieved performance improvement will still be valid after placement has been performed. This paper presents an approach for integrating retiming into a timing-driven placement environment. The experimental results show the benefit of the proposed approach on circuit performance in comparison with design flows using retiming only as a pre- or postplacement optimization method.
The languages of the world differ with respect to argument extraction possibilities. In languages such as English, wh-movement is possible from Spec IP and from the complement position, whereas in languages such as Malagasy only extraction from Spec IP is possible. This difference correlates with the fact that these language types obey different island constraints and behave differently with respect to wh-in situ and superiority effects. The goal of this paper is to outline an analysis for these differences. The basic idea is that in contrast to languages such as English, in Malagasy-type languages every argument can be merged in the complement position of the selecting head.
Flight paths of homing pigeons were measured with a newly developed recorder based on GPS. The device consists of a GPS receiver board, a logging facility, an antenna, a power supply, a DCDC converter and a casing. It has a weight of 33g and works reliably with a sampling rate of 1/s with an operation time of about 3 h, providing timeindexed data on geographic positions, ground speed and altitude. The data are downloaded when the bird is recaptured. The devices are fixed to the birds with a harness. The measured complete flight paths show many details: e.g. initial loops flown immediately after release and large detours flown by some pigeons. We are here presenting 3 examples of flight paths from a release site 17.3 km Northeast of the home loft in Frankfurt. Mean speed in flight, duration of breaks and length of the flight path were calculated. The pigeons chose different routes and have different individual tendencies to fly loops over the village close to the release site.
The early prediction of mortality is one of the unresolved tasks in intensive care medicine. This contribution models medical symptoms as observations cased by transitions between hidden markov states. Learning the underlying state transition probabilities results in a prediction probability success of about 91%. The results are discussed and put in relation to the model used. Finally, the rationales for using the model are reflected: Are there states in the septic shock data?
This article examines the expression of natural gender in Icelandic nouns denoting human beings. Particular attention will be paid to the system's symmetry with regards to nouns denoting women and men. Our society consists more or less exactly of half women and half men. One would therefore assume that systems for terms denoting persons would also be symmetrically organised. Yet this assumption could not be further from the truth, and not just in single isolated cases, but in many languages: I will attempt to show that Icelandic has numerous methods for referring to women, but also many barriers and idiosyncrasies.
The present study poses the question on what phonetic and phonological grounds postalveolar fricatives in Polish can be analyzed as retroflex and whether postalveolar fricatives in other Slavic languages are retroflex as well. Velarization and incompatibility with front vowels are introduced as articulatory criteria for retroflexion, based on crosslinguistic data. According to these criteria, Polish and Russian have retroflex fricatives, whereas Bulgarian and Czech do not. In a phonological representation of these Slavic retroflexes, the necessity of perceptual features is shown. Lastly, it is illustrated that palatalization of retroflex fricatives both in Slavic languages and more generally causes a phonetic and phonological change to a non-retroflex sound.
Vortrag gehalten an der Tagung "The XVI International Conference on Ultrarelativistic Nucleus-Nucleus Collisions, organized by SUBATECH Laboratory", in Nantes, France, 18-24 Juli 2002.
Düsseldorf is the capital of North Rhine-Westphalia, the state with the highest number of inhabitants in Germany. The city has a rich cultural history: The theatre history started in 1485 (the festivities in the context of a princely marriage at Düsseldorf). Theatre historiography marks three great periods for Düsseldorf (Immermann, 1834-1837; Dumont-Lindemann, 1905-1933; Gründgens, 1947-1955). The city has a long history of involvment with film, too. For instance the first German film journal „Der Kinematogaph“ began publishing here in 1907. Düsseldorf became after 1945 a distribution center and served for decades as site of all major German and foreign distributors‘ headquartes. It offers still a lot of cultural events: performing arts in different forms (theatre at the Düsseldorfer Schauspielhaus, musical at the Capitol, opera and ballet at the German Opera Düsseldorf-Duisburg, dance at the Tanzhaus (Dance House) North Rhine-Westphalia, free and independant theatre groups, private theatres, cinema, media, museums, cultural institutions, representing other countries like France, Poland ...
Based on a detailed case study of parallel grammar development distributed across two sites, we review some of the requirements for regression testing in grammar engineering, summarize our approach to systematic competence and performance profiling, and discuss our experience with grammar development for a commercial application. If possible, the workshop presentation will be organized around a software demonstration.
We present a broad coverage Japanese grammar written in the HPSG formalism with MRS semantics. The grammar is created for use in real world applications, such that robustness and performance issues play an important role. It is connected to a POS tagging and word segmentation tool. This grammar is being developed in a multilingual context, requiring MRS structures that are easily comparable across languages.
We present an architecture for the integration of shallow and deep NLP components which is aimed at flexible combination of different language technologies for a range of practical current and future applications. In particular, we describe the integration of a high-level HPSG parsing system with different high-performance shallow components, ranging from named entity recognition to chunk parsing and shallow clause recognition. The NLP components enrich a representation of natural language text with layers of new XML meta-information using a single shared data structure, called the text chart. We describe details of the integration methods, and show how information extraction and language checking applications for realworld German text benefit from a deep grammatical analysis.
Rating agencies state that they take a rating action only when it is unlikely to be reversed shortly afterwards. Based on a formal representation of the rating process, I show that such a policy provides a good explanation for the puzzling empirical evidence: Rating changes occur relatively seldom, exhibit serial dependence, and lag changes in the issuers’ default risk. In terms of informational losses, avoiding rating reversals can be more harmful than monitoring credit quality only twice per year.
In the last years, much effort went into the design of robust anaphor resolution algorithms. Many algorithms are based on antecedent filtering and preference strategies that are manually designed. Along a different line of research, corpus-based approaches have been investigated that employ machine-learning techniques for deriving strategies automatically. Since the knowledge-engineering effort for designing and optimizing the strategies is reduced, the latter approaches are considered particularly attractive. Since, however, the hand-coding of robust antecedent filtering strategies such as syntactic disjoint reference and agreement in person, number, and gender constitutes a once-for-all effort, the question arises whether at all they should be derived automatically. In this paper, it is investigated what might be gained by combining the best of two worlds: designing the universally valid antecedent filtering strategies manually, in a once-for-all fashion, and deriving the (potentially genre-specific) antecedent selection strategies automatically by applying machine-learning techniques. An anaphor resolution system ROSANA-ML, which follows this paradigm, is designed and implemented. Through a series of formal evaluations, it is shown that, while exhibiting additional advantages, ROSANAML reaches a performance level that compares with the performance of its manually designed ancestor ROSANA.
A common topic in recent literature on phonology is the question of whether phonological processes and segments are licensed by prosodic position or by perceptual cues. The former is the traditional view, as represented by e.g. Lombardi (1995) and Beckman (1998), and holds that segments occur in specific prosodic positions such as the coda. In a licensing by cue approach, as represented by Steriade (1995, 1999), on the other hand, segments are assumed to occur in those positions only where their perceptual cues are prominent, independent of the prosodic position. In positions where the cues are not salient, neutralization occurs.
A new semantics for number
(2003)
Semantic research over the past three decades has provided impressive confirmation of Donald Davidsons famous claim that “there is a lot of language we can make systematic sense of if we suppose events exist” (Davidson 1980:137). Nowadays, Davidsonian event arguments are no longer reserved only for action verbs (as Davidson originally proposed) or even only for the category of verbs, but instead are widely assumed to be associated with any kind of predicate (e.g. Higginbotham 2000, Parsons 2000).1 The following quotation from Higginbotham and Ramchand (1997) illustrates the reasoning that motivates this move: "Once we assume that predicates (or their verbal, etc. heads) have a position for events, taking the many consequences that stem therefrom, as outlined in publications originating with Donald Davidson (1967), and further applied in Higginbotham (1985, 1989), and Terence Parsons (1990), we are not in a position to deny an event-position to any predicate; for the evidence for, and applications of, the assumption are the same for all predicates. (Higginbotham and Ramchand 1997:54)" In fact, since Davidson’s original proposal the burden of proof for postulating event arguments seems to have shifted completely, leading Raposo and Uriagereka (1995), for example, to the following verdict: "it is unclear what it means for a predicate not to have a Davidsonian argument (Raposo and Uriagereka 1995:182)" That is, Davidsonian eventuality arguments apparently have become something like a trademark for predicates in general. The goal of the present paper is to subject this view of the relationship between predicates and events to real scrutiny. By taking a closer look at the simplest independent predicational structure – viz. copula sentences – I will argue that current Davidsonian approaches tend to stretch the notion of events too far, thereby giving up much of its linguistic and ontological usefulness. More specifically, the paper will tackle the following three questions: 1. Do copula sentences support the current view of the inherent event-relatedness of predicates? 2. If not, what is a possible alternative to an event-based analysis of copula sentences? 3. What does this tell us about Davidsonian events? The paper is organized as follows: Section 2 first reviews current event-based analyses of copula sentences and then gives a brief summary of the Davidsonian notion of events. Section 3 examines the behavior of copula sentences with respect to some standard (as well as some new) eventuality diagnostics. Copula expressions will turn out to fail all eventuality tests. They differ sharply from state verbs like stand, sit, sleep in this respect. (The latter pass all eventuality tests and therefore qualify as true “Davidsonian state” expressions.) On the basis of these observations, section 4 provides an alternative account of copula sentences that combines Kim’s (1969, 1976) notion of property exemplifications with Ashers (1993, 2000) conception of abstract objects. Specifically, I will argue that the copula introduces a referential argument for a temporally bound property exemplification (= “Kimian state”). The proposal is implemented within a DRT framework. Finally, section 5 offers some concluding remarks and suggests that supplementing Davidsonian eventualities by Kimian states not only yields a more adequate analysis for copula expressions and the like but may also improve our treatment of events.
The paper is structured as follows. Section 2.1 introduces the basic classes of adjectives that constitute the factual core of the paper. Section 2.2 summarizes in greater detail the X° and the XP movement approaches to word order variation within the DP. Section 3 briefly discusses problems for both approaches. Sections 4.1, 5.1, and 5.2 draw from Alexiadou (2001) and contain a discussion of Greek DS and its relevance for a re-analysis of the word order variation in the Romance DP. Section 4.2 introduces refinements to Alexiadou & Wilder (1998) and Alexiadou (2001). Section 5.3. discusses certain issues that arise from the analysis of postnominal adjectives in Romance as involving raising of XPs. Section 6 discusses phenomena found in other languages, which at first sight seem similar to DS. However, I show that double definiteness in e.g. Hebrew, Scandinavian or other Balkan languages constitutes a different type of phenomenon from Greek DS, thus making a distinction between determiners that introduce CPs (Greek) and those that are merely morphological/agreement markers (Hebrew, Scandinavian, Albanian).
On the basis of perceptual experiments we show that alveolo-palatal fricatives and palatalized post-alveolars are two separate sounds which are distinguished not only by Polish native speakers but also by German ones. This claim is partly attested by centre of gravity measurements of the two sibilants. In this paper we revise the claim made by Halle & Stevens [1] and Maddieson & Ladefoged [2] that the Polish alveolo-palatal fricatives [˛, ¸] are palatalized postalveolars [SJ, ZJ]. On the basis of perceptual experiments we show that alveolo-palatal fricatives and palatalized post-alveolars are two separate sounds which are distinguished not only by Polish native speakers but also by German ones. This claim is partly attested by centre of gravity measurements of the two sibilants.
Twenty years ago I discussed the oldest isoglosses in the South Slavic linguistic area (1982). Subscribing to Van Wijk’s view that the bundle of isoglosses which separates Bulgarian from Serbo-Croatian was the result of an early split in South Slavic and that the transitional dialects originated from a later mixture of Serbian and Bulgarian dialects when the contact between the two languages had been restored (1927), I argued that the shared innovations of Bulgarian and Serbo-Croatian must be dated to a period when the dialects were still spoken in the original Trans-Carpathian homeland of the Slavs. I concluded that there is no evidence for common innovations of South Slavic which were posterior to the end of what I have called the Late Middle Slavic period, which I dated to the 4th through 6th centuries AD. At that time, the major dialect divisions of Slavic were already established.
Presentation at the Università di Pisa, Pisa, Itlay 3 July 2002, the conference on Irreversible Quantum Dynamics', the Abdus Salam ICTP, Trieste, Italy, 29 July - 2 August 2002, and the University of Natal, Pietermaritzburg, South Africa, 14 May 2003. Version of 24 April 2003: examples added; 16 December 2002: revised; 12 Sptember 2002. See the corresponding papers "Zeno Dynamics of von Neumann Algebras", "Zeno Dynamics in Quantum Statistical Mechanics" and "Mathematics of the Quantum Zeno Effect"
We present a novel practical algorithm that given a lattice basis b1, ..., bn finds in O(n exp 2 *(k/6) exp (k/4)) average time a shorter vector than b1 provided that b1 is (k/6) exp (n/(2k)) times longer than the length of the shortest, nonzero lattice vector. We assume that the given basis b1, ..., bn has an orthogonal basis that is typical for worst case lattice bases. The new reduction method samples short lattice vectors in high dimensional sublattices, it advances in sporadic big jumps. It decreases the approximation factor achievable in a given time by known methods to less than its fourth-th root. We further speed up the new method by the simple and the general birthday method. n2
Twenty years ago (1983), I severely criticized Halle and Kiparsky’s review (1981) of Garde’s history of Slavic accentuation (1976). I concluded that Halle and Ki-parsky’s theoretical framework “rests upon an unwarranted limitation of the available evidence, obscures the chronological perspective, and yields results which are partly not new and partly incorrect. It is harmful because it does not give the facts their proper due and thereby blocks the road to empirical study, giving a free hand to unrestrained speculation” (1983: 40). As Halle has recently returned to the subject (2001), it may be interesting to see if there has been some progress in his thinking over the last two decades. In the following I shall try to avoid repeating what I have said in my earlier discussion.
Bilabial stops undergoing Surface Palatalization (SP) were analyzed in an EMMA/EPG study. Articulatorily, the point of maximal palatal contact and the labial opening movement were analyzed. The acoustic analysis pertained to stop related timing and the point of the highest F2-value. Results show (i) that SP yields a higher F2 at vowel onset and a lengthened opening gesture and (ii) that morphemeinduced palatalizations are distinguished from word initial ones and sandhi-palatalizations articulatorily and acoustically by a shorter delay of palatal target position with respect to stop production; (iii) no differences are found between ‘repalatalized’ and plain segments in case of sandhi palatalization.
US investors hold much less foreign stocks than mean/variance analysis applied to historical data predicts. In this article, we investigate whether this home bias can be explained by Bayesian approaches to international asset allocation. In contrast to mean/variance analysis, Bayesian approaches employ different techniques for obtaining the set of expected returns. They shrink sample means towards a reference point that is inferred from economic theory. We also show that one of the Bayesian approaches leads to the same implications for asset allocation as mean-variance/tracking error criterion. In both cases, the optimal portfolio is a combination the market portfolio and the mean/variance efficient portfolio with the highest Sharpe ratio.
Applying the Bayesian approaches to the subject of international diversification, we find that substantial home bias can be explained when a US investor has a strong belief in the global mean/variance efficiency of the US market portfolio and when he has a high regret aversion falling behind the US market portfolio. We also find that the current level of home bias can justified whenever regret aversion is significantly higher than risk aversion.
Finally, we compare the Bayesian approaches to mean/variance analysis in an empirical out-ofsample study. The Bayesian approaches prove to be superior to mean/variance optimized portfolios in terms of higher risk-adjusted performance and lower turnover. However, they not systematically outperform the US market portfolio or the minimum-variance portfolio.
Over-allotment arrangements are nowadays part of almost any initial public offering. The underwriting banks borrow stocks from the previous shareholders to issue more than the initially announced number of shares. This is combined with the option to cover this short position at the issue price. We present empirical evidence on the value of these arrangements to the underwriters of initial public offerings on the Neuer Markt. The over-allotment arrangement is regarded as a portfolio of a long call option and a short position in a forward contract on the stock, which is different from other approaches presented in the literature.
Given the economically substantial values for these option- like claims we try to identify benefits to previous shareholders or new investors when the company is using this instrument in the process of going public. Although we carefully control for potential endogeneity problems, we find virtually no evidence for a reduction in underpricing for firms using over-allotment arrangements. Furthermore, we do not find evidence for more pronounced price stabilization activities or better aftermarket performance for firms granting an over-allotment arrangement to the underwriting banks.
EFM Classification: 230, 410
We present an effort for the development of multilingual named entity grammars in a unification-based finite-state formalism (SProUT). Following an extended version of the MUC7 standard, we have developed Named Entity Recognition grammars for German, Chinese, Japanese, French, Spanish, English, and Czech. The grammars recognize person names, organizations, geographical locations, currency, time and date expressions. Subgrammars and gazetteers are shared as much as possible for the grammars of the different languages. Multilingual corpora from the business domain are used for grammar development and evaluation. The annotation format (named entity and other linguistic information) is described. We present an evaluation tool which provides detailed statistics and diagnostics, allows for partial matching of annotations, and supports user-defined mappings between different annotation and grammar output formats.
Coreference-Based Summarization and Question Answering: a Case for High Precision Anaphor Resolution
(2003)
Approaches to Text Summarization and Question Answering are known to benefit from the availability of coreference information. Based on an analysis of its contributions, a more detailed look at coreference processing for these applications will be proposed: it should be considered as a task of anaphor resolution rather than coreference resolution. It will be further argued that high precision approaches to anaphor resolution optimally match the specific requirements. Three such approaches will be described and empirically evaluated, and the implications for Text Summarization and Question Answering will be discussed.
0. Introduction 1. Observations concerning the structure of morphosyntactically marked focus constructions 1.1 First observation: SF vs. NSF asymmetry 1.2 Second observation: NSF-NAR parallelism 1.3 Affirmative ex-situ focus constructions (SF, NSF), and narrative clauses (NAR) 2. Grammaticalization 2.1 Cleft hypothesis 2.2 Movement hypothesis 2.3 Narrative hypothesis 2.3.1 Back- or Foregrounding? 2.3.2 Converse directionality of FM and conjunction 3. Language specific analysis 4. Conclusionary remarks References
A pragmatic explanation of the stage level/individual level contrast in combination with locatives
(2004)
One important difference between stage level predicates (SLPs) and individual level predicates (ILPs) is their behavior with respect to locative modifiers. It is commonly assumed that SLPs but not ILPs combine with locatives. The present study argues against a semantic account for this behavior (as advanced by e.g. Kratzer 1995, Chierchia 1995) and proposes a genuinely pragmatic explanation of the observed stage level/individual level contrast instead. The proposal is spelled out using Blutners (1998, 2000) optimality theoretic version of the Gricean maxims. Building on the observation that the respective locatives are not event-related but frame-setting modifiers, the preference for main predicates that express temporary properties is explained as a side-effect of “synchronizing” the main predicate with the locative frame in the course of finding an optimal interpretation. By emphasizing the division of labor between grammar and pragmatics, the proposed solution takes a considerable load off of semantics.
Research on dialectal varieties was for a long time concentrated on phonetic aspects of language. While there was a lot of work done on segmental aspects, suprasegmentals remained unexploited until the last few years, despite the fact that prosody was remarked as a salient aspect of dialectal variants by linguists and by naive speakers. Actual research on dialectal prosody in the German speaking area often deals with discourse analytic methods, correlating intonations curves with communicative functions (P. Auer et al. 2000, P. Gilles & R. Schrambke 2000, R. Kehrein & S. Rabanus 2001). The project I present here has another focus. It looks at general prosodic aspects, abstracted from actual situations. These global structures are modelled and integrated in a speech synthesis system. Today, mostly intonation is being investigated. However, rhythm, the temporal organisation of speech, is not a core of actual research on prosody. But there is evidence that temporal organisation is one of the main structuring elements of speech (B. Zellner 1998, B. Zellner Keller 2002). Following this approach developed for speech synthesis, I will present the modelling of the timing of two Swiss German dialects (Bernese and Zurich dialect) that are considered quite different on the prosodic level. These models are part of the project on the "development of basic knowledge for research on Swiss German prosody by means of speech synthesis modelling" founded by the Swiss National Science Foundation.
Teaching information literacy: substance and process This presentation explores the concept of information literacy within the broader context of higher education. It argues that, certain assertions in the library literature notwithstanding, the concepts associated with information literacy are not new, but rather very closely resemble the qualities traditionally considered to characterize a well-educated person. The presentation also considers the extent to which the higher education system does indeed foster the attributes commonly associated with information literacy. The term information literacy has achieved the immediacy it currently enjoys within the library community with the advent of the so-called "information age" The information age is commonly touted in the literature, both popular and professional, as constituting nothing short of a revolution. Academic librarians and other educators have of course felt called upon to make their teaching reflect both the growing proliferation of information formats and the major transformations affecting the process of information seeking. Faced with so much novelty and uncertainty, it is no surprise that many have felt that these changes call for a revolution in teaching. It is within this context that the concept of information literacy has flourished. It is argued in this presentation, however, that by treating information literacy as an essentially new specialty that owes much of its importance to the plethora of electronic information, we risk obscuring some of the most fundamental and enduring educational values we should be imparting to our students. Much of the literature on information literacy assumes - rather than argues - that recent changes in the way we approach education are indications of progress. Indeed, much of the self-narrative that institutions produce (in bulletins, mission statements, web sites, etc.) endorses an approach to education that will result in lifelong learners who are critical consumers of information. After critically examining the degree to which such statements of educational approach reflect reality, this presentation concludes by considering the effects of certain changes in the culture of higher education. It considers particularly the transformation - at least in North America - of the traditional model of higher education as a public good to a market-driven business model. It poses the question of whether a change of this significance might in fact detract from, rather than promote, the development of information literate students.
Using faculty-librarian partnerships to ensure that students become information fluent in the 21st century In the 21st century educators in partnership with librarians must prepare students effectively for productive use of information especially in higher education. Students will need to graduate from universities with appropriate information and technology skills to enable them to become productive citizens in the workplace and in society. Technology is having a major impact on society; in economics e-business is moving to the forefront; in communication e-mail, the Internet and cellular telephones have reformed how people communicate; in the work environment computers and web utilizations are emphasized and in education virtual learning and teaching are becoming more important. These few examples indicate how the 21st century information environment requires future members of the workforce to be information fluent so they will have the ability to locate information efficiently, evaluate information for specific needs, organize information to address issues, apply information skillfully to solve problems, use information to communicate effectively, and use information responsibly to ensure a productive work environment. Individuals can achieve information fluency by acquiring cultural, visual, computer, technology, research and information management skills to enable them to think critically.
Information literacy is a mosaic of attitudes, understandings, capabilities and knowledge about which there are three myths. The first myth is that it is about the ability to use ICTs to access a wealth of information. The second is that students entering higher education are information literate because student centred, resource based, and ICT focused learning are now pervasive in secondary education. The third myth is that information literacy development can be addressed by library-centric generic approaches. This paper addresses those myths and emphasises the need for information literacy to be recognised as the critical whole of education and societal issue, fundamental to an information-enabled and better world. In formal education, information literacy can only be developed by infusion into curriculum design, pedagogies, and assessment.
Navigating information, facilitating knowledge: the library, the academy, and student learning
(2004)
Understanding the nature and complementarity of the phenomena of information and knowledge lend not only epistemological clarity to their relationship, but also reaffirms the place of the library in the academic mission of knowledge transfer, acquisition, interpretation, and creation. These in turn reassert the legitimacy of the academic library as necessary participant in the teaching enterprise of colleges and universities. Such legitimacy induces an obligation to teach, and that obligation needs to be explored and implemented with adequate vigor and reach. Librarians and the academy must, however, concede that the scope of the task calls for a solution that goes beyond shared responsibilities. Academic libraries should assume a full teaching function even as they continue their exploration and design of activities and programs aimed at reinforcing information literacy in the various disciplines on campus. All must concede that need for collaboration cannot provide grounds for questioning the desirability of autonomous teaching status for the academic library in information literacy education