Universitätspublikationen
Refine
Year of publication
- 2012 (158) (remove)
Document Type
- Conference Proceeding (158) (remove)
Language
- English (133)
- German (24)
- Multiple languages (1)
Has Fulltext
- yes (158)
Is part of the Bibliography
- no (158)
Keywords
- Democracy (8)
- Law (6)
- human rights (6)
- law (6)
- democracy (5)
- Human Rights (3)
- Internet (3)
- Verfassung (3)
- educational freedom (3)
- homeschooling (3)
Institute
- Rechtswissenschaft (110)
- Physik (14)
- Medizin (8)
- Informatik (6)
- Zentrum für Weiterbildung (5)
- Biochemie und Chemie (3)
- Frankfurt Institute for Advanced Studies (FIAS) (3)
- Erziehungswissenschaften (2)
- Geschichtswissenschaften (2)
- Kulturwissenschaften (2)
The role of experts grows in the present and that is, in part, justifiable: as complexity rises, the ones who deliberate feel the need of the help of those who have know-how in specific fields. The question that must be asked revolves around the type of expectations developed in modern societies regarding what experts can do. Though specialization is not a peculiarity of our time (the process can be observed since human beings became sedentary); it has presently gained specific characteristics. Two aspects of modern life are particularly significant on that matter: (i.) the fact that the economic system is based on excitation of new needs (and no longer on the demand for satisfaction of needs); (ii.) the growing pursuit for total administration of conflicts. These factors are constitutive of what Gadamer sees as a great threat to our civilization: the excessive emphasis given in our time to the human ability to adapt. A specific ability is demanded from individuals: the capability of making an apparatus functions properly. Less resistance and more adaptability is requested, and because of that, autonomous thought - that is, not determined by the function it has in a system – is devalued. The threat we currently face is that the abilities of a good technocrat become the only qualities demanded from those who are responsible for practical decisions (especially in politics and law). Teleological reason, that guides the activity of specialists (and requires know-how in a specific area and consists in choosing means to reach a previously established goal), should not substitute practical reason, as the former requires adaptability to experience (not to a plan that was previously established) and is grounded on solidarity. In order to discuss the limits of the activity of specialists, the paper looks back to phrónesis and the way ancient Greeks set boundaries - this exercise should help raising new questions revolving the matter.
In this article the author, in the context of the fiftieth anniversary of H.L.A. Hart’s “The Concept of Law”, reconsiders the moderate indeterminacy of law thesis, which derives from the open texture of language. For that purpose, he intends: first, to analyze Hart’s moderate indeterminacy thesis, i.e. determinacy in “easy cases” and indeterminacy in “hard cases”, which resembles Aristotle’s "doctrine of the mean"; second, to criticize his moderate indeterminacy thesis as failing to embody the virtues of a center in between the vices of the extremes, by insisting that the exercise of discretion required constitutes an “interstitial” legislation; and, third, to reorganize an argument for a truly “mean” position, which requires a form of weak interpretative discretion, instead of a strong legislative discretion.
In order to understand the impact of new technologies on the law through the science of law, it is essential to observe how Law researches are done. This paper pursues the following models of legal science: analytical (theory of formal rule); hermeneutics (interpretation theory) and empirical (decision theory) to appraise methodological procedures used in monograph researches in some Brazilian Law courses. This study was to detect which model of law science was used in the development of Law researches. The study was conducted, through Juris Doctors’ interviews. All of these respondents have written a monograph, which is a requirement to complete a Law course in Brazil. The main conclusions of this study were the following: 1) most of the monographs produced do not specify the methodology used for developing the work; 2) when the papers indicate the methodology used, the analytical model was prevalent. In these cases, the science of law appears as a systematization of rules for obtaining possible decisions. 3) Hermeneutic and empirical models were also used, but on a smaller scale. These researches revealed the inaccuracy of the methodological tools used to apprehend the reality. However, these strategies are significant to define the objects of study of law in the contemporary time. Answering the question about how Law researches are done in some Brazilian Law schools, this paper discusses the construction of classical models of science of law, which were taken as the theoretical framework of this work before the hypercomplex current problems.
It is widely believed that chiral symmetry is spontaneously broken at zero temperature in the strong coupling limit of staggered fermions, for any number of colors and flavors. Using Monte Carlo simulations, we show that this conventional wisdom, based on a mean-field analysis, is wrong. For sufficiently many fundamental flavors, chiral symmetry is restored via a bulk, first-order transition. This chirally symmetric phase appears to be analytically connected with the expected conformal window of manyflavor continuum QCD. We perform simulations in the chirally symmetric phase at zero quark mass for various system sizes L, and measure the torelon mass and the Dirac spectrum. We find that all observables scale with L, which is hence the only infrared length scale. Thus, the strong-coupling chirally restored phase appears as a convenient laboratory to study IR-conformality. Finally, we present a conjecture for the phase diagram of lattice QCD as a function of the bare coupling and the number of quark flavors.
In the intersection between law, science and technology lies the debate on the overcoming of the boundaries of the biological structure of the human being and its implications on the idea of human rights, on the concept of person and on the conception of equality – being the latter a fundamental tenet of a democracy.
Posthumanism assumes a biological inadequacy of the human body regarding the quantity, complexity and quality of information which it can muster. The same occurs with the needs of accuracy, speed or strength demanded by the contemporary environment. Under such perspective, the body is considered to be an inefficient structure, with a short lifespan, easy to break and hard to fix.
The body, always seen as the locus for the definition of human, emerges as the object of a commodification process that seeks to exonerate men from their burden - by declination towards a virtual existence, totally free and rational - or to enhance them with bionic devices or drugs.
This issue has already been the subject of attention by many scholars like Savulescu, Rodotà, Broston, Fukuyama and even Habermas.
Therefore, the aim of this paper is to seek, by criticism and revision of the positions on the foreseen problems of this process, an adequate theoretical approach on issues like the concept of person and its connection with the idea of human rights in order to promote the fundamental statement that all men are equal without disregard to the values of diversity and personal identity.
Tagungsübersicht im Auftrag der Veranstalter: Ordnungen des Sehens. Innovationsfelder der kunsthistorischen Niederlandeforschung. Internationale Konferenz des Arbeitskreises Niederländische Kunst- & Kulturgeschichte e.V. (ANKK) in Zusammenarbeit mit dem Kunsthistorischen Institut der Goethe-Universität, Frankfurt am Main und dem Städel Museum, Frankfurt am Main, 30.09. - 02.10.2011
In the debate on how the new information and communication technologies impact on democratic politics the role played by the digital architecture seems to be surprisingly underrated. In particular, while a lot of attention has been paid to the possibilities that new technologies open up to democratic theory, few works have attempted to look at how democracy may help in shaping technologies. By adopting as a starting point the approach known as ‘code as law’, the paper aims at two objectives: to re-affirm the importance of discussing normative principles to guide the process of code writing in order to reinvigorate the debate; to claim the importance of input reasons when deciding which principles should be chosen. After having remarked that code is relevant for establishing democratic norms, the paper briefly tackles with the main attempts by European scholars to deal with this issue. Then, a couple of practical examples of how code impacts on democratic rights are sketched out. In the last section of the paper a shift from an output-based approach to the legitimacy of code to an input-based is openly advocated: an inquiry into the legitimacy of code should focus on its production.
Die Hauptthese dieses Papers geht von dem Konzept der normativen Verfassung der Nachkriegzeit aus und setzt sich kritisch mit dem Konzept des 19. Jahrhunderts „Verfassungswandlung“ auseinander. Das Konzept des Verfassungswandels ist mit der Verfassungsdemokratie inkompatibel. Statt von einem Verfassungswandel zu sprechen, sollte man die Entwicklung des Sinns der Normen in der Zeit als dynamische Interpretation bezeichnen.
The aim of this paper is to explore the case of the Spanish ‘indignants’ movement of May 2011 as an example of the structural changes occurring in the public sphere after the emergence of a new type of social movement characterized by the widespread use of the ICTs. First I focus on the ideological dimension of discourse of the ‘indignants’ movement, so as to reconstruct the protesters’ self-image. They thought that ICTs were playing a prominent role in a wider trend towards a regeneration of democracy, but they were rather misguided because they lack an accurate description of what really happened. In the second part of this paper I will challenge some features of my case study, emphasizing three basic elements of a democratic public sphere. I aim to call into question the idea that a ‘truly’ democratic public may be hosted by the emergent communicative environment.
Einer der zahlreichen Vorträge von Gertrude Lübbe-Wolff beginnt folgendermaßen: "Herr Gerhardt hat mich eingeladen, etwas über die aktuelle Bedeutung von Hegels Rechtsphilosophie zu sagen. Nichts lieber als das. An der Aktualität der Hegelschen Rechtsphilosophie leide ich geradezu, und über das, woran man leidet, spricht man ja gern. Die Aktualität der Hegelschen Rechtsphilosophie zeigt sich mir darin, dass ich öfter an Hegel denken muss, als mir lieb ist. Ich muss so oft an ihn denken, weil in unserer öffentlichen Kultur das Hegelwidrige so präsent ist." ...
This paper describes the ongoing efforts of the authors to present ancient Greek and Roman numismatic data on the public internet, with an emphasis on efforts to integrate information from multiple sources using Linked Data and Semantic Web techniques. By way of very modern metaphor, it is useful to think of coins as intentionally created packages of 'named entities'. Each coin was struck by a particular authority, often at a known site, and coins often make reference to familiar concepts such as deities, historical events, or symbols that were widely recognized in the ancient world. The institutions represented among the authors have deployed search interfaces that allow users to take advantage of this aspect of numismatic databases. The American Numismatic Society's database provides faceted search to its collection of over 550,000 objects. The Portable Antiquities Scheme (PAS) in the UK presents individual finds (and hoards) recorded throughout the country. The Römisch-Germanische Kommission and the University of Frankfurt (DBIS) are developing a prototype metaportal (INTERFACE) that accesses national databases of coin finds held in in Frankfurt, Vienna and Utrecht. Each of these resources is beginning to explore Semantic Web/Linked data approaches so that the role of numismatic standards is immediately coming to the fore. DBIS and INTERFACE are developing a numismatic ontology. At the ANS and PAS, the public database already presents RDF serializations based on Dublin Core. Together, the authors have begun to explore standardization of conceptual names on the basis of the vocabulary presented at the site http://nomisma.org . Nomisma.org is a collaborative effort to provide stable digital representations of numismatic concepts and entities. It provides URIs for such basic concepts as 'coin', 'mint', 'axis'. All of these are defined within the scope of numismatics but are already being linked to other stable resources where available. This is particularly the case for mints. For example, the URI http://nomisma.org/id/corinth is intended to represent that ancient city in its role as a minter/issuer of coins. The URI is linked via the SKOS ontology to the Pleiades Gazetteer of ancient places. This allows Nomisma to be the basis for a common representation of the concept that an object is a coin minted at Corinth. The ANS has already deployed such relationships in its public database. The work of all these projects is very much in progress so that this paper hopes to generate discussion on how multiple large projects can move forward in their own work while encouraging sufficient commonality to support large scale research questions undertaken by diverse audiences.
Abstract of the German original article “Rechtssubjekte und Teilrechtssubjekte des elektronischen Geschäftsverkehrs“, to be published in S. Beck (ed.): Jenseits von Mensch und Maschine: Moralische und rechtliche Aspekte des Umgangs mit Robotern, Künstlicher Intelligenz und Cyborgs. Baden-Baden: Nomos, 2012.
Doctrines developed by the EFTA Court have placed considerable demands on national courts in the EFTA States. The Court now considers the EEA Agreement to form an “international treaty sui generis which contains a distinct legal order of its own.” It would thus seem that EEA law has transformed into an independent legal order, and subsequently has a claim to validity which emulates the self-legitimising presentation of the EU legal order. This, however, is not an empirically verifiable fact, but a particular understanding which arises when one adopts the viewpoint of the EFTA Court. EEA law takes place in a different realm when interpreted and applied in the national order: this realm is essentially a construction of the constitutional order. Case law shows that the Icelandic Supreme Court is far from accepting all EEA judge-made principles. This study will describe a context of legal pluralism by reference to the Icelandic legal system and its relationship with the EEA legal order. To illustrate the discussion, the most important case law relative to the interaction between Icelandic laws and EEA law will be considered in the light of legal pluralism - particularly the principles of contrapunctual law designed by Miguel Maduro. The paper argues that the Supreme Court’s internal domestic approach to the application of EEA law will inevitably become a source of fragmentation unless it takes place within an institutional framework of judicial tolerance and judicial dialogue.
This paper considers the logic FOcard, i.e., first-order logic with cardinality predicates that can specify the size of a structure modulo some number. We study the expressive power of FOcard on the class of languages of ranked, finite, labelled trees with successor relations. Our first main result characterises the class of FOcard-definable tree languages in terms of algebraic closure properties of the tree languages. As it can be effectively checked whether the language of a given tree automaton satisfies these closure properties, we obtain a decidable characterisation of the class of regular tree languages definable in FOcard. Our second main result considers first-order logic with unary relations, successor relations, and two additional designated symbols < and + that must be interpreted as a linear order and its associated addition. Such a formula is called addition-invariant if, for each fixed interpretation of the unary relations and successor relations, its result is independent of the particular interpretation of < and +. We show that the FOcard-definable tree languages are exactly the regular tree languages definable in addition-invariant first-order logic. Our proof techniques involve tools from algebraic automata theory, reasoning with locality arguments, and the use of logical interpretations. We combine and extend methods developed by Benedikt and Segoufin (ACM ToCL, 2009) and Schweikardt and Segoufin (LICS, 2010).
Meeting abstract : Deutscher Kongress für Orthopädie und Unfallchirurgie (DKOU 2012), 23.10.-26.10.2012, Berlin.
Fragestellung: Die Behandlung langstreckiger Knochendefekte ist eine große Herausforderung. Die Masquelet-Technik zur Behandlung solcher Defekte ist eine zweizeitige Operationstechnik. Zuerst erfolgt die Insertion eines PMMA (Polymethylmethacrylat)-Zementspacers in den Knochendefekt, der die Bildung einer Membran induziert. Diese Membran enthält Wachstumsfaktoren und regenerative Zellen, die möglicherweise die Knochenheilung unterstützen. Nach einigen Wochen wird der Zementspacer entfernt und der induzierte Membranschlauch mit Beckenkammspongiosa aufgefüllt. Im weiteren Verlauf kommt es zu einer kompletten Knochenheilung. Ziele dieser Untersuchung waren die Etablierung der Masquelettechnik am Rattenmodell und die Definition eines Zeitpunkts, an welchem die Membran eine ausreichende Festigkeit sowie einen signifikanten Gehalt von Vorläuferzellen aufweist.
Methodik: Nach Genehmigung der Experimente wurden die Femura von 24 männlichen SD-Ratten osteotomiert. Die Lücke (10 mm) wurde mit PMMA-Zement aufgefüllt und mittels Miniplatte stabilisiert. Parallel wurden den Versuchstieren gleich große PMMA-Spacer subcutan unter die Rückenhaut implantiert. Nach 2, 4, bzw. 6 Wochen (W) erfolgte die Entnahme der Membranen. Ein Teil der Membran wurde für (immun)histologische Untersuchungen aufbereitet (CD34, vWF, van Giesson), ein Teil für die in vitro Kultur. Auswachsende Vorläuferzellen in vitro wurden über CD34 und STRO-1-Färbung nachgewiesen. Statistik: Mediane, Kruskal-Wallis-Test, p<0,05 ist signifikant.
Ergebnisse und Schlussfolgerungen: Im zeitlichen Verlauf nahmen die Vaskularisierung (vWF-positive Fläche [%]: 2 W: 1,8; 4 W:1.6 vs 6 W: 6,4), die Dicke der Membran ([µm]: 2 W: 350 vs 4W: 517, 6 W: 592) und der Bindegewebsanteil ([µm]: 2W: 201 vs 4W: 324, 6W: 404) signifikant zu. Der Hauptanteil elastischer Fasern war auf der dem Zement zugewandten Seite, Vaskularisierung war eher auf der Weichteil zugewandten Seite zu finden. Der Anteil CD34 positiver Zellen nahm signifikant ab (2W: 5%, 4 W: 4% vs 6 W: 1%). Auswachsende STRO-1 positive Zellen konnten nur in zweiwöchigen Membranen nachgewiesen werden. Ältere Membranen wiesen einen zunehmenden Anteil seneszenter Zellen auf. Subcutan induzierte Membranen waren vergleichbar, wiesen jedoch tendentiell eine geringere Dicke und keine STRO-1 positiven Zellen auf.
Mit dieser Studie wurde erstmalig die Induktion einer Membran nach Masquelet im Rattenmodell etabliert. Es konnte nachgewiesen werden, dass der strukturelle Aufbau sowie die zellulären Komponenten zeitlichen Änderungen unterliegen und der Ort der Induktion Einfluss auf die zellulären Komponenten der Membran hat. Junge Membranen (2W) enthielten CD34 und STRO-1 positive Zellen. 4W-Membranen enthielten nur CD34 positive Zellen wiesen aber einen signifikanten Bindegewebsanteil auf, der für eine erhöhte mechanische Stabilität spricht. Ob 2 bzw. 4 Wochen alte Membranen den Knochenheilungsprozess fördern, muss in weiterführenden Studien untersucht werden.
Der religiöse Pluralismus innerhalb der multikulturellen Gesellschaft erfordert vom Staat das Bemühen, die wechselseitige Achtung nicht nur zwischen Personen mit unterschiedlichen religiösen Glaubensüberzeugungen, sondern auch zwischen Glaubenden und Nicht-Glaubenden sicherzustellen. In diesem Kontext wird es für die vom Staat übernommene Funktion entscheidend sein, rechtzeitig zu beurteilen, ob er eine aktive und positive Rolle als eine Institution spielt, welche dafür sorgt, dass die Religionsfreiheit der Einzelnen und der Gruppen geachtet wird. Im Vorliegenden werden einige Gefahren und Bedrohungen für die Religionsfreiheit in der heutigen Gesellschaft analysiert und eine kritische Betrachtung als Antwort auf diese Krisensituation vorgelegt. Konkret werden die folgenden Punkte erörtert: 1. Der Glaube, daß die Religion nicht mit den Werten einer modernen, liberalen Gesellschaft zu vereinbaren ist. 2. Die Konfessionalisierung des Staates. 3. Der Missbrauch der staatlichen Macht, um die Präferenzen der Bürger durch absichtliche Änderungen im sozialen Kontext zu beeinflussen. 4. Die unnötigen Behinderungen und Beschränkung des Rechts auf Verweigerung aus Gewissengründen, die von einem falschen Verständnisses desselben herrühren. 5. Die Verherrlichung eines falsch verstandenen, radikalen Individualismus. 6. Ein Verständnis des Grundsatzes der Nichtdiskriminierung als ,,Gleichgültigkeit gegenüber den Unterschieden“.
The use of most if not all technologies is accompanied by negative side effects, While we may profit from today’s technologies, it is most often future generations who bear most risks. Risk analysis therefore becomes a delicate issue, because future risks often cannot be assigned a meaningful occurance probability. This paper argues that technology assessement most often deal with uncertainty and ignorance rather than risk when we include future generations into our ethical, political or juridal thinking. This has serious implications as probabilistic decision approaches are not applicable anymore. I contend that a virtue ethical approach in which dianoetic virtues play a central role may supplement a welfare based ethics in order to overcome the difficulties in dealing with uncertainty and ignorance in technology assessement.
The normative position of the judiciary under the traditional conception of democracy as self-legislation by the people is too weak to protect in an effective way the rights of suspects in the global War on Terror. Drawing on arguments elaborated by Hans Kelsen and Karl Popper, we shall attempt to devise in this paper an alternative democracy conception that could serve as a much more solid foundation for the judicial branch of government in a democratic state. Through this jurisprudential strategy, we hope to be able to maintain the balance of normative power among the Trias Politica, which, in turn, may contribute to the preservation of the legal rights of every person during the struggle against terrorists.
When judges are authorised to invalidate legal acts for being unconstitutional, the competence of the legislator is directly concerned. The question raises, if thus judges do not usurp legislative power. In the traditional doctrine of the separation of powers the parliament is the first power, based on its direct democratic legitimacy. Yet cancelling legal acts completely or partially does evoke more irritations in the public that could be expected. The people seem to have more confidence to the assumed impartiality of the judges than to the results of the parliamentary work which seems to be dominated by the struggles of the parties. The necessity of judicial review mainly is based on the consideration that individual rights even in an authentic democratic system may be violated by a legal act of the parliament. In this case constitutional courts have the very task to defend individual rights, principles of liberty and authentic equality. Therefore it is justified to speak of the “jurisdiction of liberty”, as the Italian constitutional expert Cappelletti has said. But also without such legitimacy in many countries the Courts intervene in the field of the legislator. The courts themselves discuss the limits of judicial interventions, emphasising themselves, that they have to respect the legislative decisions principally, but do not abide always by their own proclaimed principles. In Spanish recent publications it is spoken of the principle “in dubio pro legislatore”, (in case of doubt in favour the legislator), reminding of “in dubio pro reo”, in order to treat the legislative power not worse than the defendant in a criminal process..
One-photon and multi-photon absorption, spontaneous and stimulated photon emission, resonance Raman scattering and electron transfer are important molecular processes that commonly involve combined vibrational-electronic (vibronic) transitions. The corresponding vibronic transition profiles in the energy domain are usually determined by Franck-Condon factors (FCFs), the squared norm of overlap integrals between vibrational wavefunctions of different electronic states. FC profiles are typically highly congested for large molecular systems and the spectra usually become not well-resolvable at elevated temperatures. The (theoretical) analyses of such spectra are even more difficult when vibrational mode mixing (Duschinsky) effects are significant, because contributions from different modes are in general not separable, even within the harmonic approximation. A few decades ago Doktorov, Malkin and Man'ko [1979 J. Mol. Spectrosc. 77, 178] developed a coherent state-based generating function approach and exploited the dynamical symmetry of vibrational Hamiltonians for the Duschinsky relation to describe FC transitions at zero Kelvin. Recently, the present authors extended the method to incorporate thermal, single vibronic level, non-Condon and multi-photon effects in energy, time and probability density domains for the efficient calculation and interpretation of vibronic spectra. Herein, recent developments and corresponding generating functions are presented for single vibronic levels related to fluorescence, resonance Raman scattering and anharmonic transition.