Universitätspublikationen
Filtern
Erscheinungsjahr
Dokumenttyp
- Wissenschaftlicher Artikel (10808)
- Dissertation (1567)
- Preprint (1554)
- Arbeitspapier (1438)
- Ausgabe (Heft) zu einer Zeitschrift (564)
- Konferenzveröffentlichung (511)
- Bericht (299)
- Teil eines Buches (Kapitel) (107)
- Rezension (92)
- Buch (Monographie) (60)
Sprache
- Englisch (17089) (entfernen)
Volltext vorhanden
- ja (17089) (entfernen)
Schlagworte
- inflammation (92)
- COVID-19 (89)
- SARS-CoV-2 (62)
- Financial Institutions (47)
- Germany (45)
- climate change (45)
- aging (43)
- ECB (42)
- cancer (42)
- apoptosis (41)
Institut
- Medizin (5096)
- Physik (2985)
- Wirtschaftswissenschaften (1643)
- Frankfurt Institute for Advanced Studies (FIAS) (1575)
- Biowissenschaften (1399)
- Informatik (1249)
- Center for Financial Studies (CFS) (1136)
- Sustainable Architecture for Finance in Europe (SAFE) (1059)
- Biochemie und Chemie (855)
- House of Finance (HoF) (700)
Previous research on working memory (WM) in children with poor mathematical skills has yielded heterogeneous results, possibly due to inconsistent consideration of the IQ-achievement discrepancy and additional reading and spelling difficulties. To examine the impact of both, the WM of 68 average-achieving and 68 low-achieving third-graders in mathematics was assessed. Preliminary analyses showed that poor mathematical skills were associated with poor WM. Afterwards, children with isolated mathematical difficulties were separated from those with additional reading and spelling difficulties. Half of each group fulfilled the IQ-achievement discrepancy, resulting in a 2 (additional reading and spelling difficulties: yes/no) by 2 (IQ-achievement discrepancy: yes/no) factorial design. Analyses revealed that not fulfilling the IQ achievement discrepancy was associated with poor visual WM, whereas additional reading and spelling difficulties were associated with poor central executive functioning in children fulfilling the IQ-achievement discrepancy. Therefore, WM in children with poor mathematical skills differs according to the IQ-achievement discrepancy and additional reading and spelling difficulties.
BIOfid is a specialized information service currently being developed to mobilize biodiversity data dormant in printed historical and modern literature and to offer a platform for open access journals on the science of biodiversity. Our team of librarians, computer scientists and biologists produce high-quality text digitizations, develop new text-mining tools and generate detailed ontologies enabling semantic text analysis and semantic search by means of user-specific queries. In a pilot project we focus on German publications on the distribution and ecology of vascular plants, birds, moths and butterflies extending back to the Linnaeus period about 250 years ago. The three organism groups have been selected according to current demands of the relevant research community in Germany. The text corpus defined for this purpose comprises over 400 volumes with more than 100,000 pages to be digitized and will be complemented by journals from other digitization projects, copyright-free and project-related literature. With TextImager (Natural Language Processing & Text Visualization) and TextAnnotator (Discourse Semantic Annotation) we have already extended and launched tools that focus on the text-analytical section of our project. Furthermore, taxonomic and anatomical ontologies elaborated by us for the taxa prioritized by the project’s target group - German institutions and scientists active in biodiversity research - are constantly improved and expanded to maximize scientific data output. Our poster describes the general workflow of our project ranging from literature acquisition via software development, to data availability on the BIOfid web portal (http://biofid.de/), and the implementation into existing platforms which serve to promote global accessibility of biodiversity data.
Background: Only few studies deal with the workload of physical therapists and the health consequences, although this occupational group is quite important for the health care system in many industrialized countries (e.g. ca. 136 000 people are currently employed as physical therapists in Germany). Therefore, the current state of knowledge of work-related diseases and disorders of physical therapists is insufficient. The aim of the "Physical Therapist Cohort" (PTC) study is to analyze the association between work-related exposures and diseases among physical therapists in Germany. This article describes the protocol of the baseline assessment of the PTC study.
Methods/Design: A cross-sectional study will be conducted as baseline assessment and will include a representative random sample of approximately 300 physical therapists employed in Germany (exposure group), and a population-based comparison group (n = 300). The comparison group will comprise a sample of working aged (18–65 years) inhabitants of a German city. Variables of interest will be assessed using a questionnaire manual including questions regarding musculoskeletal, dermal, and infectious diseases and disorders as well as psychosocial exposures, diseases and disorders. In addition to subjective measures, a clinical examination will be used to objectify the questionnaire-based results (n = 50).
Discussion: The study, which includes extensive data collection, provides a unique opportunity to study the prospective association of work-related exposures and associated complaints of physical therapists. Baseline results will give first clues with regard to whether and how prevalent main exposures of physiotherapeutic work and typical work areas of physical therapists are associated with the development of work-related diseases. Thereby, this baseline assessment provides the basis for further investigations to examine causal relationships in accordance with a longitudinal design.
Background: The aim of this study was to investigate the acceptance and assessment of work shadowing carried out by students and dentists in dental practices. Furthermore, the extent to which students perceive an improvement in their specialised, communication and social competencies, was to be examined.
Methods: 61 dental students in their clinical semesters at a German university participated in work shadowing placements at 27 different general dental practices. Before beginning, they received checklists of various competencies that they self-assessed using school grades (from 1 = "very good", to 6 = "failed"), which they also repeated after completion. The dentists supplemented this with their external assessments. In addition, the students were requested to fill out a 54-item questionnaire and compose a freely-structured report after the work shadowing; the dentists filled out a questionnaire containing 16 items. The statistical analysis was carried out by means of the Friedman Test, including a post-hoc test (Bonferroni-Holm correction).
Results: The analysis showed a significant overall improvement in the students’ self-assessed competencies by 0.71* ± 0.43 grades. With an average of 0.33* ± 0.36, the dentists’ external assessment proved significantly higher than the self-assessment. The greatest improvements were perceived by the students in the areas of accounting (1.17* ± 0.77), practice organisation (1.05* ± 0.61) and dentist’s discussions (0.94* ±0.80) [*p < 0.05]. The students confirmed experiencing an expansion of knowledge, an improvement in their communication skills and indicated a high degree of satisfaction in regard to the dentists (school grade 1.58 ± 0.93). A maximum amount of satisfaction towards the work shadow students was demonstrated by the dentists, and this form of teaching was assessed with a school grade of 1.69 ± 0.89.
Conclusion: Both students and dental practitioners demonstrated a high level of satisfaction in regard to the work shadowing. The students felt their knowledge had increased, viewed the dentists as motivating role models and acknowledged a significant improvement in their specialised, communication and social competencies. Work shadowing in dental teaching practices presents a sensible addition to academic teaching at a university.
The information and communication technology (ICT) sector within the Netherlands is a major driver of globalization, the country’s economic growth and innovation. The Dutch ICT sector’s performance is increasingly becoming dependent upon employee driven innovations in order to address the needs of the sectors they service. In other words, the ICT sector within the Netherlands is largely dependent upon the performance and innovative capacity of its employees; both of which are functions of employee engagement. Given the high demand, and low supply of talent within this sector, ICT organizations need to develop innovative ways to enhance the performance capacities of its people. Developing an engaged and highly innovative workforce seems to be an efficient way to activate employees’ performance. As such, the aim of this paper was to investigate the mediating function of employee driven innovative work behaviors in the relationship between work engagement and task performance within the a Dutch ICT consulting firm. A cross-sectional survey-based research design, employing a census-based sampling method, was employed to obtain data from a global ICT consulting firm within the Netherlands (n = 232). The Utrecht Work Engagement Scale, the Innovative Work Behavior Scale and the Task Performance Scale was used to assess the associative subjective experiences of ICT employees. The results showed that work engagement is a significant driver for innovative work behaviors, which in turn affects the task performance of employees. Further, innovative work behaviors are therefore important to translate the engaging energies of employees into performance. This paper discusses the theoretical and practical implications of these findings.
We test two hypotheses, based on sexual selection theory, about gender differences in costly social interactions. Differential selectivity states that women invest less than men in interactions with new individuals. Differential opportunism states that women’s investment in social interactions is less responsive to information about the interaction’s payoffs. The hypotheses imply that women’s social networks are more stable and path dependent and composed of a greater proportion of strong relative to weak links. During their introductory week, we let new university students play an experimental trust game, first with one anonymous partner, then with the same and a new partner. Consistent with our hypotheses, we find that women invest less than men in new partners and that their investments are only half as responsive to information about the likely returns to the investment. Moreover, subsequent formation of students’ real social networks is consistent with the experimental results: being randomly assigned to the same introductory group has a much larger positive effect on women’s likelihood of reporting a subsequent friendship.
Background: Scientifically evaluated cognitive intervention programs are essential to meet the demands of our increasingly aging society. Currently, one of the “hottest” topics in the field is the improvement of working memory function and its potential impact on overall cognition. The present study evaluated the efficacy of WOME (WOrking MEmory), a theory-based working memory training program, in a double-blind, placebo-controlled, and randomized controlled trial (www.drks.de, DRKS00013162).
Methods: N = 60 healthy older adults were allocated to (1) the WOME intervention, (2) an active low-level intervention, or (3) a passive control group. Overall, the intervention groups practiced twelve sessions of 45 min within 4 weeks of their respective training. Transfer effects were measured via an extensive battery of neuropsychological tests and questionnaires both pre-/post-training and at a 3-month follow-up.
Results: WOME led to a significant improvement in working memory function, demonstrated on a non-trained near transfer task and on two different composite scores with moderate to large effect sizes. In addition, we found some indication of relevant impact on everyday life. The effects were short-term rather than stable, being substantially diminished at follow-up with only little evidence suggesting long-term maintenance. No transfer effects on other cognitive functions were observed.
Conclusion: WOME is an appropriate and efficient intervention specifically targeting the working memory system in healthy older adults.
Trial Registration: German Clinical Trials Register (DRKS), Identifier: DRKS00013162.
Reproducing the characteristics and the functional responses of the blood–brain barrier (BBB) in vitro represents an important task for the research community, and would be a critical biotechnological breakthrough. Pharmaceutical and biotechnology industries provide strong demand for inexpensive and easy-to-handle in vitro BBB models to screen novel drug candidates. Recently, it was shown that canonical Wnt signaling is responsible for the induction of the BBB properties in the neonatal brain microvasculature in vivo. In the present study, following on from earlier observations, we have developed a novel model of the BBB in vitro that may be suitable for large scale screening assays. This model is based on immortalized endothelial cell lines derived from murine and human brain, with no need for co-culture with astrocytes. To maintain the BBB endothelial cell properties, the cell lines are cultured in the presence of Wnt3a or drugs that stabilize β-catenin, or they are infected with a transcriptionally active form of β-catenin. Upon these treatments, the cell lines maintain expression of BBB-specific markers, which results in elevated transendothelial electrical resistance and reduced cell permeability. Importantly, these properties are retained for several passages in culture, and they can be reproduced and maintained in different laboratories over time. We conclude that the brain-derived endothelial cell lines that we have investigated gain their specialized characteristics upon activation of the canonical Wnt pathway. This model may be thus suitable to test the BBB permeability to chemicals or large molecular weight proteins, transmigration of inflammatory cells, treatments with cytokines, and genetic manipulation.
This article aims to shed light on some core challenges of liberating social criticism. Its centerpiece is an intuitively attractive account of the nature and difficulty of critical social thought that nevertheless goes missing in many philosophical conversations about critique. This omission at bottom reflects the fact that the account presupposes a philosophically contentious conception of rationality. Yet the relevant conception of rationality does in fact inform influential philosophical treatments of social criticism, including, very prominently, a left Hegelian strand of thinking within contemporary Critical Theory. Moreover, it is possible to mount a defense of the conception by reconstructing, if with various qualifications and additions, an argument from classic—i.e., mid twentieth-century—Anglo-American philosophy of the social sciences, in particular, the argument that forms the backbone of Peter Winch’s The Idea of a Social Science. Winch draws his guiding insights from the later philosophy of Wittgenstein, and one of the payoffs of considering Winch’s Wittgenstein-inspired work against the backdrop of Hegel-inspired work in Critical Theory is to contest the artificial professional strictures that are sometimes taken to speak against reaching across the so-called ‘Continental Divide’ in philosophy. The larger payoff is advancing, by means of this philosophically ecumenical approach, the enterprise of liberating social thought.
Background: Using data from the COHERE collaboration, we investigated whether primary prophylaxis for pneumocystis pneumonia (PcP) might be withheld in all patients on antiretroviral therapy (ART) with suppressed plasma human immunodeficiency virus (HIV) RNA (≤400 copies/mL), irrespective of CD4 count.
Methods: We implemented an established causal inference approach whereby observational data are used to emulate a randomized trial. Patients taking PcP prophylaxis were eligible for the emulated trial if their CD4 count was ≤200 cells/µL in line with existing recommendations. We compared the following 2 strategies for stopping prophylaxis: (1) when CD4 count was >200 cells/µL for >3 months or (2) when the patient was virologically suppressed (2 consecutive HIV RNA ≤400 copies/mL). Patients were artificially censored if they did not comply with these stopping rules. We estimated the risk of primary PcP in patients on ART, using the hazard ratio (HR) to compare the stopping strategies by fitting a pooled logistic model, including inverse probability weights to adjust for the selection bias introduced by the artificial censoring.
Results: A total of 4813 patients (10 324 person-years) complied with eligibility conditions for the emulated trial. With primary PcP diagnosis as an endpoint, the adjusted HR (aHR) indicated a slightly lower, but not statistically significant, different risk for the strategy based on viral suppression alone compared with the existing guidelines (aHR, .8; 95% confidence interval, .6–1.1; P = .2).
Conclusions: This study suggests that primary PcP prophylaxis might be safely withheld in confirmed virologically suppressed patients on ART, regardless of their CD4 count.
In 1905, the managing editor of the Jewish Encyclopedia, Isidore Singer (1859–1939), published an article in the journal Ost und West from a "bird’s eye perspective on the development of American Jewry in the last 250 years." In this historical overview, Singer eventually attested that Jewish scholarship in America had an "absolute dependency on the European motherland." This judgment was based on his disapproving view of the two American rabbinical seminaries that existed at that time. According to Singer, there were still no scholars at the Hebrew Union College (HUC) in Cincinnati of the "already American[-born] generation of Israel." In fact, Singer’s observation was appropriate because it applied to the Jewish Theological Seminary of America (JTSA) in New York as much as to the HUC.3 Despite the history of Jewish settlement in America, around 1900 there was still no native Jewish scholarship in America. The scene was dominated by scholars educated in Europe, who often came with broken English and a strict academic sense of mission. In 1903, Kaufmann Kohler (1843–1926), born in Bavaria and trained at German universities, was chosen as the president of HUC. And a year earlier, Solomon Schechter (1847–1915) had been called to the JTSA in New York as its new president. ...
The influence and power of some OECD states is under threat but China appears to remain astonishingly flexible, economically potent, and politically strong. How accurate is this view? To answer this question, major aspects of Chinese economic regulation that were adopted in the country’s progress towards capitalist modernization are examined. The analysis requires a historical reconstruction of how China changed the way it intervenes economically and politically, especially with regard to the institutions of the central state. Such a reconstruction reveals that, since the 1990s, the central state has indeed increased its steering capacities. These capacities have a distinctive basis that includes acceptance of a state-centered approach, idiosyncratic innovation policies taking place in the "shadow" of the state’s hierarchy, and the ongoing influence of the communist party. An all-embracing controlling power is, however, not detectable. What does exist in China’s competition-driven system of “statecapitalist” regulation, is a set of limits on the state’s capacity to govern.
This article provides an overview and critical assessment of WIPO ALERT. It locates this initiative in the broader context of transnational IP enforcement schemes on the Internet. These initiatives are classified into two categories according to their point of attachment and geographical effect. Whereas source-related measures (e.g. website takedowns) tend to have a transnational and possibly even a global effect, recipient-related measures (e.g. website and ad blockings) typically mirror the territorially fragmented IPR landscape. This fragmentation is where WIPO ALERT comes into play. It can be understood as a matching service which interconnects holders of information about copyright infringing websites (“Authorized Contributors”) and actors of the online ad industry who want to avoid these outlets (“Authorized Users”). The critical assessment of WIPO ALERT calls for more transparency and the establishment of uniform substantive and procedural standards that have to be met if a new “site of concern” is added to the global ad blacklist.
Background: Correct species identification of blow flies is a crucial step for understanding their biology, which can be used not only for designing fly control programs, but also to determine the minimum time since death. Identification techniques are usually based on morphological and molecular characters. However, the use of classical morphology requires experienced entomologists for correct identification; while molecular techniques rely on a sound laboratory expertise and remain ambiguous for certain taxa. Landmark-based geometric morphometric analysis of insect wings has been extensively applied in species identification. However, few wing morphometric analyses of blow fly species have been published.
Methods: We applied a landmark-based geometric morphometric analysis of wings for species identification of 12 medically and forensically important blow fly species of Thailand. Nineteen landmarks of each right wing of 372 specimens were digitised. Variation in wing size and wing shape was analysed and evaluated for allometric effects. The latter confirmed the influence of size on the shape differences between species and sexes. Wing shape variation among genera and species were analysed using canonical variates analysis followed by a cross-validation test.
Results: Wing size was not suitable for species discrimination, whereas wing shape can be a useful tool to separate taxa on both, genus and species level depending on the analysed taxa. It appeared to be highly reliable, especially for classifying Chrysomya species, but less robust for a species discrimination in the genera Lucilia and Hemipyrellia. Allometry did not affect species separation but had an impact on sexual shape dimorphism.
Conclusions: A landmark-based geometric morphometric analysis of wings is a useful additional method for species discrimination. It is a simple, reliable and inexpensive method, but it can be time-consuming locating the landmarks for a large scale study and requires non-damaged wings for analysis.
Ebola virus (EBOV) infection causes a high death toll, killing a high proportion of EBOV-infected patients within 7 days. Comprehensive data on EBOV infection are fragmented, hampering efforts in developing therapeutics and vaccines against EBOV. Under this circumstance, mathematical models become valuable resources to explore potential controlling strategies. In this paper, we employed experimental data of EBOV-infected nonhuman primates (NHPs) to construct a mathematical framework for determining windows of opportunity for treatment and vaccination. Considering a prophylactic vaccine based on recombinant vesicular stomatitis virus expressing the EBOV glycoprotein (rVSV-EBOV), vaccination could be protective if a subject is vaccinated during a period from one week to four months before infection. For the case of a therapeutic vaccine based on monoclonal antibodies (mAbs), a single dose might resolve the invasive EBOV replication even if it was administrated as late as four days after infection. Our mathematical models can be used as building blocks for evaluating therapeutic and vaccine modalities as well as for evaluating public health intervention strategies in outbreaks. Future laborator experiments will help to validate and refine the estimates of the windows of opportunity proposed here.
Chern numbers can be calculated within a frame of vortex fields related to phase conventions of a wave function. In a band protected by gaps the Chern number is equivalent to the total number of flux carrying vortices. In the presence of topological defects like Dirac cones this method becomes problematic, in particular if they lack a well-defined winding number. We develop a scheme to include topological defects into the vortex field frame. A winding number is determined by the behavior of the phase in reciprocal space when encircling the defect's contact point. To address the possible lack of a winding number we utilize a more general concept of winding vectors. We demonstrate the usefulness of this ansatz on Dirac cones generated from bands of the Hofstadter model.
Chern numbers can be calculated within a frame of vortex fields related to phase conventions of a wave function. In a band protected by gaps the Chern number is equivalent to the total number of flux carrying vortices. In the presence of topological defects like Dirac cones this method becomes problematic, in particular if they lack a well-defined winding number. We develop a scheme to include topological defects into the vortex field frame. A winding number is determined by the behavior of the phase in reciprocal space when encircling the defect's contact point. To address the possible lack of a winding number we utilize a more general concept of winding vectors. We demonstrate the usefulness of this ansatz on Dirac cones generated from bands of the Hofstadter model.
Chern numbers can be calculated within a frame of vortex fields related to phase conventions of a wave function. In a band protected by gaps the Chern number is equivalent to the total number of flux carrying vortices. In the presence of topological defects like Dirac cones this method becomes problematic, in particular if they lack a well-defined winding number. We develop a scheme to include topological defects into the vortex field frame. A winding number is determined by the behavior of the phase in reciprocal space when encircling the defect's contact point. To address the possible lack of a winding number we utilize a more general concept of winding vectors. We demonstrate the usefulness of this ansatz on Dirac cones generated from bands of the Hofstadter model.
One of the crucial steps during trials for Zika and other vaccines is to recruit participants and to understand how participants’ attitudes and sociodemographic characteristics affect willingness to participate (WTP). This study was conducted to assess WTP, its explanatory variables, and the impact of financial compensation on WTP in Indonesia. A health facility-based cross-sectional study was conducted in eleven regencies in the Aceh and West Sumatra provinces of Indonesia. Participants were recruited via a convenience sampling method and were interviewed. The associations between explanatory variables and WTP were assessed using a two-step logistic regression analysis. A total of 1,102 parents were approached, and of these 956 (86.8%) completed the interview and were included in analysis. Of those, 144 (15.1%) were willing to participate in a Zika vaccine trial without a financial compensation. In the multivariate analysis, WTP was tied to an age of more than 50 years old, compared to 20–29 years (odds ratio (OR): 5.0; 95% confidence interval (CI): 2.37–10.53), to being female (OR: 2.20; 95% CI: 1.11–4.37), and to having heard about Zika (OR: 2.41; 95% CI: 1.59–3.65). Participants’ WTP increased gradually with higher financial compensation. The rate of WTP increased to 62.3% at the highest offer (US$ 350.4), and those who were still unwilling to participate (37.7%) had a poorer attitude towards childhood vaccination. This study highlights that pre-existing knowledge about Zika and attitudes towards childhood vaccination are important in determining community members being willing to participate in a vaccine trial. Financial incentives are still an important factor to enhance participant recruitment during a vaccine trial.
This paper investigates whether exchanging the Social Security delayed retirement credit, currently paid as an increase in lifetime annuity benefits, for a lump sum would induce later claiming and additional work. We show that people would voluntarily claim about half a year later if the lump sum were paid for claiming any time after the Early Retirement Age, and about two-thirds of a year later if the lump sum were paid only for those claiming after their Full Retirement Age. Overall, people will work one-third to one-half of the additional months, compared to the status quo. Those who would currently claim at the youngest ages are likely to be most responsive to the offer of a lump sum benefit.
The paper presents a study which was based on the hypothesis that wikis that are initiated bottom up by students might be used more deliberately than wikis which are introduced top down by teachers. Therefore it examines the specific effects observed in nine different wiki projects at the university of Frankfurt ranging from student wiki projects up to wikis used in seminars and as information tool for institutions.
This paper provides a theoretical assessment of gestures in the context of authoring image-related hypertexts by example of the museum information system WikiNect. To this end, a first implementation of gestural writing based on image schemata is provided (Lakoff in Women, fire, and dangerous things: what categories reveal about the mind. University of Chicago Press, Chicago, 1987). Gestural writing is defined as a sort of coding in which propositions are only expressed by means of gestures. In this respect, it is shown that image schemata allow for bridging between natural language predicates and gestural manifestations. Further, it is demonstrated that gestural writing primarily focuses on the perceptual level of image descriptions (Hollink et al. in Int J Hum Comput Stud 61(5):601–626, 2004). By exploring the metaphorical potential of image schemata, it is finally illustrated how to extend the expressiveness of gestural writing in order to reach the conceptual level of image descriptions. In this context, the paper paves the way for implementing museum information systems like WikiNect as systems of kinetic hypertext authoring based on full-fledged gestural writing.
We study the Wigner function for massive spin-1/2 fermions in electromagnetic fields. The Wigner function is analytically solved in five cases when electromagnetic fields are constants. For a general space-time dependent field configuration, we use the method of semi-classical expansion and solved the Wigner function at linear order in the Planck's constant. At the same order, we obtained a generalized Boltzmann equation for particle distribution, and a generalized BMT equation for spin polarization. Using the Wigner function, we calculated some physical quantities in a thermal equilibrium system.
Several studies suggested that transcription factor (TF) binding to DNA may be impaired or enhanced by DNA methylation. We present MeDeMo, a toolbox for TF motif analysis that combines information about DNA methylation with models capturing intra-motif dependencies. In a large-scale study using ChIP-seq data for 335 TFs, we identify novel TFs that are affected by DNA methylation. Overall, we find that CpG methylation decreases the likelihood of binding for the majority of TFs. For a considerable subset of TFs, we show that intra-motif dependencies are pivotal for accurately modelling the impact of DNA methylation on TF binding.
Previous reports of improved oral reading performance for dyslexic children but not for regular readers when between-letter spacing was enlarged led to the proposal of a dyslexia-specific deficit in visual crowding. However, it is in this context also critical to understand how letter spacing affects visual word recognition and reading in unimpaired readers. Adopting an individual differences approach, the present study, accordingly, examined whether wider letter spacing improves reading performance also for non-impaired adults during silent reading and whether there is an association between letter spacing and crowding sensitivity. We report eye movement data of 24 German students who silently read texts presented either with normal or wider letter spacing. Foveal and parafoveal crowding sensitivity were estimated using two independent tests. Wider spacing reduced first fixation durations, gaze durations, and total fixation time for all participants, with slower readers showing stronger effects. However, wider letter spacing also reduced skipping probabilities and elicited more fixations, especially for faster readers. In terms of words read per minute, wider letter spacing did not provide a benefit, and faster readers in particular were slowed down. Neither foveal nor parafoveal crowding sensitivity correlated with the observed letter-spacing effects. In conclusion, wide letter spacing reduces single word processing time in typically developed readers during silent reading, but affects reading rates negatively since more words must be fixated. We tentatively propose that wider letter spacing reinforces serial letter processing in slower readers, but disrupts parallel processing of letter chunks in faster readers. These effects of letter spacing do not seem to be mediated by individual differences in crowding sensitivity.
wo assumptions underlie current models of the geographical ranges of perennial plant species: 1. current ranges are in equilibrium with the prevailing climate, and 2. changes are attributable to changes in macroclimatic factors, including tolerance of winter cold, the duration of the growing season, and water stress during the growing season, rather than to biotic interactions. These assumptions allow model parameters to be estimated from current species ranges. Deterioration of growing conditions due to climate change, e.g. more severe drought, will cause local extinction. However, for many plant species, the predicted climate change of higher minimum temperatures and longer growing seasons means, improved growing conditions. Biogeographical models may under some circumstances predict that a species will become locally extinct, despite improved growing conditions, because they are based on an assumption of equilibrium and this forces the species range to match the species-specific macroclimatic thresholds. We argue that such model predictions should be rejected unless there is evidence either that competition influences the position of the range margins or that a certain physiological mechanism associated with the apparent improvement in growing conditions negatively affects the species performance. We illustrate how a process-based vegetation model can be used to ascertain whether such a physiological cause exists. To avoid potential modelling errors of this type, we propose a method that constrains the scenario predictions of the envelope models by changing the geographical distribution of the dominant plant functional type. Consistent modelling results are very important for evaluating how changes in species areas affect local functional trait diversity and hence ecosystem functioning and resilience, and for inferring the implications for conservation management in the face of climate change.
Last November, the media organisation of the „Islamic State“ (IS) published a video, the sole purpose of which was to prove that the „caliphate“ which the IS has established in June 2014 was in fact a proper state. The video highlighted a host of institutions in order to drive home the claim of real statehood, including examples like a working judiciary, a prison administration, a schooling system, and so on. At one point in the video, the IS claimed that it was also financially independent and had apt resources at its disposal, namely oil and gas.
However, while it is true that the IS controls a number of oil and gas fields in Syria as well as in Iraq, we have by now enough evidence to be rather sure that the economic base of the „caliphate“ is by no means sustainable...
This policy letter provides evidence for the crucial importance of the initial regulatory treatment for the further development of financial innovations by exploring the emergence and initial legal framing of off-balance-sheet leasing in Germany. Due to a missing legal framework, lease contracts occurred as an innovative social practice of off-balance-sheet financing. However, this lacking legal framing impeded the development of this financial innovation as it also created legal uncertainties. This was about to change after the initial legal framing of leasing in the 1970’s which eliminated those legal uncertainties and off-balance-sheet leasing entered into a stunning period of growth while laying the foundation of a regulatory resiliency against efforts that seek to abandon the off-balance-sheet treatment of leases. As the initial legal framing is crucial for the further development of a financial innovation, we propose the French approach for the initial vindication of new financial products in which the principles-based rules are aligned with the capabilities of regulators to intervene, even when a financial innovation complies with the letter of the law. In this way, regulators could regulate the frontier of financial innovations and weed out those which are entirely or mainly driven by regulatory arbitrage considerations while maintaining the beneficial elements of those products.
canning tunneling microscopy (STM) is perhaps the most promising way to detect the superconducting gap size and structure in the canonical unconventional superconductor Sr2RuO4 directly. However, in many cases, researchers have reported being unable to detect the gap at all in simple STM conductance measurements. Recently, an investigation of this issue on various local topographic structures on a Sr-terminated surface found that superconducting spectra appeared only in the region of small nanoscale canyons, corresponding to the removal of one RuO surface layer. Here, we analyze the electronic structure of various possible surface structures using first principles methods, and argue that bulk conditions favorable for superconductivity can be achieved when removal of the RuO layer suppresses the RuO4 octahedral rotation locally. We further propose alternative terminations to the most frequently reported Sr termination where superconductivity surfaces should be observed.
A link between populism and social media is often suspected. This paper spells out a set of possible mechanisms underpinning this link: that social media changes the communication structure of the public sphere, making it harder for citizens to obtain evidence that refutes populist assumptions. By developing a model of the public sphere, four core functions of the public sphere are identified: exposing citizens to diverse information, promoting equality of deliberative opportunity, creating deliberative transparency, and producing common knowledge. A wellworking public sphere allows citizens to learn that there are genuine disagreements among citizens that are held in good faith. Social media makes it harder to gain this insight, opening the door for populist ideology.
Why MREL won’t help much
(2017)
The bail-in tool as implemented in the European bank resolution framework suffers from severe shortcomings. To some extent, the regulatory framework can remedy the impediments to the desirable incentive effect of private sector involvement (PSI) that emanate from a lack of predictability of outcomes, if it compels banks to issue a sufficiently sized minimum of high-quality, easy to bail-in (subordinated) liabilities. Yet, even the limited improvements any prescription of bail-in capital can offer for PSI’s operational effectiveness seem compromised in important respects.
The main problem, echoing the general concerns voiced against the European bail-in regime, is that the specifications for minimum requirements for own funds and eligible liabilities (MREL) are also highly detailed and discretionary and thus alleviate the predicament of investors in bail-in debt, at best, only insufficiently. Quite importantly, given the character of typical MREL instruments as non-runnable long-term debt, even if investors are able to gauge the relevant risk of PSI in a bank’s failure correctly at the time of purchase, subsequent adjustment of MREL-prescriptions by competent or resolution authorities potentially change the risk profile of the pertinent instruments. Therefore, original pricing decisions may prove inadequate and so may market discipline that follows from them.
The pending European legislation aims at an implementation of the already complex specifications of the Financial Stability Board (FSB) for Total Loss Absorbing Capacity (TLAC) by very detailed and case specific amendments to both the regulatory capital and the resolution regime with an exorbitant emphasis on proportionality and technical fine-tuning. What gets lost in this approach, however, is the key policy objective of enhanced market discipline through predictable PSI: it is hardly conceivable that the pricing of MREL-instruments reflects an accurate risk-assessment of investors because of the many discretionary choices a multitude of agencies are supposed to make and revisit in the administration of the new regime. To prove this conclusion, this chapter looks in more detail at the regulatory objectives of the BRRD’s prescriptions for MREL and their implementation in the prospectively amended European supervisory and resolution framework.
The bail-in tool as implemented in the European bank resolution framework suffers from severe shortcomings. To some extent, the regulatory framework can remedy the impediments to the desirable incentive effect of private sector involvement (PSI) that emanate from a lack of predictability of outcomes, if it compels banks to issue a sufficiently sized minimum of high-quality, easy to bail-in (subordinated) liabilities. Yet, even the limited improvements any prescription of bail-in capital can offer for PSI’s operational effectiveness seem compromised in important respects.
The main problem, echoing the general concerns voiced against the European bail-in regime, is that the specifications for minimum requirements for own funds and eligible liabilities (MREL) are also highly detailed and discretionary and thus alleviate the predicament of investors in bail-in debt, at best, only insufficiently. Quite importantly, given the character of typical MREL instruments as non-runnable long-term debt, even if investors are able to gauge the relevant risk of PSI in a bank’s failure correctly at the time of purchase, subsequent adjustment of MREL-prescriptions by competent or resolution authorities potentially change the risk profile of the pertinent instruments. Therefore, original pricing decisions may prove inadequate and so may market discipline that follows from them.
The pending European legislation aims at an implementation of the already complex specifications of the Financial Stability Board (FSB) for Total Loss Absorbing Capacity (TLAC) by very detailed and case specific amendments to both the regulatory capital and the resolution regime with an exorbitant emphasis on proportionality and technical fine-tuning. What gets lost in this approach, however, is the key policy objective of enhanced market discipline through predictable PSI: it is hardly conceivable that the pricing of MREL-instruments reflects an accurate risk-assessment of investors because of the many discretionary choices a multitude of agencies are supposed to make and revisit in the administration of the new regime. To prove this conclusion, this chapter looks in more detail at the regulatory objectives of the BRRD’s prescriptions for MREL and their implementation in the prospectively amended European supervisory and resolution framework.
The prefix cyber, prepended onto terms like war, peace, security, and so on, results in interesting word combinations which we construct with our spoken language. Many scholars, from political to social science, have discussed the terms and the semantics of it in order to understand the problem and to create some scientific value out of it. But this article will not be another endless discussion on whether cyberfoo exists somewhere in any computer network at the moment or not...
Myocardial injury as induced by myocardial infarction results in tissue ischemia, which critically incepts cardiomyocyte death. Endothelial cells play a crucial role in restoring oxygen and nutrient supply to the heart. Latest advances in single-cell multi-omics, together with genetic lineage tracing, reveal a transcriptional and phenotypical adaptation to the injured microenvironment, which includes alterations in metabolic, mesenchymal, hematopoietic and pro-inflammatory signatures. The extent of transition in mesenchymal or hematopoietic cell lineages is still debated, but it is clear that several of the adaptive phenotypical changes are transient and endothelial cells revert back to a naïve cell state after resolution of injury responses. This resilience of endothelial cells to acute stress responses is important for preventing chronic dysfunction. Here, we summarize how endothelial cells adjust to injury and how this dynamic response contributes to repair and regeneration. We will highlight intrinsic and microenvironmental factors that contribute to endothelial cell resilience and may be targetable to maintain a functionally active, healthy microcirculation.
The article presents a brief overview of research and publication in the history of international law in Europe today. The upsurge of interest in historical studies is traced back to a sense of present transformation, with historical studies seeking to explore both aspects of continuity and change in the international legal system. The article outlines three tasks for the discipline in the future: to begin work for international law’s Ideengeschichte, to focus on the relationship between the West and its "Other", and to undertake studies in the historical sociology of international law.
In this article, I question the use of the notion of ‘constituent power’ as a tool for the democratization of the European Union (EU). Rather than seeing the absence of a transnational constituent power as a cause of the EU’s ‘democratic deficit’, I identify it as an opportunity for unfettered democratic participation. Against the reification of power-in-action into a power-constituted-in-law, I argue that the democratization of the EU can only be achieved through the multiplication of ‘constituent moments’. I begin by deconstructing the normative justifications surrounding the concept of constituent power. Here I analyze the structural aporia of constituent power and question the autonomous and emancipatory dimension of this notion. I then test the theoretical hypothesis of this structural aporia of the popular constituent power by comparing it with the historical experiments of a European popular constituent power. Finally, based on these theoretical and empirical observations, I propose to replace the ambivalence of the concept of popular constituent power with a more cautious approach to the bottom-up democratization of European integration: that of a multiplication of transnational constituent moments.
Why does the schooling gap close while the wage gap persists across country income comparisons?
(2023)
The schooling gap diminishes because the services sector becomes more pronounced for high-income countries, and the paid hours gap closes. Although gender wage inequality persists across country income groups, differences in schooling years between females and males diminish. We assemble a novel dataset, calibrate a general equilibrium, multi-sector, -gender, and -production technology model, and show that gender-specific sectoral comparative advantages explain the paid hours and schooling gap decline from low- to high-income economies even when the wage gap persists. Additionally, our counterfactual analyses indicate that consumption subsistence and production share heterogeneity across both income groups and genders are essential to explain the co-decline of the schooling and paid hours gaps. Our results highlight effective mechanisms for policies aiming to reduce gender inequality in schooling and suggest that the schooling gap decline and the de-invisibilization of female paid work observed in high-income countries are linked by structural sector movements instead of wage inequality reductions.
From 1963 through 2015, idiosyncratic risk (IR) is high when market risk (MR) is high. We show that the positive relation between IR and MR is highly stable through time and is robust across exchanges, firm size, liquidity, and market-to-book groupings. Though stock liquidity affects the strength of the relation, the relation is strong for the most liquid stocks. The relation has roots in fundamentals as higher market risk predicts greater idiosyncratic earnings volatility and as firm characteristics related to the ability of firms to adjust to higher uncertainty help explain the strength of the relation. Consistent with the view that growth options provide a hedge against macroeconomic uncertainty, we find evidence that the relation is weaker for firms with more growth options.
Identifying the cause of discrimination is crucial to design effective policies and to understand discrimination dynamics. Building on traditional models, this paper introduces a new explanation for discrimination: discrimination based on motivated reasoning. By systematically acquiring and processing information, individuals form motivated beliefs and consequentially discriminate based on these beliefs. Through a series of experiments, I show the existence of discrimination based on motivated reasoning and demonstrate important differences to statistical discrimination and taste-based discrimination. Finally, I demonstrate how this form of discrimination can be alleviated by limiting individuals’ scope to interpret information.
There is mounting evidence that retail investors make predictable, costly investment mistakes, including underinvestment, naïve diversification, and payment of excessive fund fees. Over the past thirty-five years, however, participant-directed 401(k) plans have largely replaced professionally managed pension plans, requiring unsophisticated retail investors to navigate the financial markets themselves. Policy-makers have struggled with regulatory interventions designed to improve the quality of investment decisions without a clear understanding of the reasons for investor mistakes. Absent such an understanding, it is difficult to design effective regulatory responses. This article offers a first step in understanding the investor decision-making process. We use an internet-based experiment to disentangle possible explanations for inefficient investment decisions. The experiment employs a simplified construct of an employee’s allocation among the options in a retirement plan coupled with technology that enables us to collect data on the specific information that investors choose to view. In addition to collecting general information about the process by which investors choose among mutual fund options, we employ an experimental manipulation to test the effect of an instruction on the importance of mutual fund fees. Pairing this instruction with simplified fee disclosure allows us to distinguish between motivation-limits and cognition-limits as explanations for the widespread findings that investors ignore fees in their investment decisions. Our results offer partial but limited grounds for optimism. On the one hand, within our simplified experimental construct, our subjects allocated more money, on average, to higher-value funds. Furthermore, subjects who received the fees instruction paid closer attention to mutual fund fees and allocated their investments into funds with lower fees. On the other hand, the effects of even a blunt fees instruction were limited, and investors were unable to identify and avoid clearly inferior fund options. In addition, our results suggest that excessive, naïve diversification strategies are driving many investment decisions. Although our findings are preliminary, they suggest valuable avenues for future research and important implications for regulation of retail investing.
According to disposition effect theory, people hold losing investments too long. However, many investors eventually sell at a loss, and little is known about which psychological factors contribute to these capitulation decisions. This study integrates prospect theory, utility maximization theory, and theory on reference point adaptation to argue that the combination of a negative expectation about an investment’s future performance and a low level of adaptation to previous losses leads to a greater capitulation probability. The test of this hypothesis in a dynamic experimental setting reveals that a larger total loss and longer time spent in a losing position lead to downward adaptations of the reference point. Negative expectations about future investment performance lead to a greater capitulation probability. Consistent with the theoretical framework, empirical evidence supports the relevance of the interaction between adaptation and expectation as a determinant of capitulation decisions. Keywords: Investments , Adaptation , Reference Point , Capitulation , Selling Decisions , Disposition Effect , Financial Markets JEL Classification: D91, D03, D81
We provide a comprehensive analysis of the determinants of trading in the sovereign credit default swaps (CDS) market, using weekly data for single-name sovereign CDS from October 2008 to September 2015. We describe the anatomy of the sovereign CDS market, derive a law of motion for gross positions and their components, and identify the key factors that drive the cross-sectional and time-series properties of trading volume and net notional amounts outstanding. While a single principal component accounts for 54 percent of the variation in sovereign CDS spreads, the largest common factor explains only 7 percent of the variation in sovereign CDS net notional amounts outstanding. Moreover, unlike for CDS spreads, common global factors explain very little of the variation in sovereign CDS trading and net notional amounts outstanding, suggesting that it is driven primarily by idiosyncratic country risk. We analyze several local and regional channels that may explain the trading in sovereign CDS: (a) country-specific credit risk shocks, including changes in a country's credit rating and related outlook changes, (b) the announcement and issuance of domestic and international debt, (c) macroeconomic sentiment derived from conventional and unconventional monetary policy, macro-economic news and shocks, and (d) regulatory channels, such as changes in bank capital adequacy requirements. All our findings suggest that sovereign CDS are more likely used for hedging than for speculative purposes.
This Policy Letter presents two event studies based on the pre-war data that foreshadows the remarkable way in which Russian economy was able to withstand the pressure from unprecedented package of international sanctions. First, it shows that a sudden stop of one of the two domestic producers of zinc in 2018 did not lead to a slowdown in the steel industry, which heavily relied on this input. Second, it demonstrates that a huge increase in cost of fuel called mazut in 2020 had virtually no impact on firms that used it, even in the regions where it was hard to substitute it for alternative fuels. This Policy Letter argues that such stability in production can be explained by the fact that Russian economy is heavily oriented toward commodities. It is much easier to replace a commodity supplier than a supplier of manufacturing goods, and many commodity producers operate at high profit margins that allow them to continue to operate even after big increases in their costs. Thus, sanctions had a much smaller impact on Russia than they would have on an economy with larger manufacturing sector, where inputs are less substitutable and profit margins are smaller.
A large empirical literature has shown that user fees signicantly deter public service utilization in developing #countries. While most of these results reflect partial equilibrium analysis, we find that the nationwide abolition of public school fees in Kenya in 2003 led to no increase in net public enrollment rates, but rather a dramatic shift toward private schooling. Results suggest this divergence between partial- and general-equilibrium effects is partially explained by social interactions: the entry of poorer pupils into free education contributed to the exit of their more affluent peers.
It is increasingly recognized that neuroscience has not delivered the revolutionary clinical possibilities for psychiatry that had been promised. Explanations differ, however: some proponents emphasize the divide between biopsychosocial psychiatry and mechanistic neurology. Others rely on further basic experimental neuroscience as only the most elementary level of explanation will allow us to fully understand and treat mental disorders. From a clinical-neuropsychological perspective, I shall argue that both views are mistaken. Diagnosis and treatment of neurological diseases demands a biopsychosocial perspective similar to psychiatry. Acknowledging this might help to bring both disciplines together and improve clinical outcome.
Why bank money creation?
(2022)
We provide a rationale for bank money creation in our current monetary system by investigating its merits over a system with banks as intermediaries of loanable funds. The latter system could result when CBDCs are introduced. In the loanable funds system, households limit banks’ leverage ratios when providing deposits to make sure they have enough “skin in the game” to opt for loan monitoring. When there is unobservable heterogeneity among banks with regard to their (opportunity) costs from monitoring, aggregate lending to bank-dependent firms is inefficiently low. A monetary system with bank money creation alleviates this problem, as banks can initiate lending by creating bank deposits without relying on household funding. With a suitable regulatory leverage constraint, the gains from higher lending by banks with a high repayment pledgeability outweigh losses from banks which are less diligent in monitoring. Bank-risk assessments, combined with appropriate risk-sensitive capital requirements, can reduce or even eliminate such losses.
Organized running events have gained substantial popularity. This study aimed to elucidate the prevalence of musculoskeletal pain, knowledge about injury prevention as well as the attitudes and motivations of individuals participating in the JP Morgan Corporate Challenge in Frankfurt (Germany). A total of 720 recreational runners completed a digital questionnaire immediately prior to the start. The majority of them displayed low to moderate physical activity levels and were rather unambitious regarding targeted finishing time. One quarter (25.3%) participated for the first time in an organized race. The most stated reasons to register were team building (76.4%) and experiencing the run’s atmosphere (50.6%). In contrast, improving health played a minor role (19.4%). More than one in five individuals (n = 159 runners) reported pain, with the most common locations being the knee and lower back. Both at rest (3.2/10 on a numerical rating scale) and during activity (4.7/10), average pain intensity was clinically relevant. Almost three thirds of the participants believed that stretching and wearing appropriate shoes would be effective for injury prevention while other methods such as resistance training, balance exercise or wearing of orthoses were rarely named. Musculoskeletal pain is a significant burden in runners participating in an urban mass event. In view of the poor knowledge about injury prevention, organizers and coaches may consider offering structured preparation programs as well as tailored running-related health education.
A(syn)-U/T and G(syn)-C+ Hoogsteen (HG) base pairs (bps) are energetically more disfavored relative to Watson–Crick (WC) bps in A-RNA as compared to B-DNA by >1 kcal/mol for reasons that are not fully understood. Here, we used NMR spectroscopy, optical melting experiments, molecular dynamics simulations and modified nucleotides to identify factors that contribute to this destabilization of HG bps in A-RNA. Removing the 2′-hydroxyl at single purine nucleotides in A-RNA duplexes did not stabilize HG bps relative to WC. In contrast, loosening the A-form geometry using a bulge in A-RNA reduced the energy cost of forming HG bps at the flanking sites to B-DNA levels. A structural and thermodynamic analysis of purine-purine HG mismatches reveals that compared to B-DNA, the A-form geometry disfavors syn purines by 1.5–4 kcal/mol due to sugar-backbone rearrangements needed to sterically accommodate the syn base. Based on MD simulations, an additional penalty of 3–4 kcal/mol applies for purine-pyrimidine HG bps due to the higher energetic cost associated with moving the bases to form hydrogen bonds in A-RNA versus B-DNA. These results provide insights into a fundamental difference between A-RNA and B-DNA duplexes with important implications for how they respond to damage and post-transcriptional modifications.
In the euro area, monetary policy is conducted by a single central bank for 20 member countries. However, countries are heterogeneous in their economic development, including their inflation rates. This paper combines a New Keynesian model and a neural network to assess whether the European Central Bank (ECB) conducted monetary policy between 2002 and 2022 according to the weighted average of the inflation rates within the European Monetary Union (EMU) or reacted more strongly to the inflation rate developments of certain EMU countries.
The New Keynesian model first generates data which is used to train and evaluate several machine learning algorithms. They authors find that a neural network performs best out-of-sample. They use this algorithm to generally classify historical EMU data, and to determine the exact weight on the inflation rate of EMU members in each quarter of the past two decades. Their findings suggest disproportional emphasis of the ECB on the inflation rates of EMU members that exhibited high inflation rate volatility for the vast majority of the time frame considered (80%), with a median inflation weight of 67% on these countries. They show that these results stem from a tendency of the ECB to react more strongly to countries whose inflation rates exhibit greater deviations from their long-term trend.
Public health authorities in Germany regard communication as a crucial part of infectious disease prevention and control strategies. Communication becomes even more important during public health crises such as pandemics. Drawing on Briggs and Hallin’s concept of biocommunicability, we analysed the German National Pandemic Plan and key informant interviews with public health experts, critical infrastructure providers and ambulance services. We examined the projected expectations towards the behaviour of the audiences and the projected ways of information circulation informing public health communication strategies during a pandemic. Participants shared the expectation that the population would react towards an influenza pandemic with panic and fear due to a lack of information or a sensationalist media coverage. They associated the information uptake of their target audience with trust in their expertise. While our informants from public health conceptualised trust in terms of a face-to-face interaction, they sought to gain trust through transparency in their respective institutional settings. Our analysis suggests that this moved health information into a political register where their medical authority was open to debate. In response to this, they perceived the field of communication as a struggle for hegemony.
External linkages allow nascent ventures to access crucial resources during the process of new product development. Forming external linkages can substantially contribute to a venture’s performance. However, little is known about the paths of external linkage formation, as well as the circumstances that drive the choice to pursue one rather than another path. This gap deserves further investigation, because we do not know whether insights developed for incumbent firms also apply to nascent ventures: To address this gap, we explore a novel dataset of 370 venture creation processes. Using sequence analyses based on optimal matching techniques and cluster analyses, we reveal that nascent ventures pursue one of overall four distinct paths of linkage formation activities during new product development. Contrary to the findings of the strategy literature, we find that if nascent ventures engage in external linkages at all, they do not combine exploration- and exploitation-oriented linkages but form either exploration- or exploitation-oriented linkages. Additional regression analyses highlight the circumstances that lead nascent ventures to pursue one rather than the other pathways. Taken together, our analyses point out that resource scarcity constitutes an important factor shaping the linkage formation activities of nascent ventures. Accordingly, we show that nascent ventures tend not to optimize by adding complementary knowledge to the firm’s knowledge base but rather to extend the existing knowledge base—a strategy which we call bricolage.
Reconstructing the evolution of baleen whales (Mysticeti) has been problematic because morphological and genetic analyses have produced different scenarios. This might be caused by genomic admixture that may have taken place among some rorquals. We present the genomes of six whales, including the blue whale (Balaenoptera musculus), to reconstruct a species tree of baleen whales and to identify phylogenetic conflicts. Evolutionary multilocus analyses of 34,192 genome fragments reveal a fast radiation of rorquals at 10.5 to 7.5 million years ago coinciding with oceanic circulation shifts. The evolutionarily enigmatic gray whale (Eschrichtius robustus) is placed among rorquals, and the blue whale genome shows a high degree of heterozygosity. The nearly equal frequency of conflicting gene trees suggests that speciation of rorqual evolution occurred under gene flow, which is best depicted by evolutionary networks. Especially in marine environments, sympatric speciation might be common; our results raise questions about how genetic divergence can be established.
Species is the fundamental taxonomic unit in biology and its delimitation has implications for conservation. In giraffe (Giraffa spp.), multiple taxonomic classifications have been proposed since the early 1900s.1 However, one species with nine subspecies has been generally accepted,2 likely due to limited in-depth assessments, subspecies hybridizing in captivity,3,4 and anecdotal reports of hybrids in the wild.5 Giraffe taxonomy received new attention after population genetic studies using traditional genetic markers suggested at least four species.6,7 This view has been met with controversy,8 setting the stage for debate.9,10 Genomics is significantly enhancing our understanding of biodiversity and speciation relative to traditional genetic approaches and thus has important implications for species delineation and conservation.11 We present a high-quality de novo genome assembly of the critically endangered Kordofan giraffe (G. camelopardalis antiquorum)12 and a comprehensive whole-genome analysis of 50 giraffe representing all traditionally recognized subspecies. Population structure and phylogenomic analyses support four separately evolving giraffe lineages, which diverged 230–370 ka ago. These lineages underwent distinct demographic histories and show different levels of heterozygosity and inbreeding. Our results strengthen previous findings of limited gene flow and admixture among putative giraffe species6,7,9 and establish a genomic foundation for recognizing four species and seven subspecies, the latter of which should be considered as evolutionary significant units. Achieving a consensus over the number of species and subspecies in giraffe is essential for adequately assessing their threat level and will improve conservation efforts for these iconic taxa.
Current theories of schizophrenia (ScZ) posit that the symptoms and cognitive dysfunctions arise from a dysconnection syndrome. However, studies that have examined this hypothesis with physiological data at realistic time scales are so far scarce. The current study employed a state-of-the-art approach using Magnetoencephalography (MEG) to test alterations in large-scale phase synchronization in a sample of n = 16 chronic ScZ patients, 10 males and n = 19 healthy participants, 10 males, during a perceptual closure task. We identified large-scale networks from source reconstructed MEG data using data-driven analyses of neuronal synchronization. Oscillation amplitudes and interareal phase-synchronization in the 3–120 Hz frequency range were estimated for 400 cortical parcels and correlated with clinical symptoms and neuropsychological scores. ScZ patients were characterized by a reduction in γ-band (30–120 Hz) oscillation amplitudes that was accompanied by a pronounced deficit in large-scale synchronization at γ-band frequencies. Synchronization was reduced within visual regions as well as between visual and frontal cortex and the reduction of synchronization correlated with elevated clinical disorganization. Accordingly, these data highlight that ScZ is associated with a profound disruption of transient synchronization, providing critical support for the notion that core aspect of the pathophysiology arises from an impairment in coordination of distributed neural activity.
We show strong overall and heterogeneous economic incidence effects, as well as distortionary effects, of only shifting statutory incidence (i.e., the agent on which taxes are levied), without any tax rate change. For identification, we exploit a tax change and administrative data from the credit market: (i) a policy change in 2018 in Spain shifting an existing mortgage tax from being levied on borrowers to being levied on banks; (ii) some areas, for historical reasons, were exempt from paying this tax (or have different tax rates); and (iii) an exhaustive matched credit register. We find the following robust results: First, after the policy change, the average mortgage rate increases consistently with a strong – but not complete – tax pass-through. Second, there is a large heterogeneity in such pass-through: larger for borrowers with lower income, a smaller number of lending relationships, not working for the lender, or facing less banks in their zip-code, thereby suggesting a bargaining power mechanism at work. Third, despite no variation in the tax rate, and consistent with the non-full tax pass-through, the tax shift increases banks’ risk-taking. More affected banks reduce costly mortgage insurance in case of loan default (especially so if banks have weaker ex-ante balance sheets) and expand into non-affected but (much) ex-ante riskier consumer lending, experiencing even higher ex-post defaults within consumer loans.
Who should hold bail-inable debt and how can regulators police holding restrictions effectively?
(2023)
This paper analyses the demand-side prerequisites for the efficient application of the bail-in tool in bank resolution, scrutinises whether the European bank crisis management and deposit insurance (CMDI) framework is apt to establish them, and proposes amendments to remedy identified shortcomings.
The first applications of the new European CMDI framework, particularly in Italy, have shown that a bail-in of debt holders is especially problematic if they are households or other types of retail investors. Such debt holders may be unable to bear losses, and the social implications of bailing them in may create incentives for decision makers to refrain from involving them in bank resolution. In turn, however, if investors can expect resolution authorities (RAs) to behave inconsistently over time and bail-out bank capital and debt holders despite earlier vows to involve them in bank rescues, the pricing and monitoring incentives that the crisis management framework seeks to invigorate would vanish. As a result, market discipline would be suboptimal and moral hazard would persist. Therefore, the policy objectives of the CMDI framework will only be achieved if critical bail-in capital is not held by retail investors without sufficient loss-bearing capacity. Currently, neither the CMDI framework nor capital market regulation suffice to assure that this precondition is met. Therefore, some amendments are necessary. In particular, debt instruments that are most likely to absorb losses in resolution should have a high minimum denomination and banks should not be allowed to self-place such securities.
In recent decades, the assessment of instructional quality has grown into a popular and well-funded arm of educational research. The present study contributes to this field by exploring first impressions of untrained raters as an innovative approach of assessment. We apply the thin slice procedure to obtain ratings of instructional quality along the dimensions of cognitive activation, classroom management, and constructive support based on only 30 s of classroom observations. Ratings were compared to the longitudinal data of students taught in the videos to investigate the connections between the brief glimpses into instructional quality and student learning. In addition, we included samples of raters with different backgrounds (university students, middle school students and educational research experts) to understand the differences in thin slice ratings with respect to their predictive power regarding student learning. Results suggest that each group provides reliable ratings, as measured by a high degree of agreement between raters, as well predictive ratings with respect to students’ learning. Furthermore, we find experts’ and middle school students’ ratings of classroom management and constructive support, respectively, explain unique components of variance in student test scores. This incremental validity can be explained with the amount of implicit knowledge (experts) and an attunement to assess specific cues that is attributable to an emotional involvement (students).
We consider an additively time-separable life-cycle model for the family of power period utility functions u such that u0(c) = c−θ for resistance to inter-temporal substitution of θ > 0. The utility maximization problem over life-time consumption is dynamically inconsistent for almost all specifications of effective discount factors. Pollak (1968) shows that the savings behavior of a sophisticated agent and her naive counterpart is always identical for a logarithmic utility function (i.e., for θ = 1). As an extension of Pollak’s result we show that the sophisticated agent saves a greater (smaller) fraction of her wealth in every period than her naive counterpart whenever θ > 1 (θ < 1) irrespective of the specification of discount factors. We further show that this finding extends to an environment with risky returns and dynamically inconsistent Epstein-Zin-Weil preferences.
This paper studies discrete time finite horizon life-cycle models with arbitrary discount functions and iso-elastic per period power utility with concavity parameter θ. We distinguish between the savings behavior of a sophisticated versus a naive agent. Although both agent types have identical preferences, they solve different utility maximization problems whenever the model is dynamically inconsistent. Pollak (1968) shows that the savings behavior of both agent types is nevertheless identical for logarithmic utility (θ = 1). We generalize this result by showing that the sophisticated agent saves in every period a greater fraction of her wealth than the naive agent if and only if θ ≥ 1. While this result goes through for model extensions that preserve linearity of the consumption policy function, it breaks down for non-linear model extensions.
We present novel evidence on the value of cross-border political access. We analyze data on meetings of US multinational enterprises (MNEs) with European Commission (EC) policymakers. Meetings with Commissioners are associated with positive abnormal equity returns. We study channels of value creation through political access in the areas of regulation and taxation. US enterprises with EC meetings are more likely to receive favorable outcomes in their European merger decisions and have lower effective tax rates on foreign income than their peers without meetings. Our results suggest that access to foreign policymakers is of substantial value for MNEs.
This paper uses a novel account of non-ideal political action that can justify radical responses to severe climate injustice, including and especially deliberate attempts to engineer the climate system in order reflect sunlight into space and cooling the planet. In particular, it discusses the question of what those suffering from climate injustice may do in order to secure their fundamental rights and interests in the face of severe climate change impacts. Using the example of risky geoengineering strategies such as sulfate aerosol injections, I argue that peoples that are innocently subject to severely negative climate change impacts may have a special permission to engage in large-scale yet risky climate interventions to prevent them. Furthermore, this can be true even if those interventions wrongly harm innocent people.
Mobilizations in defence of ‘companion animals’ have become major sites of contestation in Chinese society in recent years. They often reject the existing ambiguity between the use of these animals as pets and as meat, demanding unambiguous respect for and protection of dogs. However, in a society where inequalities are as significant as in China, where the level of poverty, sickness, and environmental and industrial tragedies appears overwhelming, one may ask how pets’ destinies have become such a symbolic focus and source of occasional fury – for both Chinese and foreign audiences. Taking this question seriously, this article aims to examine such mobilizations in China – demanding the protection of dogs – as a starting point to theoretically unwrap the more general problem of how the perception of certain beings as ‘weak’ and as deserving the protection of society is socially constructed, and what the related choices imply. I argue that to better understand these mobilizations to protect dogs, we should not separate the focus of the calls for protection from the social web of relationships and oppositions in which they are entrenched.
Homestead exemptions to personal bankruptcy allow households to retain their home equity up to a limit determined at the state level. Households that may experience bankruptcy thus have an incentive to bias their portfolios towards home equity. Using US household data for the period 1996 to 2006, we find that household demand for real estate is relatively high if the marginal investment in home equity is covered by the exemption. The home equity bias is more pronounced for younger households that face more financial uncertainty and therefore have a higher ex ante probability of bankruptcy.
Homestead exemptions to personal bankruptcy allow households to retain their home equity up to a limit determined at the state level. Households that may experience bankruptcy thus have an incentive to bias their portfolios towards home equity. Using US household data from the Survey of Income and Program Participation for the period 1996-2006, we find that especially households with low net worth maintain a larger share of their wealth as home equity if a larger homestead exemption applies. This home equity bias is also more pronounced if the household head is in poor health, increasing the chance of bankruptcy on account of unpaid medical bills. The bias is further stronger for households with mortgage finance, shorter house tenures, and younger household heads, which taken together reflect households that face more financial uncertainty.
Who gains from inter-corporate credit? To answer this question we measure the impact of the announcements of inter-corporate loans in China on the stock prices of the firms involved. We find that the average abnormal return for the issuers of inter-corporate loans is significantly negative, whereas it is positive for the receivers. Issuing firms may be perceived by investors to have run out of worthwhile projects to finance, while receiving firms are being certified as creditworthy. Subsequent firm performance and investment confirms these valuations as overall accurate.
Manipulative communications touting stocks are common in capital markets around the world. Although the price distortions created by so-called “pump-and-dump” schemes are well known, little is known about the investors in these frauds. By examining 421 “pump-and-dump” schemes between 2002 and 2015 and a proprietary set of trading records for over 110,000 individual investors from a major German bank, we provide evidence on the participation rate, magnitude of the investments, losses, and the characteristics of the individuals who invest in such schemes. Our evidence suggests that participation is quite common and involves sizable losses, with nearly 6% of active investors participating in at least one “pump-and-dump” and an average loss of nearly 30%. Moreover, we identify several distinct types of investors, some of which should not be viewed as falling prey to these frauds. We also show that portfolio composition and past trading behavior can better explain participation in touted stocks than demographics. Our analysis offers insights into the challenges associated with designing effective investor protection against market manipulation.
The future of work has become a pressing matter of concern: Researchers, business consultancies, and industrial companies are intensively studying how new work models could be best implemented to increase workplace flexibility and creativity. In particular, the agile model has become one of the “must-have” elements for re-organizing work practices, especially for technology development work. However, the implementation of agile work often comes together with strong presumptions: it is regarded as an inevitable tool that can be universally integrated into different workplaces while having the same outcome of flexibility, transparency, and flattened hierarchies everywhere. This paper challenges such essentializing assumptions by turning agile work into a “matter of care.” We argue that care work occurs in contexts other than feminized reproductive work, namely, technology development. Drawing on concepts from feminist Science and Technology Studies and ethnographic research at agile technology development workplaces in Germany and Kenya, we examine what work it takes to actually keep up with the imperative of agile work. The analysis brings the often invisibilized care practices of human and nonhuman actors to the fore that are necessary to enact and stabilize the agile promises of flexibilization, co-working, and rapid prototyping. Revealing the caring sociotechnical relationships that are vital for working agile, we discuss the emergence of power asymmetries characterized by hierarchies of skills that are differently acknowledged in the daily work of technology development. The paper ends by speculating on the emancipatory potential of a care perspective, by which we seek to inspire careful Emancipatory Technology Studies.
When experienced in-person, engagement with art has been associated with positive outcomes in well-being and mental health. However, especially in the last decade, art viewing, cultural engagement, and even ‘trips’ to museums have begun to take place online, via computers, smartphones, tablets, or in virtual reality. Similarly, to what has been reported for in-person visits, online art engagements—easily accessible from personal devices—have also been associated to well-being impacts. However, a broader understanding of for whom and how online-delivered art might have well-being impacts is still lacking. In the present study, we used a Monet interactive art exhibition from Google Arts and Culture to deepen our understanding of the role of pleasure, meaning, and individual differences in the responsiveness to art. Beyond replicating the previous group-level effects, we confirmed our pre-registered hypothesis that trait-level inter-individual differences in aesthetic responsiveness predict some of the benefits that online art viewing has on well-being and further that such inter-individual differences at the trait level were mediated by subjective experiences of pleasure and especially meaningfulness felt during the online-art intervention. The role that participants' experiences play as a possible mechanism during art interventions is discussed in light of recent theoretical models.
This paper compares the shareholder-value-maximizing capital structure and pricing policy of insurance groups against that of stand-alone insurers. Groups can utilise intra-group risk diversification by means of capital and risk transfer instruments. We show that using these instruments enables the group to offer insurance with less default risk and at lower premiums than is optimal for standalone insurers. We also take into account that shareholders of groups could find it more difficult to prevent inefficient overinvestment or cross-subsidisation, which we model by higher dead-weight costs of carrying capital. The tradeoff between risk diversification on the one hand and higher dead-weight costs on the other can result in group building being beneficial for shareholders but detrimental for policyholders.
This is the 23. article in our series Trouble on the Far-Right.
ccording to several observers new waves of refugees’ arrivals could increase the popularity of far right organizations.1 In these interpretations electoral and political support should be promoted by societal resonance of ethnocentric discourses. Recent data from the Eurobarometer illustrates that in EU-member states migration from non-EU countries is now considered to be the most important concern that the Union is facing. This is a sudden shift with respect to the results of the 2013 Eurobarometer where – in the middle of the euro crisis – EU citizens seemed to be more concerned about the economy and unemployment. I propose to place the magnifying glass on the arguments developed by these organizations by focusing on the least researched members of the far right family: nonparty organizations. After introducing CasaPound Italia (CPI) it will be discussed what fuels its anti-migrant’s discourse by highlighting continuities and changes with respect to classic nativist far right rhetoric. Digging into the arguments is crucial to getting a better assessment of their potential appeal especially in a favorable context...
This paper investigates the determinants of value and growth investing in a large administrative panel of Swedish residents over the 1999-2007 period. We document strong relationships between a household’s portfolio tilt and the household’s financial and demographic characteristics. Value investors have higher financial and real estate wealth, lower leverage, lower income risk, lower human capital, and are more likely to be female than the average growth investor. Households actively migrate to value stocks over the life-cycle and, at higher frequencies, dynamically offset the passive variations in the value tilt induced by market movements. We verify that these results are not driven by cohort effects, financial sophistication, biases toward popular or professionally close stocks, or unobserved heterogeneity in preferences. We relate these household-level results to some of the leading explanations of the value premium.
Cryptocurrencies have received growing attention from individuals, the media, and regulators. However, little is known about the investors whom these financial instruments attract. Using administrative data, we describe the investment behavior of individuals who invest in cryptocurrencies with structured retail products. We find that cryptocurrency investors are active traders, prone to investment biases, and hold risky portfolios. In line with attention effects and anticipatory utility, we find that the average cryptocurrency investor substantially increases log-in and trading activity after his or her first cryptocurrency purchase. Our results document which investors are more likely to adopt new financial products and help inform regulators about investors' vulnerability to cryptocurrency investments.
Whither artificial intelligence? Debating the policy challenges of the upcoming transformation
(2018)
Whiteout: animal traces in Werner Herzog’s Grizzly man and encounters at the end of the world
(2017)
Literary animal studies are confronted with a systematic question: How can writing, as a human-made sign system, represent the nonhuman animal as an autonomous agent without falling back into the pitfalls of anthropomorphism? Against the backdrop of this problem, this paper asks how the medium of film allows for a different representation of the animal and analyzes two of Werner Herzog’s later documentary films. Although the depiction of animals and landscapes has always played a significant part in Herzog’s films, critical assessments of his work—including those of Herzog himself—tended to view the role of nature imagery as purely allegorical: it expresses the inner nature, the inner landscapes of the film’s human protagonists. This paper tries to open up a different view. It argues that both Grizzly Man and Encounters at the End of the World develop an aesthetic that depicts nonhuman nature as an autonomous and lively presence. In the close proximity amongst camera, human, and nonhuman agents, a clear distinction between nature and culture is increasingly blurred.
The current management of a primary IgE-mediated peanut allergy consists of the two basic pillars “exposure prophylaxis” with avoidance of the allergen and “emergency therapy” with short-term treatment of an acute allergic reaction after accidental ingestion. Accidental reactions are common despite attempted avoidance. The severity of an allergic or even anaphylactic reaction after accidental ingestion is difficult to assess prior to reaction. In addition, reaction thresholds may vary depending on the accompanying augmentation factor. Therefore, every peanut allergic patient should receive individual dietary counseling as well as instructions for the use of the emergency kit and a structured patient education program (anaphylaxis group training), if necessary. For the first time, since fall 2021 a causal treatment option with a drug for oral immunotherapy will now be available for 4‑ to 17-year-old peanut-allergic children and adolescents. The oral immunotherapy with peanut protein as defatted powder of Arachis hypogaea L., semen (peanuts) leads to desensitization with a good efficacy record and an acceptable safety profile. Other treatment options with different therapeutic approaches are also under development and will probably expand the range for treatment in the coming years.
White paper peanut allergy
(2022)
The current management of a primary IgE-mediated peanut allergy consists of the two basic pillars “exposure prophylaxis” with avoidance of the allergen and “emergency therapy” with short-term treatment of an acute allergic reaction after accidental ingestion. Accidental reactions are common despite attempted avoidance. The severity of an allergic or even anaphylactic reaction after accidental ingestion is difficult to assess prior to reaction. In addition, reaction thresholds may vary depending on the accompanying augmentation factor. Therefore, every peanut allergic patient should receive individual dietary counseling as well as instructions for the use of the emergency kit and a structured patient education program (anaphylaxis group training), if necessary. For the first time, since fall 2021 a causal treatment option with a drug for oral immunotherapy will now be available for 4‑ to 17-year-old peanut-allergic children and adolescents. The oral immunotherapy with peanut protein as defatted powder of Arachis hypogaea L., semen (peanuts) leads to desensitization with a good efficacy record and an acceptable safety profile. Other treatment options with different therapeutic approaches are also under development and will probably expand the range for treatment in the coming years.
Background: Peanuts are a member of the legume family (botanical family Leguminosae) and peanut allergies are the most common cause of food anaphylaxis in many countries. The prevalence of peanut allergy is increasing.
Methods: Experts from Germany and Austria performed a standardized literature search and published their consensus recommendations in a White Paper on Peanut Allergy, which this care pathway is based upon, thus, providing a comprehensive diagnosis and treatment algorithm.
Results: The most important diagnostic key elements include a detailed clinical medical history, evidence of peanut-specific sensitization by means of skin prick testing and/or in vitro determination of the peanut (extract)-specific IgE and/or the molecular component diagnostics (most important Ara h 2-specific IgE, sometimes also Ara h1-, 3-, 6-, 8- and 9-specific IgE) as well as the gold standard, the double-blind, placebo-controlled food challenge. The diagnostic algorithms were created for the following constellations: Suspected primary peanut allergy with a clear history of systemic immediate-type reaction, suspected primary peanut allergy with questionable symptoms, suspected secondary (possibly pollen-associated) peanut allergy with a history of solely oropharyngeal symptoms and incidental finding of sensitization and no peanut ingestion so far.
Conclusions: After established diagnosis the standard of care is counseling to avoid peanut contact and prescription of emergency medications (oral antihistamines, oral steroids, inhaled β2-agonists, injectable intramuscular epinephrine) as needed. Instruction on the use of these emergency medications should be provided. A preparation for oral immunotherapy (OIT) for 4 to 17 years old peanut allergic children/ adolescents has been recently approved by the regulatory authorities. OIT for peanut allergy shows high efficacy and an acceptable safety profile, improves quality of life, and health economic aspects. Thus it offers a therapeutic option for peanut allergic children and adolescents.
Objective: Studies using diffusion tensor imaging (DTI) to investigate white matter (WM) microstructure in youths with conduct disorder (CD) have reported disparate findings. We investigated WM alterations in a large sample of youths with CD, and examined the influence of sex and callous-unemotional (CU) traits.
Method: DTI data were acquired from 124 youths with CD (59 female) and 174 typically developing (TD) youths (103 female) 9 to 18 years of age. Tract-based spatial statistics tested for effects of diagnosis and sex-by-diagnosis interactions. Associations with CD symptoms, CU traits, a task measuring impulsivity, and the impact of comorbidity, and age- and puberty-related effects were examined.
Results: Youths with CD exhibited higher axial diffusivity in the corpus callosum and lower radial diffusivity and mean diffusivity in the anterior thalamic radiation relative to TD youths. Female and male youths with CD exhibited opposite changes within the internal capsule, fornix, posterior thalamic radiation, and uncinate fasciculus. Within the CD group, CD symptoms and callous traits exerted opposing influences on corpus callosum axial diffusivity, with callous traits identified as the unique clinical feature predicting higher axial diffusivity and lower radial diffusivity within the corpus callosum and anterior thalamic radiation, respectively. In an exploratory analysis, corpus callosum axial diffusivity partially mediated the association between callous traits and impulsive responses to emotional faces. Results were not influenced by symptoms of comorbid disorders, and no age- or puberty-related interactions were observed.
Conclusion: WM alterations within the corpus callosum represent a reliable neuroimaging marker of CD. Sex and callous traits are important factors to consider when examining WM in CD.
White matter microstructural changes and episodic memory disturbances in late-onset bipolar disorder
(2018)
Background: Bipolar disorder (BD) has been associated with distributed network disruption, but little is known on how different clinical subtypes, particularly those with an earlier and later onset of disease, are related to connectivity changes in white matter (WM) tracts.
Methods: Diffusion tensor imaging (DTI) and volumetric measures were carried out in early-onset bipolar patients [(EOD) (n = 16)], late-onset bipolar disorder [(LOD)(n = 14)] and healthy controls (n = 32). We also computed ROI analysis of gray matter (GM) and white matter (WM) volumes using the regions with significant group differences in the DTI parameters. Cognitive and behavior measurements were analyzed between groups.
Results: Lower fraction of anisotropy (FA) in the right hemisphere comprising anterior thalamic radiation, fornix, posterior cingulate, internal capsule, splenium of corpus callosum was observed in the LOD in comparison with EOD; additionally, lower FA was also found in the LOD in comparison with healthy controls, mostly in the right hemisphere and comprising fibers of the splenium of the corpus callosum, cingulum, superior frontal gyrus and posterior thalamic radiation; LOD also showed worse episodic memory performance than EOD; no statistical significant differences between mood symptoms, WM and GM volumes were found between BD groups.
Conclusion: Even after correcting for age differences, LOD was associated with more extensive WM microstructural changes and worse episodic memory performance than EOD; these findings suggest that changes in the WM fiber integrity may be associated with a later presentation of BD, possibly due to mechanisms other than neuroprogression. However, these findings deserve replication in larger, prospective, studies.