Refine
Year of publication
Document Type
- Working Paper (1496)
- Part of Periodical (568)
- Article (205)
- Report (141)
- Book (100)
- Doctoral Thesis (70)
- Contribution to a Periodical (44)
- Conference Proceeding (21)
- Part of a Book (13)
- Periodical (12)
Is part of the Bibliography
- no (2698)
Keywords
- Deutschland (98)
- Financial Institutions (90)
- Capital Markets Union (65)
- ECB (65)
- Financial Markets (59)
- Banking Union (50)
- Banking Regulation (49)
- Household Finance (45)
- Monetary Policy (41)
- Banking Supervision (40)
Institute
- Wirtschaftswissenschaften (2698) (remove)
This study looks at the interrelationship between fiscal policy and safe assets as there is surprisingly little analysis about this beyond fleeting references. The study argues that from a certain point more public debt will not “buy” more safety: countries face a kind of “safe-assets Laffer curve” with a maximum amount of safe assets at some level of indebtedness. The position and “stability” of this curve depend on a number of national and international factors, including the international risk appetite and, as a more recent factor, QE policies by central banks. The study also finds evidence of declining safe assets as reflected in government debt ratings.
Using experimental data from a comprehensive field study, we explore the causal effects of algorithmic discrimination on economic efficiency and social welfare. We harness economic, game-theoretic, and state-of-the-art machine learning concepts allowing us to overcome the central challenge of missing counterfactuals, which generally impedes assessing economic downstream consequences of algorithmic discrimination. This way, we are able to precisely quantify downstream efficiency and welfare ramifications, which provides us a unique opportunity to assess whether the introduction of an AI system is actually desirable. Our results highlight that AI systems’ capabilities in enhancing welfare critically depends on the degree of inherent algorithmic biases. While an unbiased system in our setting outperforms humans and creates substantial welfare gains, the positive impact steadily decreases and ultimately reverses the more biased an AI system becomes. We show that this relation is particularly concerning in selective-labels environments, i.e., settings where outcomes are only observed if decision-makers take a particular action so that the data is selectively labeled, because commonly used technical performance metrics like the precision measure are prone to be deceptive. Finally, our results depict that continued learning, by creating feedback loops, can remedy algorithmic discrimination and associated negative effects over time.
The theoretical derivation of credit market segmentation as the result of a free market process
(2003)
Information asymmetries make it difficult for banks to assess accurately whether specific entrepreneurs are able and/or willing to repay their loans. This leads to implicit interest rate ceilings, i.e. banks "refuse" to increase their interest rates beyond this ceiling as this would lower their net returns. Although the maximum interest rate increases as the size of enterprises decreases, such ceilings nonetheless constrain the banks’ ability to set interest rates at a level that would enable them to cover costs. If transaction costs are high, the total costs associated with granting small and medium-sized loans will exceed the maximum average return which the banks can earn by issuing such loans. For this reason, banks do not lend to small and medium-sized enterprises, and, as a consequence, these businesses have no access to formal sector loans. Because micro and small enterprises have a very high RoI, it is worthwhile for them to rely on expensive informal loans to finance their operations, at least until they reach a certain size. Once they have reached this size, however, it does not make economic sense for them to continue taking out informal credits, and thus they face a growth constraint imposed by the credit market. Medium-sized enterprises earn a lower RoI than small ones, which is why borrowing in the informal credit market is not a worthwhile option for them. Moreover, they do not have access to credit from formal financial institutions, and are thus excluded from obtaining any kind of financing in either of the two credit markets. As the result of free, unregulated market forces we get a stable equilibrium in which the credit market is segmented into an informal (small loan) segment, a formal (large loan) segment and, in between, a "non-market" (medium loan) segment.
After the Global Financial Crisis a controversial rush to fiscal austerity followed in many countries. Yet research on the effects of austerity on macroeconomic aggregates was and still is unsettled, mired by the difficulty of identifying multipliers from observational data. This paper reconciles seemingly disparate estimates of multipliers within a unified and state-contingent framework. We achieve identification of causal effects with new propensity-score based methods for time series data. Using this novel approach, we show that austerity is always a drag on growth, and especially so in depressed economies: a one percent of GDP fiscal consolidation translates into 4 percent lower real GDP after five years when implemented in the slump rather than the boom. We illustrate our findings with a counterfactual evaluation of the impact of the U.K. government’s shift to austerity policies in 2010 on subsequent growth.
We use data from a German online brokerage and a survey to show that retail investors sharply reduce risk-taking in response to nearby firm bankruptcies, which are not pre- dictive of returns. The effects on trading are spatially highly concentrated, immediate and not persistent. They seem to operate through more pessimistic expected returns and increased risk aversion and do not reflect wealth effects or changes in background risks. Investors learn about bankruptcies through immediate coverage in local newspapers. Our findings suggest that non-informative local experiences that make downside risks of stock investment more salient contribute to idiosyncratic short-term fluctuations in trading.
This paper defends The Transformation of Values into Prices on the Basis of Random Systems, published in EIER, by answering to the Comments made in the same journal by Professors Mori, Morioka and Yamazaki. The clarifications mainly concern the justification of the randomness assumptions, the conditions needed to obtain the equality of total profit with total surplus value in the simplified one-industry system and the invariance of the results to changes in the units of measurement.
In recent methodological work the well known ACD approach, originally introduced by Engle and Russell (1998), has been supplemented by the involvement of an unobservable stochastic process which accompanies the underlying process of durations via a discrete mixture of distributions. The Mixture ACD model, emanating from the specialized proposal of De Luca and Gallo (2004), has proved to be a moderate tool for description of financial duration data. The use of one and the same family of ordinary distributions has been common practice until now. Our contribution incites to use the rich parameterized comprehensive family of distributions which allows for interacting different distributional idiosyncrasies. JEL classification: C41, C22, C25, C51, G14
In recent methodological work the well known ACD approach, originally introduced by Engle and Russell (1998), has been supplemented by the involvement of an unobservable stochastic process which accompanies the underlying process of durations via a discrete mixture of distributions. The Mixture ACD model, emanating from the specialized proposal of De Luca and Gallo (2004), has proved to be a moderate tool for description of financial duration data. The use of one and the same family of ordinary distributions has been common practice until now. Our contribution incites to use the rich parameterized comprehensive family of distributions which allows for interacting different distributional idiosyncrasies. JEL classification: C41, C22, C25, C51, G14.
This study contributes to the valuation of employee stock options (ESO) in two ways: First, a new pricing model is presented, admitting a major part of calculations to be solved in closed form. Designed with a focus on good replication of empirics, the model fits with publicly observable exercise characteristics better than earlier models. In particular, it is able to account for the correlation of the time of exercise and the stock price at exercise, suspected of being crucial for the option value. The impact of correlation is weak, however, whereas cancellations play a central role. The second contribution of this paper is an examination to what extent the ESO pricing method of SFAS 123 is subject to discretion of the accountant. Given my model were true, the SFAS price would be a good proxy. Yet, outside shareholders usually cannot observe one of the SFAS input parameters. On behalf of an example I show that there is wide latitude left to the accountant.
This study contributes to the valuation of employee stock options (ESO) in two ways: First, a new pricing model is presented, admitting a major part of calculations to be solved in closed form. Designed with a focus on good replication of empirics, the model fits with publicly observable exercise characteristics better than earlier models. In particular, it is able to account for the correlation of the time of exercise and the stock price at exercise, suspected of being crucial for the option value. The impact of correlation is weak, however, whereas cancellations play a central role. The second contribution of this paper is an examination to what extent the ESO pricing method of SFAS 123 is subject to discretion of the accountant. Given my model were true, the SFAS price would be a good proxy. Yet, outside shareholders usually cannot observe one of the SFAS input parameters. On behalf of an example I show that there is wide latitude left to the accountant.
This paper presents causal evidence of the effects of boardroom networks on firm value. We exploit exogenous variation in network centrality arising from a ban on interlocking directorates of Italian financial and insurance companies. We leverage this shock to show that firms that become more central in the network as a result of the shock experience positive abnormal returns around the announcement date. We find that information dissemination plays a central role: results are driven by firms that have higher idiosyncratic volatility, low analyst coverage, and more uncertainty surrounding their earnings forecasts. We also find that firms benefit more from boardroom centrality when they are more central in the input-output network, as this reinforces information complementarities, or when they are less central in the cross-ownership network, as well as when they suffer from low profitability and low growth opportunities. Network centrality also results in higher compensation for board directors.
Public employee pension systems throughout the developed world have traditionally been of the pay-as-you-go (PAYGO) defined benefit (DB) variety, where pensioner payments are financed by taxes (contributions) levied on the working generation. But as the number of retirees rises relative to the working-age group, such systems have begun to face financial distress. This trend has been exacerbated in many countries, among them Germany, by high unemployment rates producing further deterioration of the contribution base. In the long run, public sector pension benefits will have to be cut or contributions increased, if the systems are to be maintained. An alternative path sometimes offered to ease the crunch of paying for public employee pensions is to move toward funding: here, plan assets are gradually built up, invested, and enhanced returns devoted to partly defray civil servants’ pension costs. In this study, we evaluate the impact of introducing partial prefunding, paired with a strategic investment policy for the German federal state of Hesse. The analysis assesses the impact of introducing a supplementary tax-sponsored pension fund whose contributions are invested in the capital market and used to relieve the state budget from (some) pension payments. Our model determines the expectation and the Conditional Value-at-Risk of economic pension costs using a stochastic simulation process for pension plan assets. This approach simultaneously determines the optimal contribution rate and asset allocation that controls the expected economic costs of providing the promised pensions, while at the same time controlling investment risk. Specifically, we offer answers to the following questions: 1. How can the plan be designed to control cash-flow shortfall risk, so as to mitigate the potential burden borne by future generations of taxpayers? 2. What is the optimal asset allocation for this fund as it is built up, to generate a maximum return while simultaneously restricting capital market and liability risk? 3. What are reasonable combinations of annual contribution rates and asset allocation to a state-managed pension fund, which will limit costs of providing promised public sector pensions? We anticipate that this research will interest several sorts of policymaker groups. First, focusing on the German case, the state and Federal governments should find it relevant, as these entities face considerable public sector pension liabilities. Second, our findings will also be of interest to other European countries, as most have substantial underfunded defined benefit plans for civil servants. In what follows, we first offer a brief description of the structure of civil servant pensions in Germany, focusing on their benefit formulas, their financing, and the resulting current as well as future plan obligations for taxpayers. Next, we turn to an analysis of the actuarial status of the Hesse civil servants’ pension plan and evaluate how much would have to be contributed to fund this plan in a nonstochastic context. Subsequently we evaluate the asset-liability and decision-making process from the viewpoint of the plan sponsor, to determine sensible plan asset allocation behavior. A final section summarizes findings and implications.
We investigate the effect of overreaction in the fine art market. Using a unique sample of auction prices of modern prints, we define an overvalued (undervalued) print as a print that was bought for a price above (below) its high (low) auction pricing estimate. Based on the overreaction hypothesis, we predict that overvalued (undervalued) prints generate a negative (positive) excess return at a subsequent sale. Our empirical findings confirm our expectations. We report that prints that were bought for a price 10 percent above (below) its high (low) pricing estimate generate a positive (negative) excess return of 12 percent (17 percent) after controlling for the general price movement on the prints market. The price correction for overvalued (undervalued) prints is more pronounced during recessions (expansions).
The Wirecard scandal is a wake-up call alerting German politics to the importance of securities market integrity. The role of market supervision is to ensure the smooth functioning of capital markets and their integrity, creating trust among and acceptance by investors locally and globally. The existing patchwork of national supervisory practice in Europe is under discussion today, in the wake of Brexit that will end the role of London as a de-facto lead supervisor in stock and bond markets. A fundamental overhaul of a fragmented securities markets supervisory regime in Europe would offer the potential to lead to the establishment of an independent European Single Market Supervisor (ESMS). Endowed with strong enforcement powers, and supported by the existing national agencies, the ESMS would be entrusted with ensuring a uniform market standard as to transparency and other issues of market integrity across Europe. This would not rule out maintaining a variety of market organization structures at the national level. The ESMS would need executive powers in the world of markets (i.e. securities and trading), much like the SSM in the world of banking. To fill this new role, ESMS would have to be established as a new, independent institution, including an enormously scaled up staff if compared, e.g., to ESMA.
Die traditionelle Begründung der Bankenregulierung basiert auf der Vorstellung, daß regulatorische Vorschriften in erster Linie dem Schutz der Bankeinlagen und der Vermeidung von Bankkonkursen dienen. In Kapitel 2 wurde argumentiert, daß diese Begründung nicht den Kern des Regulierungsproblems trifft. Der einzige letztlich überzeugende Grund für die Regulierung der Eigenkapitalausstattung von Banken ist in Externalitäten zu sehen, die aus den derzeit praktizierten Einlagensicherungssystemen bzw. der impliziten Einlagensicherung durch den Staat resultieren. Diese Begründung für die Bankenregulierung führt dazu, daß die Aktiva der Bank ins Zentrum der Betrachtung rücken: Die Bankenregulierung muß danach dafür sorgen, daß die von Banken getroffenen Entscheidungen in bezug auf die Übernahme von Risiken in einem wohlfahrtstheoretisch effizienten Sinn getroffen werden. Wenn die Verminderung des Konkursrisikos die alleinige Zielgröße der Regulierung darstellen würde, müßte man schlußfolgern, daß ein einfaches Verbot riskanter Geschäftsaktivitäten die Erreichung des Regulierungsziels sicherstellt. Dies kann natürlich keine ernstzunehmende Option sein. Das Geschäft der Banken besteht in der Evaluierung und dem Management von Risiken. Die Regulierung muß dafür sorgen, daß Banken diese Funktion trotz bestehender Externalitäten in einer wohlfahrtstheoretisch effizienten Weise erfüllen. Um die Frage der Effizienz von Risikoentscheidungen in einem formalen Modellrahmen analysieren zu können, wurde in Kapitel 4.2.4 ein neuer Optionsbewertungsansatz für das Eigenkapital der Bank entwickelt. Dies war erforderlich, da sich die bisher in der Literatur verwendeten Ansätze nicht zur Beurteilung dieser Frage eignen. Im μ-σ-Rahmen kann zwar analysiert werden, ob die Bank ein effizientes Portfolio wählt. Der Modellansatz basiert aber auf der kritischen Annahme normalverteilter Renditen. Da die beschränkte Haftung der Eigenkapitalgeber aber eine untere Schranke der Eigenkapitalrendite bewirkt, ist diese Annahme zur Analyse der Risikowahl problematisch. Das in Kapitel 4.2.1 vorgestellte Optionsbewertungsmodell von Merton berücksichtigt zwar die beschränkte Haftung der Eigenkapitalgeber, eignet sich aber nicht zur Analyse der Effizienz der Risikowahl, weil die Wahl des Anlagerisikos irrelevant für die erreichbare Wohlfahrt ist. Als problemadäquater Modellrahmen wurde ein Modell mit risikoabhängiger Rendite von Bankanlagen entwickelt. Dieser Ansatz beruht auf der Idee, daß Banken in Spezialbereichen über superiore Informationen verfügen, die sie zur Erzielung von über den Kapitalkosten liegenden Renditen einsetzen können. Da die Bank aber nur über begrenzte Informationen verfügt, ist die Anzahl der verfügbaren Anlagen mit positivem Kapitalwert begrenzt und definiert so das effiziente Maß der Risikoübernahme. Die Analyse der Auswirkungen der Eigenkapitalregulierung zeigt, daß die risikoabhängige Eigenkapitalregulierung zwar der einfachen Verschuldungsgradregulierung überlegen ist, das Risikoanreizproblem der Banken aber nicht vollständig lösen kann. Sie stellt aber im Vergleich zu den anderen in dieser Arbeit diskutierten Regulierungsansätzen die robusteste Lösung des Problems dar. Nach weitverbreiteter Ansicht leisten Wettbewerbsbeschränkungen einen wichtigen Beitrag zur Behebung des Risikoanreizproblems, weil eine Bank bei Erhöhung des Anlagerisikos den Verlust der Monopolrente (Charterwert ) im Konkurs befürchten muß. Es wurde aber gezeigt, daß diese Kraft nur für unplausibel hohe Monopolrenten stark genug ist, um die aus der Einlagensicherung resultierenden Anreize zu kompensieren. Auch der in Kapitel 4.5 analysierte Vorschlag von John, John und Senbet (1991) einer Nutzung des Steuersystems zur Korrektur des Risikoanreizproblems erweist sich bei näherer Analyse für praktische Regulierungszwecke als unbrauchbar, da hierfür ein marginaler Steuersatz auf die Investitionsrückflüsse in der Größenordnung von 90% notwendig ist. Schließlich erweist sich auch die in Kapitel 4.4 und 4.6.1 untersuchte Idee, durch Einsatz von sicheren Reserven den Effekt einer Erhöhung der Anlagerisiken zu kompensieren, als wenig sinnvoll. Der das Merton-Modell treibende Effekt einer Anlagersikoerhöhung kann durch diese Regulierungsform erfolgreich ausgeschaltet werden. An seine Stelle treten aber unerwünschte Nebeneffekte, die im Ergebnis nach wie vor zu einer nicht effizienten Risikowahl führen. Die Nebeneffekte können nur durch einen Verzicht auf eine Eigenkapitalunterlegung ausgeschaltet werden. Obwohl prinzipiell auch ohne Eigenkapitalunterlegung die Stabilität von Banken durch entsprechend hohe sichere Reservehaltung gewährleistet werden kann, ist eine solche Regulierung aufgrund des enormen Bedarfs an sicheren Reserveanlagen nicht praktikabel. Eine Regulierungsform, die theoretisch in der Lage ist, die wohlfahrtseffiziente Risi kowahl zu implementieren, ist ein Einlagensicherungssystem mit "fair" berechneter risikoabhängiger Prämienkalkulation. Ein solches Einlagensicherungssystem verhindert effektiv jegliche Externalitäten der Risikoentscheidung der Bank auf die Einlagensicherung und führt damit automatisch zur effizienten Wahl des Anlagerisikos. Angesichts dieses Ergebnisses überrascht die Tatsache, daß sich die Aufsichtsbehörden fast vollständig auf das Instrument der Eigenkapitalregulierung konzentrieren. In Kapitel 4.6.2.2 wurden argumentiert, daß die Gründe dafür in den politischen und praktischen Problemen der Umsetzung von risikoabhängigen Einlagensicherungsprämien liegen. Da die Einlagensicherungsprämie nur in diskreten Zeitabständen fällig wird, hat eine eigenkapitalschwache Bank nach wie vor einen starken Anreiz, nach Zahlung der Prämie ihr Anlagerisiko zu erhöhen. Daneben zeigen die Erfahrungen der Savings & Loan-Krise, daß die Einlagensicherung mit erheblichen Agency-Kosten verbunden ist, weil die Mitarbeiter der Einlagensicherungsbehörde dazu neigen, notwendige Liquidationen von Banken zu unterlassen oder hinauszuzögern. In Kapitel 5 wurde das Zusammenwirken von Einlagensicherung und Eigenkapitalregulierung untersucht. Dazu wurde ein Modell verwendet, in dem eine optimale Kapitalstruktur von Banken existiert, die sich aus einem Trade Off zwischen Finanzierungskostenvorteilen der Einlagenfinanzierung und Liquidationskosten im Insolvenzfall ergibt. Das Modell zeigt, daß die Kombination einer risikoabhängigen Eigenkapitalregulierung mit einer traditionellen volumenabhängigen Einlagensicherungsprämie zwar die effiziente Stabilität von Banken, nicht dagegen effiziente Anlageentscheidungen erreicht. Wenn alternativ eine traditionelle Verschuldungsgradregulierung mit einer risikoabhängigen Einlagensicherungsprämie kombiniert wird, resultieren effiziente Anlageentscheidungen, die Bank operiert jedoch nicht im Kapitalkostenminimum. Erst die Kombination von risikoabhängiger Eigenkapitalregulierung und risikoabhängiger Einlagensicherungsprämie führt dazu, daß das insgesamt wohlfahrtsoptimale Marktergebnis implementiert wird. Die risikoabhängige Kapitalregulierung steht danach in einem komplementären Verhältnis zum risikoabhängigen Einlagensicherungssystem. Entgegen der teilweise geäußerten Warnung vor einer "doppelten" Bestrafung der Risikoübernahme durch Eigenkapitalregulierung und Einlagensicherung führt allein die Risikoabhängigkeit bei der Regulierungen zu effizienten Marktergebnissen. In Kapitel 6 wurde argumentiert, daß auch die Berücksichtigung der Interessendivergenz zwischen Kapitalgebern und Managern nichts an der Beurteilung der risikoabhängigen Eigenkapitalregulierung ändert. Die Eigeninteressen von Managern können zwar zu einer Abweichung der tatsächlich verfolgten Risikopolitik von der durch die Eigenkapitalgeber bevorzugten Risikopolitik führen. Allerdings können die Eigeninteressen des Managements in nicht prognostizierbarer Weise sowohl risikoerhöhende als auch risikosenkende Wirkung haben. Aus Sicht der Aufsichtsbehörden erscheint daher die Orientierung an Modellen sinnyoll, die ein marktwertmaximierendes Verhalten der Bank unterstellen. Schließlich zeigt die Analyse der Wechselwirkungen von Eigenkapitalregulierung und den Arbeitsanreizen des Managements, daß die risikobasierte Eigenkapitalregulierung auch im Hinblick auf diesen Agency-Konflikt positive Wirkungen entfaltet. Wenn die Drohung des Konkurses ein Motivationsinstrument für Manager darstellt. dann ist es aus Sicht der Kapitalgeber unter plausiblen Annahmen sinnvoll, auch im Hinblick auf die Auswirkung auf die Anreizsituation des Managements bei hohen Anlagerisiken eine höhere Eigenkapitalquote zu wählen. Die risikoabhängige Eigenkapitalregulierung kann allerdings den Agency-Konflikt auch verstärken, wenn die Aufsichtsbehörde ein zu hohes Konfidenzniveau vorgibt. Dann versagt die Anreizwirkung der Konkursdrohung, weil Konkurse auch bei "Fehlverhalten" des Managements so unwahrscheinlich sind, daß die Konkursdrohung keinen Einfluß auf das Verhalten von Managern hat. Zusammenfassend kann festgestellt werden, daß die risikoabhängige Eigenkapitalregulierung von Banken nicht nur die Sorgen der Aufsichtsbehörden verringert, sondern auch auf einem soliden theoretischen Fundament steht. Sie stellt die notwendige Antwort auf die durch die Einlagensicherung generierten Fehlanreize zur Übernahme exzessiver Risiken dar und leistet damit einen wichtigen Beitrag für die Funktionsfähigkeit der Kapitalmärkte. Gleichzeitig zeigt die Analyse aber auch, daß überzogenes Stabilitätsdenken der Aufsichtsbehörden nicht angebracht ist. Gerade im Fall von Banken verfügen Manager über ausgesprochen große Verhaltensspielräume und Eigenkapitalgeber über wenige Instrumente, um eine Geschäftspolitik im Sinne der Kapitalgeber durchzusetzen. Die Konkursdrohung stellt in dieser Situation ein wichtiges Instrument zur Durchsetzung des Interessengleichlaufs zwischen Management und Kapitalgebern dar. Die Bankenregulierung darf daher die Konkurswahrscheinlichkeit von Banken nicht so stark senken, daß die Drohung des Konkurses seine positive Anreizwirkung verliert.
Experiments are an important tool in economic research. However, it is unclear to which extent the control of experiments extends to the perceptions subjects form of such experimental decision situations. This paper is the first to explicitly elicit perceptions of the dictator and trust game and shows that there is substantial heterogeneity in how subjects perceive the same game. Moreover, game perceptions depend not only on the game itself but also on the order of games (i.e., the broader experimental context in which the game is embedded) and the subject herself. This highlights that the control of experiments does not necessarily extend to game perceptions. The paper also demonstrates that perceptions are correlated with game behavior and moderate the relationship between game behavior and field behavior, thereby underscoring the importance and relevance of game perceptions for economic research.
In light of the failed negotiations with Greece, Jan Krahnen argues that an effective reform agenda for Greece can only be designed by the elected government. Fundamental reforms will take time to take full effect and euro area member states will, in the meantime, have to offer Greece a basic level of economic security.
Krahnen demands that policy makers and the professional public involved view the Greek crisis as an opportunity to take the next necessary steps to formulate a reform agenda for the European Monetary Union. A community of supranational and non-party researchers and intellectuals could take the initiative and in a structured process develop a trustworthy and realistic concept that drafts the next big step towards a political union of Europe, including elements of a fiscal union.
This paper examines heterogeneity in time discounting among a representative sample of elderly Americans, as well as its role in explaining key economic behaviors at older ages. We show how older Americans evaluate simple (hypothetical) inter-temporal choices in which payments today are compared with payments in the future. Using the indicators derived from this measure, we then demonstrate that differences in discounting patterns are associated with characteristics of particular importance in elderly populations. For example, cognitive deficits are associated with greater impatience, whereas bequest motives are associated with less impatience. We then relate our discounting measure to key economic outcomes and find that impatience is associated with lower wealth, fewer investments in health, and less planning for end of life care.
Households regularly fail to make optimal financial decisions. But what are the underlying reasons for this? Using two conceptually distinct measures of time inconsistency based on bank account transaction data and behavioral measurement experiments, we show that the excessive use of bank account overdrafts is linked to time inconsistency. By contrast, there is no correlation between a survey-based measure of financial literacy and overdraft usage. Our results indicate that consumer education and information may not suffice to overcome mistakes in households’ financial decision-making. Rather, behaviorally motivated interventions targeting specific biases in decision-making should also be considered as effective policy tools.
We study whether and how time preferences change over the life cycle, exploiting representative long-term panel data. We estimate the age patterns of discount rates from age 25 to 80. In order to identify age effects, we have to disentangle them from cohort and period factors. We address this identification problem by estimating individual fixed effects models, where we substitute period effects with determinants of time preferences that depend on calendar years. We find that discount rates decrease with age and the decline is remarkably linear over the life cycle.
Discussions about the banking union have restarted. Its success so far is limited: national banking sectors are still overwhelmingly exposed to their own countries’ economies, cross border banking has not increased and capital and liquidity remain locked within national boundaries. The policy letter highlights that the current debate, centered on sovereign exposures and deposit insurance, misses critical underlying problems in the supervision and resolution frameworks. The ECB supervisors’ efforts to facilitate cross-border banking have been hampered by national ringfencing. The resolution framework is not up to its task: limited powers of the SRB, prohibitive access conditions and limited size of the Single Resolution Fund limit its effectiveness. A lack of a coherent European framework for insolvency unlevels the regulatory field and creates incentives to bypass European rules. The new Commission and European Parliament, with the new ECB leadership, provide a unique opportunity to address these shortcomings and make the banking union work.
Analysing causality among oil prices and, in general, among financial and economic variables is of central relevance in applied economics studies. The recent contribution of Lu et al. (2014) proposes a novel test for causality— the DCC-MGARCH Hong test. We show that the critical values of the test statistic must be evaluated through simulations, thereby challenging the evidence in papers adopting the DCC-MGARCH Hong test. We also note that rolling Hong tests represent a more viable solution in the presence of short-lived causality periods.
We use a unique data set from the Trade Reporting and Compliance Engine (TRACE) to study liquidity effects in the US structured product market. Our main contribution is the analysis of the relation between the accuracy in measuring liquidity and the potential degree of disclosure. Having access to all relevant trading information, we provide evidence that transaction cost measures that use dealer specific information such as trader identity and trade direction can be efficiently proxied by measures that use less detailed information. This finding is important for all market participants in the context of OTC markets, as it fosters our understanding of the information contained in transaction data. Thus, our results provide guidance for improving transparency while maintaining trader confidentiality. In addition, we analyze liquidity in the structured product market in general and show that securities that are mainly institutionally traded, guaranteed by a federal authority, or have low credit risk, tend to be more liquid.
This paper analyzes the bail-in tool under the Bank Recovery and Resolution Directive (BRRD) and predicts that it will not reach its policy objective. To make this argument, this paper first describes the policy rationale that calls for mandatory private sector involvement (PSI). From this analysis, the key features for an effective bail-in tool can be derived.
These insights serve as the background to make the case that the European resolution framework is likely ineffective in establishing adequate market discipline through risk-reflecting prices for bank capital. The main reason for this lies in the avoidable embeddedness of the BRRD’s bail-in tool in the much broader resolution process, which entails ample discretion of the authorities also in forcing private sector involvement. Moreover, the idea that nearly all positions on the liability side of a bank’s balance sheet should be subjected to bail-in is misguided. Instead, a concentration of PSI in instruments that fall under the minimum requirements for own funds and eligible liabilities (MREL) is preferable.
Finally, this paper synthesized the prior analysis by putting forward an alternative regulatory approach that seeks to disentangle private sector involvement as a precondition for effective bank-resolution as much as possible form the resolution process as such.
This paper analyses the bail-in tool under the BRRD and predicts that it will not reach its policy objective. To make this argument, this paper first describes the policy rationale that calls for mandatory PSI. From this analysis the key features for an effective bail-in tool can be derived. These insights serve as the background to make the case that the European resolution framework is likely ineffective in establishing adequate market discipline through risk-reflecting prices for bank capital. The main reason for this lies in the avoidable embeddedness of the BRRD’s bail-in tool in the much broader resolution process which entails ample discretion of the authorities also in forcing private sector involvement. Finally, this paper synthesized the prior analysis by putting forward an alternative regulatory approach that seeks to disentangle private sector involvement as a precondition for effective bank-resolution as much as possible form the resolution process as such.
The banking system is highly interconnected and these connections can be conveniently represented as an interbank network. This survey presents a systematic overview of the recent advances in the theoretical literature on interbank networks. We assess our current understanding of the structure of interbank networks, of how network characteristics affect contagion in the banking system and of how banks form connections when faced with the possibility of contagion and systemic risk. In particular, we highlight how the theoretical literature on interbank networks offers a coherent way of studying interconnections, contagion processes and systemic risk, while emphasizing at the same time the challenges that must be addressed before general results on the link between the structure of the interbank network and financial stability can be established. The survey concludes with a discussion of the policy relevance of interbank network models with a special focus on macroprudential policies and monetary policy.
A premise of the capabilities perspective in strategy is that firm-specific capabilities allow some firms to be unusually adept at exploiting growth opportunities. Since few firms have the capacity to internally generate the quantity or variety of strategic resources needed to exploit growth opportunities, the ability to externally acquire complementary resources is critical to the acquisition of competitive advantage. However, the external sourcing of resources exposes the firm’s strategic resources to risks of expropriation. We argue this threat gives capable firms incentive to use internally generated strategic resources to pursue growth opportunities before turning to external sources. A pecking order theory of strategic resource deployment is implied. Data from a 22-year sample of cross-border investment partnership decisions made by U.S.-based venture capital firms lend support to our theory.
Armstrong et al. (2022) review the empirical methods used in the accounting literature to draw causal inferences. They document a growing number of studies using quasi-experimental methods and provide a critical perspective on this trend as well as the use of these methods in the accounting literature. In this discussion, I complement their review by broadening the perspective. I argue for a design-based approach to accounting research that shifts attention from methods to the entire research design. I also discuss why studies that aim to draw causal inferences are important, how these studies fit into the scientific process, and why assessing the strength of the research design is important when evaluating studies and aggregating research findings.
There is a large, but yet growing debate about the need to complement the European monetary union with a stronger fiscal union. This paper reviews the potential trade-offs between effectiveness, moral hazard problems, and permanent redistribution. In particular, we contribute to the question of how member states may be willing to enter into a stronger fiscal union if the evolution of this union may imply large redistribution under incomplete contracting. We discuss clawback mechanisms that have been suggested in the literature, but conclude that clawbacks are undesirable, as they would essentially destroy the insurance value of a fiscal union. Instead, we propose that a clearly defined exit option as a guarantee against involuntary redistribution can make entry into a stronger fiscal union less risky and hence more attractive for member states.
Although the world of banking and finance is becoming more integrated every day, in most aspects the world of financial regulation continues to be narrowly defined by national boundaries. The main players here are still national governments and governmental agencies. And until recently, they tended to follow a policy of shielding their activities from scrutiny by their peers and members of the academic community rather than inviting critical assessments and an exchange of ideas. The turbulence in international financial markets in the 1980s, and its impact on U.S. banks, gave rise to the notion that academics working in the field of banking and financial regulation might be in a position to make a contribution to the improvement of regulation in the United States, and thus ultimately to the stability of the entire financial sector. This provided the impetus for the creation of the “U.S. Shadow Financial Regulatory Committee”. In the meantime, similar shadow committees have been founded in Europe and Japan. The specific problems associated with financial regulation in Europe, as well as the specific features which distinguish the European Shadow Financial Regulatory Committee from its counterparts in the U.S. and Japan, derive from the fact that while Europe has already made substantial progress towards economic and political integration, it is still primarily a collection of distinct nation-states with differing institutional set-ups and political and economic traditions. Therefore, any attempt to work towards a European approach to financial regulation must include an effort to promote the development of a European culture of co-operation in this area, and this is precisely what the European Shadow Financial Regulatory Committee (ESFRC) seeks to do. In this paper, Harald Benink, chairman of the ESFRC, and Reinhard H. Schmidt, one of the two German members, discuss the origin, the objectives and the functioning of the committee and the thrust of its recommendations.
Tractable hedging - an implementation of robust hedging strategies : [This Version: March 30, 2004]
(2004)
This paper provides a theoretical and numerical analysis of robust hedging strategies in diffusion–type models including stochastic volatility models. A robust hedging strategy avoids any losses as long as the realised volatility stays within a given interval. We focus on the effects of restricting the set of admissible strategies to tractable strategies which are defined as the sum over Gaussian strategies. Although a trivial Gaussian hedge is either not robust or prohibitively expensive, this is not the case for the cheapest tractable robust hedge which consists of two Gaussian hedges for one long and one short position in convex claims which have to be chosen optimally.
Using a unique data set on trade credit defaults among French firms, we investigate whether and how trade credit is used to relax financial constraints. We show that firms that face idiosyncratic liquidity shocks are more likely to default on trade credit, especially when the shocks are unexpected, firms have little liquidity, are likely to be credit constrained or are close to their debt capacity. We estimate that credit constrained firms pass more than one fourth of the liquidity shocks they face on to their suppliers down the trade credit chain. The evidence is consistent with the idea that firms provide liquidity insurance to each other and that this mechanism is able to alleviate the consequences of credit constraints. In addition, we show that the chain of defaults stops when it reaches firms that are large, liquid, and have access to financial markets. This suggests that liquidity is allocated from large firms with access to outside finance to small, credit constrained firms through trade credit chains.
We propose a novel approach to the study of international trade based on a theory of country integration that embodies a broad systemic viewpoint on the relationship between trade and growth. Our model leads to an indicator of country openness that measures a country's level of integration through the full architecture of its connections in the trade network. We apply our methodology to a sample of 204 countries and find a sizable and significant positive relationship between our integration measure and a country's growth rate, while that of the traditional measures of outward orientation is only minor and statistically insignificant.
The recent success of some alternative trading systems (ATSs) has had a strong impact on the traditional stock exchange industry and many observers expect even more dramatic changes in the future. This paper investigates the nature of competition between stock exchanges and ATSs and argues that the introduction of ATSs weakens the level of competition between market-makers in the traditional exchange...
On November 8, 2013, several members of the British House of Lords’ Subcommittee A conducted a hearing at the ECB in Frankfurt, Germany, on “Genuine Economic and Monetary Union and its Implications for the UK”. Professors Otmar Issing and Jan Pieter Krahnen were called as expert witnesses.
The testimony began with a general discussion on the elements considered necessary for a functioning internal market. Do economic union and monetary union require a fiscal union or even a political union, beyond the elements of the banking union currently being prepared? In this context, also the critique of the German current account surplus and the international expectations that Germany stimulate internal demand to support growth in crisis countries, were discussed.
With regard to the monetary union, the members of the subcommittee asked for an assessment of how European nations and the banking industry would have fared in the banking crisis that followed the Lehman collapse, had there not been a common currency. Given the important role that the ECB has played in the course of the crisis management, the members further asked for an evaluation of the OMT-program of the ECB and also if the monetary union is in need of common debt instruments, in order to provide the ECB with the possibility of buying EU liabilities, comparable to the Fed buying US Treasury bonds. Finally, the dual role of the ECB for monetary policy and banking supervision was an issue touched on by several questions.
Climate change has become one of the most prominent concerns globally. In this paper, the authors study the transition risk of greenhouse gas emission reduction in structural environmental-macroeconomic DSGE models. First, they analyze the uncertainty in model prediction on the effect of unanticipated and pre-announced carbon price increases. Second, they conduct optimal model-robust policy in different settings. They find that reducing emissions by 40% causes 0.7% to 4% output loss with 2% on average. Pre-announcement of carbon prices affects the inflation dynamics significantly. The central bank should react slightly less to inflation and output growth during the transition risk. With optimal carbon price designs, it should react even less to inflation, and more to output growth.
Telemonitoring devices can be used to screen consumers' characteristics and mitigate information asymmetries that lead to adverse selection in insurance markets. However, some consumers value their privacy and dislike sharing private information with insurers. In the second-best efficient Wilson-Miyazaki-Spence framework, we allow for consumers to reveal their risk type for an individual subjective cost and show analytically how this affects insurance market equilibria as well as utilitarian social welfare. Our analysis shows that the choice of information disclosure with respect to revelation of their risk type can substitute deductibles for consumers whose transparency aversion is sufficiently low. This can lead to a Pareto improvement of social welfare and a Pareto efficient market allocation. However, if all consumers are offered cross-subsidizing contracts, the introduction of a transparency contract decreases or even eliminates cross-subsidies. Given the prior existence of a WMS equilibrium, utility is shifted from individuals who do not reveal their private information to those who choose to reveal. Our analysis provides a theoretical foundation for the discussion on consumer protection in the context of digitalization. It shows that new technologies bring new ways to challenge crosssubsidization in insurance markets and stresses the negative externalities that digitalization has on consumers who are not willing to take part in this development.
The paper looks at the determinants of fiscal adjustments as reflected in the primary surplus of countries. Our conjecture is that governments will usually find it more attractive to pursue fiscal adjustments in a situation of relatively high growth, but based on a simple stylized model of government behavior the expectation is that mainly high trust governments will be in a position to defer consolidation to years with higher growth. Overall, our analysis of a panel of European countries provides support for this expectation. The difference in fiscal policies depending on government trust levels may help explaining why better governed countries have been found to have less severe business cycles. It suggests that trust and credibility play an important role not only in monetary policy, but also in fiscal policy.
We analyze efficient risk-sharing arrangements when the value from deviating is determined endogenously by another risk sharing arrangement. Coalitions form to insure against idiosyncratic income risk. Self-enforcing contracts for both the original coalition and any coalition formed (joined) after deviations rely on a belief in future cooperation which we term "trust". We treat the contracting conditions of original and deviation coalitions symmetrically and show that higher trust tightens incentive constraints since it facilitates the formation of deviating coalitions. As a consequence, although trust facilitates the initial formation of coalitions, the extent of risk sharing in successfully formed coalitions is declining in the extent of trust and efficient allocations might feature resource burning or utility burning: trust is indeed a double-edged sword.
Trust in policy makers fluctuates signi
cantly over the cycle and affects the transmission mechanism. Despite this it is absent from the literature. We build a monetary model embedding trust cycles; the latter emerge as an equilibrium phenomenon of a game-theoretic interaction between atomistic agents and the monetary authority. Trust affects agents' stochastic discount factors, namely the price of future risk, and through this it interacts with the monetary transmission mechanism. Using data from the Eurobarometer surveys, we analyze the link between trust and the transmission mechanism of macro and monetary shocks: Empirical results are in line with theoretical ones.
The efficacy of monetary authority actions depends primarily on the ability of the monetary authority to affect inflation expectations, which ultimately depend on agents' trust. We propose a model embedding trust cycles, as emerging from sequential coordination games between atomistic agents and the policy maker, in a monetary model. Trust affects agents' stochastic discount factor, namely the price of future risk, and their expectation formation process: these effects in turn interact with the monetary transmission mechanism. Using data from the Eurobarometer survey we analyze the link between trust on the one side and the transmission mechanism of shocks and of the policy rate on the other: data show that the two interact significantly and in a way comparable to the obtained in our model.
In the aftermath of the financial crisis, the ECB has experienced an unprecedented deterioration in the level of trust. This raises the question as to what factors determine trust in central banking. We use a unique cross-country dataset which includes a rich set of socio-economic characteristics and supplement it with variables meant to reflect a country’s macroeconomic condition. We find that besides individual socio-economic characteristics, macroeconomic conditions play a crucial role in the trust-building process. Our results suggest that agents are boundedly rational in the trust-building process and that current ECB market operations may even be beneficial for trust in the ECB in the long-run.
We examine trust and trustworthiness of individuals with varying professional preferences and experiences. Our subjects study business and economics in Frankfurt, the financial center of Germany and continental Europe. In the trust game, subjects with a high interest in working in the financial industry return 25 percent less than subjects with a low interest. We find no evidence that the extent of professional experience in the financial industry has a negative impact on trustworthiness. We also do not find any evidence that the financial industry screens out less trustworthy individuals in the hiring process. In a prediction game that is strategically equivalent to the trust game, the amount sent by first-movers was significantly smaller when the second-mover indicated a high interest in working in finance. These results suggest that the financial industry attracts less trustworthy individuals, which may contribute to the current lack of trust in its employees.
This paper investigates risk-taking in the liquid portfolios held by a large panel of Swedish twins. We document that the portfolio share invested in risky assets is an increasing and concave function of financial wealth, leading to different risk sensitivities across investors. Human capital, which we estimate directly from individual labor income, also drives risk-taking positively, while internal habit and expenditure commitments tend to reduce it. Our micro findings lend strong support to decreasing relative risk aversion and habit formation preferences. Furthermore, heterogeneous risk sensitivities across investors help reconcile individual preferences with representative-agent models.
We present a thought-provoking study of two monetary models: the cash-in-advance and the Lagos and Wright (2005) models. We report that the different approach to modeling money — reduced-form vs. explicit role — neither induces theoretical nor quantitative differences in results. Given conformity of preferences, technologies and shocks, both models reduce to one difference equation. The equations do not coincide only if price distortions are differentially imposed across models. To illustrate, when cash prices are equally distorted in both models equally large welfare costs of inflation are obtained in each model. Our insight is that if results differ, then this is due to differential assumptions about the pricing mechanism that governs cash transactions, not the explicit microfoundation of money.
Die Empfehlung des Corporate Governance-Kodex (Ziff. 5.4.2), „dem Aufsichtsrat soll eine nach seiner Einschätzung angemessene Anzahl unabhängiger Mitglieder angehören“, wirft in der Praxis nach wie vor Fragen auf. Im Folgenden sollen einige Thesen zur Auslegung dieser Empfehlung aufgestellt werden. Eine rechtspolitische Auseinandersetzung mit ihr und Änderungsvorschläge sind an dieser Stelle nicht beabsichtigt.
We assemble a data set of more than eight million German Twitter posts related to the war in Ukraine. Based on state-of-the-art methods of text analysis, we construct a daily index of uncertainty about the war as perceived by German Twitter. The approach also allows us to separate this index into uncertainty about sanctions against Russia, energy policy and other dimensions. We then estimate a VAR model with daily financial and macroeconomic data and identify an exogenous uncertainty shock. The increase in uncertainty has strong effects on financial markets and causes a significant decline in economic activity as well as an increase in expected inflation. We find the effects of uncertainty to be particularly strong in the first months of the war.
We analyze the repercussions of different kinds of uncertainty on cash demand, including uncertainty of the digital infrastructures, confidence crises of the financial system, natural disasters, political uncertainties, and inflationary crises. Based on a comprehensive literature survey, theoretical considerations and complemented by case studies, we derive a classification scheme how cash holdings typically evolve in each of these types of uncertainty by separating between demand for domestic and international cash as well as between transaction and store of value balances. Hereby, we focus on the stabilizing macroeconomic properties of cash and recommend guidelines for cash supply by central banks and the banking system. Finally, we exemplify our analysis with five case studies from the developing world, namely Venezuela, Zimbabwe, Afghanistan, Iraq, and Libya.
We find that high macroeconomic uncertainty is associated with greater accumulation of physical capital, despite a reduction in investment and valuations. To reconcile this puzzling evidence, we show that uncertainty predicts lower depreciation and utilization of existing capital, which dominates the investment slowdown. Motivated by these dynamics, we develop a quantitative production-based model in which firms implement precautionary savings through reducing utilization rather than raising invest-ment. Through this novel intensive-margin mechanism, uncertainty shocks command a quarter of the equity premium in general equilibrium, while flexibility in utilization adjustments helps explain uncertainty risk exposures in the cross-section of industry returns.
Unconventional green
(2023)
We analyze the effects of the PEPP (Pandemic Emergency Purchase Programme), the temporary quantitative easing implemented by the ECB immediately after the burst of the Covid-19 pandemic. We show that the differences in aim, size and flexibility with respect to the traditional Corporate Sector Purchase Programme (CSPP) were able to significantly involve, in addition to the directly targeted bonds, also the green bond segment. Via a standard difference- in-differences model we estimate that the yield on green bonds declined by more than 20 basis points after the PEPP. In order to take into account also the differences attributable to the eligibility to the programme, we employ a triple difference estimator. Bonds that at the same time were green and eligible benefitted of an additional premium of 39 basis points.
We assess the effect and the timing of the corporate arm of the ECB quantitative easing (CSPP) on corporate bond issuance. Because of several contemporaneous measures, to isolate the programme effects we rely on one key eligibility feature: the euro denomination of newly issued bonds. We find that the significant increase in bonds issuance by eligible firms is due to the CSPP and that this effect took at least six months to unfold. This result holds even when comparing firms with similar ratings, thus providing evidence that unconventional monetary policy can foster a financing diversification regardless of firms’ risk profile. We also highlight the impact of the programme on the real economic activity. The evidence suggests that while all firms increased investment in capital expenditures and intangible assets, the CSPP induced eligible firms to invest in marketable and equity securities, to repurchase their own stocks, to hold cash and to carry out short-term investment.
Effective market discipline incentivizes financial institutions to limit their risk-taking behavior, making it a key element for financial regulation. However, without adequate incentives to monitor and control the risk-taking behavior of financial institutions market discipline erodes. As a consequence, bailing out financial institutions, as happened unprecedentedly during the recent financial crisis, may impose indirect costs to financial stability if bailout expectations of investors change. Analyzing US data covering the period between 2004 and 2014, Hett und Schmidt (2017) find that market participants adjusted their bailout expectations in response to government interventions, undermining market discipline mechanisms. Given these findings, policymakers need to take into account the potential effects on market discipline when deciding about public support to troubled financial institutions in the future. Considering the parallelism of events and public responses during the financial crisis as well as the recent developments of Italian banks, these results not only concern the US, but also have important implications for European financial markets and policy makers.
Many consumers care about climate change and other externalities associated with their purchases. We analyze the behavior and market effects of such “socially responsible consumers” in three parts. First, we develop a flexible theoretical framework to study competitive equilibria with rational consequentialist consumers. In violation of price taking, equilibrium feedback non-trivially dampens a consumer’s mitigation efforts, undermining responsible behavior. This leads to a new type of market failure, where even consumers who fully “internalize the externality” overconsume externality-generating goods. At the same time, socially responsible consumers change the relative effectiveness of taxes, caps, and other policies in lowering the externality. Second, since consumer beliefs about and preferences over dampening play a crucial role in our framework, we investigate them empirically via a tailored survey. Consistent with our model, consumers are predominantly consequentialist, and on average believe in dampening. Inconsistent with our model, however, many consumers fail to anticipate dampening. Third, therefore, we analyze how such “naive” consumers modify our theoretical conclusions. Naive consumers behave more responsibly than rational consumers in a single-good economy, but may behave less responsibly in a multi-good economy with cross-market spillovers. A mix of naive and rational consumers may yield the worst outcomes.
Consider two independent random walks. By chance, there will be spells of association between them where the two processes move in the same direction, or in opposite direction. We compute the probabilities of the length of the longest spell of such random association for a given sample size, and discuss measures like mean and mode of the exact distributions. We observe that long spells (relative to small sample sizes) of random association occur frequently, which explains why nonsense correlation between short independent random walks is the rule rather than the exception. The exact figures are compared with approximations. Our finite sample analysis as well as the approximations rely on two older results popularized by Révész (Stat Pap 31:95–101, 1990, Statistical Papers). Moreover, we consider spells of association between correlated random walks. Approximate probabilities are compared with finite sample Monte Carlo results.
Some observers have conjectured that oil supply shocks in the United States and in other countries are behind the plunge in the price of oil since June 2014. Others have suggested that a major shock to oil price expectations occurred when in late November 2014 OPEC announced that it would maintain current production levels despite the steady increase in non-OPEC oil production. Both conjectures are perfectly reasonable ex ante, yet we provide quantitative evidence that neither explanation appears supported by the data. We show that more than half of the decline in the price of oil was predictable in real time as of June 2014 and therefore must have reflected the cumulative effects of earlier oil demand and supply shocks. Among the shocks that occurred after June 2014, the most influential shock resembles a negative shock to the demand for oil associated with a weakening economy in December 2014. In contrast, there is no evidence of any large positive oil supply shocks between June and December. We conclude that the difference in the evolution of the price of oil, which declined by 44% over this period, compared with other commodity prices, which on average only declined by about 5%-15%, reflects oil-market specific developments that took place prior to June 2014.
This paper examines the advantages and drawbacks of alternative methods of estimating oil supply and oil demand elasticities and of incorporating this information into structural VAR models. I not only summarize the state of the literature, but also draw attention to a number of econometric problems that have been overlooked in this literature. Once these problems are recognized, seemingly conflicting conclusions in the recent literature can be resolved. My analysis reaffirms the conclusion that the one-month oil supply elasticity is close to zero, which implies that oil demand shocks are the dominant driver of the real price of oil. The focus of this paper is not only on correcting some misunderstandings in the recent literature, but on the substantive and methodological insights generated by this exchange, which are of broader interest to applied researchers.
Understanding the shift from micro to macro-prudential thinking: a discursive network analysis
(2016)
While some economists argued for macro-prudential regulation pre-crisis, the macro-prudential approach and its emphasis on endogenously created systemic risk have only gained prominence post-crisis. Employing discourse and network analysis on samples of the most cited scholarly works on banking regulation as well as on systemic risk (60 sources each) from 1985 to 2014, we analyze the shift from micro to macro-prudential thinking in the shift to the post crisis period. Our analysis demonstrates that the predominance of formalism, particularly, partial equilibrium analysis along with the exclusion of historical and practitioners’ styles of reasoning from banking regulatory studies impeded economists from engaging seriously with the endogenous sources of systemic risk prior to the crisis. Post-crisis, these topics became important in this discourse, but the epistemological failures of banking regulatory studies pre-crisis were not sufficiently recognized. Recent attempts to conceptualize and price systemic risk as a negative externality point to the persistence of formalism and equilibrium thinking, with its attending dangers of incremental innovation due to epistemological barriers constrains theoretical progress, by excluding observed phenomena, which cannot yet be accommodated in mathematical models.
Since the 1970s, the overarching view in the literature has been that a Phillips curve relationship did not exist in Ireland prior to the 1979 exchange rate break with Sterling. It was argued that, as a small open economy, prices were determined externally. To test this relationship, we study the determination of inflation between 1926 and 2012, a longer sample period than any previously used. We find that the difference between unemployment and the NAIRU is a significant determinant of inflation both in the full sample and in the subsamples spanning the periods before and after the Sterling parity link.
This paper analyzes empirically the distribution of unemployment durations in West- Germany before and after the changes during the mid 1980s in the maximum entitlement periods for unemployment benefits for elderly unemployed. The analysis is based on the comprehensive IAB employment subsample containing register panel data for about 500.000 individuals in West Germany. We analyze two proxies for unemployment since the data do not precisely measure unemployment in an economic sense. We provide a theoretical analysis of the link between the durations of nonemployment and of unemployment between jobs. Our empirical analysis finds significant changes in the distributions of nonemployment durations for older unemployed individuals. At the same time, the distribution of unemployment durations between jobs did not change in response to the reforms. Our findings are consistent with an interpretation that many firms and workers used the more bene cial laws as a part of early retirement packages but those workers who were still looking for a job did not reduce their search effort in response to the extension of the maximum entitlement periods. This interpretation is consistent with our theoretical model under plausible assumptions. JEL: C24, J64, J65
Abstract: It is commonplace in the debate on Germany's labor market problems to argue that high unemployment and low wage dispersion are related. This paper analyses the relationship between unemployment and residual wage dispersion for individuals with comparable attributes. In the conventional neoclassical point of view, wages are determined by the marginal product of the workers. Accordingly, increases in union minimum wages result in a decline of residual wage dispersion and higher unemployment. A competing view regards wage dispersion as the outcome of search frictions and the associated monopsony power of the firms. Accordingly, an increase in search frictions causes both higher unemployment and higher wage dispersion. The empirical analysis attempts to discriminate between the two hypotheses for West Germany analyzing the relationship between wage dispersion and both the level of unemployment as well as the transition rates between different labor market states. The findings are not completely consistent with either theory. However, as predicted by search theory, one robust result is that unemployment by cells is not negatively correlated with the within cell wage dispersion.
We investigate the impact of uneven transparency regulation across countries and industries on the location of economic activity. Using two distinct sources of regulatory variation—the varying extent of financial-reporting requirements and the staggered introduction of electronic business registers in Europe—, we consistently document that direct exposure to transparency regulation is negatively associated with the focal industry’s economic activity in terms of inputs (e.g., employment) and outputs (e.g., production). By contrast, we find that indirect exposure to supplier and customer industries’ transparency regulation is positively associated with the focal industry’s economic activity. Our evidence suggests uneven transparency regulation can reallocate economic activity from regulated toward unregulated countries and industries, distorting the location of economic activity.
We investigate how unconventional monetary policy, via central banks’ purchases of corporate bonds, unfolds in credit-saturated markets. While this policy results in a loosening of credit market conditions as intended by policymakers, we report two unintended side effects. First, the policy impacts the allocation of credit among industries. Affected banks reallocate loans from investment-grade firms active on bond markets almost entirely to real estate asset managers. Other industries do not obtain more loans, particularly real estate developers and construction firms. We document an increase in real estate prices due to this policy, which fuels real estate overvaluation. Second, more loan write-offs arise from lending to these firms, and banks are not compensated for this risk by higher interest rates. We document a drop in bank profitability and, at the same time, a higher reliance on real estate collateral. Our findings suggest that central banks’ quantitative easing has substantial adverse effects in credit-saturated economies.
Using granular supervisory data from Germany, we investigate the impact of unconventional monetary policies via central banks’ purchase of corporate bonds. While this policy results in a loosening of credit market conditions as intended by policy makers, we document two unintended side effects. First, banks that are more exposed to borrowers benefiting from the bond purchases now lend more to high-risk firms with no access to bond markets. Since more loan write-offs arise from these firms and banks are not compensated for this risk by higher interest rates, we document a drop in bank profitability. Second, the policy impacts the allocation of loans among industries. Affected banks reallocate loans from investment grade firms active on bond markets to mainly real estate firms without investment grade rating. Overall, our findings suggest that central banks’ quantitative easing via the corporate bond markets has the potential to contribute to both banking sector instability and real estate bubbles.
This paper analyzes the relationship between monetary policy and financial stability in the Banking Union. There is no uniform global model regarding the relationship between monetary policy-making on the one hand, and prudential supervision on the other. Before the crisis, EU Member States followed different approaches, some of them uniting monetary and supervisory functions in one institution, others assigning them to different, neatly separated institutions. The financial crisis has underlined that monetary policy and prudential supervision deeply affect each other, especially in case of systemic events. Even in normal times, monetary and supervisory decisions might conflict with each other. After the crisis, some jurisdictions have moved towards a more holistic approach under which monetary policy takes supervisory considerations into account, while supervisory decisions pay due regard to monetary policy.
The Banking Union puts prudential supervision in the hands of the European Central Bank (ECB), the institution responsible for monetary policy. Nevertheless, at its establishment there was the political understanding that the ECB should follow a policy of meticulous separation in the discharge of its different functions. This raises the question whether the ECB may pursue a holistic approach to monetary policy and supervisory decision-making, respectively. On the basis of a purposive reading of the monetary policy mandate and the SSM Regulation, the paper answers this question in the affirmative. Effective monetary policy (or supervision) requires financial stability (or smooth monetary policy transmission). Moreover, without a holistic approach, the SSM Regulation is more likely to provoke the adoption of mutually defeating decisions by the Governing Board. The reputation of the ECB would suffer considerably under such a situation – in a field where reputation is of paramount importance for effective policy.
As any meticulous separation between monetary and supervisory functions turns out to be infeasible, the paper explores the reasons. Parting from Katharina Pistor’s legal theory of finance, which puts the emphasis on exogenous factors to explain the (non)enforcement of legal rules, the paper suggests a legal instability theorem which focuses on endogenous reasons, such as law’s indeterminacy, contextuality, and responsiveness to democratic deliberation. This raises the question whether the holistic approach would be democratically legitimate under the current framework of the ESCB. The idea of technocratic legitimacy that exempts the ECB from representative structures is effectively called into question by the legal instability theorem. This does not imply that the independence of the ECB should be given up, as there are no viable alternatives to protect monetary policy against the time inconsistency problem. Rather, any solution might benefit from recognizing the ECB in its mixed technocratic and political shape as a centerpiece of European integration and improving.
Questionable research practices have generated considerable recent interest throughout and beyond the scientific community. We subsume such practices involving secret data snooping that influences subsequent statistical inference under the term MESSing (manipulating evidence subject to snooping) and discuss, illustrate and quantify the possibly dramatic effects of several forms of MESSing using an empirical and a simple theoretical example. The empirical example uses numbers from the most popular German lottery, which seem to suggest that 13 is an unlucky number.
In this study, we unpack the ESG ratings of four prominent agencies in Europe and find that (i) each single E, S, G pillar explains the overall ESG score differently,(ii) there is a low co-movement between the three E, S, G pillars and (iii) there are specific ESG Key Performance Indicators (KPIs) that are driving these ratings more than others. We argue that such discrepancies might mislead firms about their actual ESG status, potentially leading to cherry-picking areas for improvement, thus raising questions about the accuracy and effectiveness of ESG evaluations in both explaining sustainability and driving capital toward sustainable companies.
Unternehmen und Ethik
(2010)
Die noch nicht völlig überwundene Finanzmarktkrise hat nicht nur den Gesetzgeber auf den Plan gerufen. Auch die Frage nach der Ethik der Akteure wird vielfach erörtert. Haben von Gier getriebene Finanzmarktakteure nicht nur Rechtsregeln, sondern auch ethische Normen, die Grundsätze des Wirtschaftens ehrbarer Kaufleute, gewissenlos beiseite geschoben, um sich zu bereichern? Wie läßt sich die Beachtung dieser Normen künftig sichern? Diese aktuelle Debatte soll Anlaß zu einigen allgemeineren Betrachtungen zum Thema „Unternehmen und Ethik“ sein.
Using a nonlinear Bayesian likelihood approach that fully accounts for the zero lower bound on nominal interest rates, the authors analyze US post-crisis business cycle dynamics and provide reference parameter estimates. They find that neither the inclusion of financial frictions nor that of household heterogeneity improve the empirical fit of the standard model, or its ability to provide a joint explanation for the post-2007 dynamics. Associated financial shocks mis-predict an increase in consumption. The common practice of omitting the ZLB period in the estimation severely distorts the analysis of the more recent economic dynamics.
Use banks the right way
(2020)
This paper compares the accuracy of credit ratings of Moody s and Standard&Poors. Based on 11,428 issuer ratings and 350 defaults in several datasets from 1999 to 2003 a slight advantage for the rating system of Moody s is detected. Compared to former research the robustness of the results is increased by using nonparametric bootstrap approaches. Furthermore, robustness checks are made to control for the impact of Watchlist entries, staleness of ratings and the effect of unsolicited ratings on the results.
When parties present divergent econometric evidence, the court may view such evidence as contradictory and thus ignore it completely, without conducting closer analysis. We develop a simple method for distinguishing between actual and merely apparent contradiction based on the statistical concept of the “severity” of the furnished evidence. Again using “severity”, we also propose a method for reconciling divergent findings in instances of mere seeming contradiction. Our chosen application is that of damage estimation in follow-on cases.
Im Rahmen des Value Based Managements sollen alle Manahmen der Unternehmensfhrung auf die Steigerung des intrinsischen Marktwertes des Eigenkapitals ausgerichtet werden. Hauptanwendungsbereiche des Value Based Managements sind die Planung, Performancemessung, Managemententlohnung sowie die Bereichssteuerung. Eine konsequente shareholder-orientierte Steuerung beinhaltet eine adquate Ermittlung von Wertbeitrgen in den einzelnen Anwendungsbereichen. Hierzu dienen insbesondere absolute Wertbeitragskennzahlen in Form von Residualgewinnen. Residualgewinne sind dadurch gekennzeichnet, dass der Gewinn um eine auf das Gesamtkapital bezogene Capital bzw. Interest Charge reduziert wird. Der magebliche Gewinn sowie das Gesamtkapital ergeben sich, indem die handelsrechtlichen Gren durch mehr oder minder weitreichende Modifikationen angepasst werden. Die Berechnung der Interest Charge erfolgt entweder auf Basis eines risikoangepassten Zinssatzes oder auf Basis eines risikofreien Zinssatzes. In der Praxis finden sich zahlreiche unterschiedliche Residualgewinnkonzepte, die sich insbesondere durch die jeweils charakteristischen Modifikationen und durch den verwendeten Zinssatz zur Berechnung der Kapitalkosten unterscheiden.1) Herkmmliche Konzepte zeichnen sich dadurch aus, dass die Interest Charge auf Basis eines risikoangepassten Zinssatzes berechnet wird. Das in Velthuis (2003c) theoretisch entworfene und von der KPMG in der Praxis umgesetzte Konzept Earnings less Riskfree Interest Charge, ERIC , ist hingegen dadurch charakterisiert, dass zur Berechnung der Interest Charge ein risikofreier Zinssatz verwendet wird. Ferner werden Modifikationen nur insofern vorgenommen, als dass eine vollstndige Bercksichtigung aller Erfolgskomponenten gewhrleistet wird. Glaubt man den Vertretern herkmmlicher Konzepte, basieren diese Konzepte auf einer hinreichenden theoretischen Fundierung. Diesem Anspruch gengen solche Konzepte tatschlich jedoch nicht, sie zeichnen sich gerade durch eine mangelnde theoretische Fundierung aus.2) In Velthuis (2003c) wird verdeutlicht, dass die in der Praxis verbreitete Vorgehensweise gerade zum Ausweis von Wertbeitrgen fhrt, die inkonsistent mit der Zielsetzung der Shareholder sind. In allen Anwendungsbereichen des VBM zeigt sich nmlich aus theoretischer Sicht die Problematik der VerIm Rahmen des Value Based Managements sollen alle Maßnahmen der Unternehmensführung auf die Steigerung des intrinsischen Marktwertes des Eigenkapitals ausgerichtet werden. Hauptanwendungsbereiche des Value Based Managements sind die Planung, Performancemessung, Managemententlohnung sowie die Bereichssteuerung. Eine konsequente shareholder-orientierte Steuerung beinhaltet eine adäquate Ermittlung von Wertbeiträgen in den einzelnen Anwendungsbereichen. Hierzu dienen insbesondere absolute Wertbeitragskennzahlen in Form von Residualgewinnen. Residualgewinne sind dadurch gekennzeichnet, dass der Gewinn um eine auf das Gesamtkapital bezogene Capital bzw. Interest Charge reduziert wird. Der maßgebliche Gewinn sowie das Gesamtkapital ergeben sich, indem die handelsrechtlichen Größen durch mehr oder minder weitreichende Modifikationen angepasst werden. Die Berechnung der Interest Charge erfolgt entweder auf Basis eines risikoangepassten Zinssatzes oder auf Basis eines risikofreien Zinssatzes. In der Praxis finden sich zahlreiche unterschiedliche Residualgewinnkonzepte, die sich insbesondere durch die jeweils charakteristischen Modifikationen und durch den verwendeten Zinssatz zur Berechnung der Kapitalkosten unterscheiden.1) Herkömmliche Konzepte zeichnen sich dadurch aus, dass die Interest Charge auf Basis eines risikoangepassten Zinssatzes berechnet wird. Das in Velthuis (2003c) theoretisch entworfene und von der KPMG in der Praxis umgesetzte Konzept Earnings less Riskfree Interest Charge, ERIC Ò , ist hingegen dadurch charakterisiert, dass zur Berechnung der Interest Charge ein risikofreier Zinssatz verwendet wird. Ferner werden Modifikationen nur insofern vorgenommen, als dass eine vollständige Berücksichtigung aller Erfolgskomponenten gewährleistet wird. Glaubt man den Vertretern herkömmlicher Konzepte, basieren diese Konzepte auf einer hinreichenden theoretischen Fundierung. Diesem Anspruch genügen solche Konzepte tatsächlich jedoch nicht, sie zeichnen sich gerade durch eine mangelnde theoretische Fundierung aus.2) In Velthuis (2003c) wird verdeutlicht, dass die in der Praxis verbreitete Vorgehensweise gerade zum Ausweis von Wertbeiträgen führt, die inkonsistent mit der Zielsetzung der Shareholder sind. In allen Anwendungsbereichen des VBM zeigt sich nämlich aus theoretischer Sicht die Problematik der Verwendung eines risikoangepassten Zinssatzes zur Berechnung von Kapitalkosten: Schon die Verwendung eines risikoangepassten Kapitalkostensatzes im Rahmen der Planung kann problematisch sein. Die Verwendung risikoangepasster Kapitalkosten als Benchmark bei der Performancemessung ist sogar ganz abzulehnen, und bei der Managemententlohnung ist die Verrechnung von risikoangepassten Kapitalkosten zwingend mit Fehlanreizen verbunden. Aber auch praxisübliche Modifikationen erweisen sich als problematisch, da Erfolgskomponenten nicht bzw. nicht vollständig berücksichtigt werden. Im Gegensatz zu herkömmlichen Konzepten weist das VBM-Konzept ERIC eine tatsächliche theoretische Fundierung auf. In Velthuis (2003c) werden grundlegende theoretische und praktische Anforderungen für die einzelnen Anwendungsbereiche herausgearbeitet und es wird gezeigt, dass ein integriertes VBM-Konzept auf Basis des Residualgewinns ERIC diese Anforderungen stets erfüllt. Ziel dieses Aufsatzes ist es, die Grundkonzeption von ERIC theoretisch weiter zu entwickeln. Während in Velthuis (2003c) stets von einer sehr einfachen Idealwelt ausgegangen wird, soll hier zum einen verdeutlicht werden, wie ein erfolgreiches Value Based Management auf Basis von ERIC in komplexeren und realitätsnäheren Situationen erfolgen kann. Zum anderen gilt es, zusätzliche praktische Problembereiche zu erkennen und Lösungsmöglichkeiten aufzuzeigen. Der Aufsatz ist wie folgt aufgebaut: Zunächst wird gezeigt, wie der Residualgewinn ERIC zu berechnen ist, und es wird ein Überblick über die Anwendungsbereiche gegeben. Anschließend wird die grundlegende Entscheidungssituation für die späteren Darstellungen erörtert und es werden fundamentale Prinzipien des Value Based Managements und ihre Bedeutung für die einzelnen Anwendungsbereichen erläutert. Sodann wird für die einzelnen Bereichen des Value Based Managements, der Planung, Performancemessung, Managemententlohnung und der Bereichssteuerung gezeigt, wie ERIC anzuwenden ist, damit eine tatsächliche Shareholder-Orientierung gewährleistet wird. Die Darstellungen erfolgen dabei jeweils zunächst vor dem Hintergrund einer Idealwelt. In einem zweiten Schritt wird die Entscheidungssituation jeweils modifiziert, um der Realität ein Stück näher zu kommen.
This paper addresses the need for transparent sustainability disclosure in the European Auto Asset-Backed Securities (ABS) market, a crucial element in achieving the EU's climate goals. It proposes the use of existing vehicle identifiers, the Type Approval Number (TAN) and the Type-Variant-Version Code (TVV), to integrate loan-level data with sustainability-related vehicle information from ancillary sources. While acknowledging certain challenges, the combined use of TAN and TVV is the optimal solution to allow all stakeholders to comprehensively assess the environmental characteristics of securitised exposure pools in terms of data protection, matching accuracy, and cost-effectiveness.
Venture capital-backed firms, unavoidable value-destroying trade sales, and fair value protections
(2020)
This paper investigates the implications of the fair value protections contemplated by the standard corporate contract (i.e., the standard contract form for which corporate law provides) for the entrepreneur–venture capitalist relationship, focusing, in particular, on unavoidable value-destroying trade sales. First, it demonstrates that the typical entrepreneur–venture capitalist contract does institutionalize the venture capitalist’s liquidity needs, allowing, under some circumstances, for counterintuitive instances of contractually-compliant value destruction. Unavoidable value-destroying
trade sales are the most tangible example. Next, it argues that fair value protections can prevent the entrepreneur and venture capitalist from allocating the value that these transactions generate as they would want. Then, it shows that the reality of venture capital-backed firms calls for a process of adaptation of the standard corporate contract that has one major step in the deactivation or re-shaping of fair value protections. Finally, it argues that a standard corporate contract aiming to promote social welfare through venture capital should feature flexible fair value protections
Ziel dieser Präsentation (anlässlich des Seminars „Die Auswirkungen von Asset Securitisation auf die Stabilität des Finanzmarktes“ Österreichische Nationalbank (ÖNB), Wien 1. Oktober 2003) ist es, eine Verbindung zwischen Verbriefung und Finanzmarkstabilität unter Berücksichtigung veränderter Finanzintermediation herzustellen. In der folgenden Abhandlung soll nun zunächst auf die Natur der Verbriefung per se eingegangen werden, um sodann anhand theoretischer Überlegungen und empirischer Beobachtungen mögliche Quelle systemischen Risikos in der Kreditverbriefung aufzuzeigen. In diesem Fall handelt es sich um die Informationsasymmetrien und die durch Handelbarkeit von Kreditrisiko bestimmte Transaktionsstruktur („security design“), die bei regulatorischer Nichtberücksichtigung eine destabilisierende Wirkung nicht nur im Bereich der Verbriefung, sondern auch hinsichtlich der gegenseitigen Zahlungsverpflichtungen von Finanzintermediären begründen könnte.
Es gehört zu den Mindestansprüchen des Sozialstaats, bedürftigen Personen ein soziokulturelles Existenzminimum zu sichern. In Deutschland sollen gegenwärtig insbesondere die Leistungen der Sozialhilfe und der Grundsicherung für Arbeitsuchende dafür sorgen, dass dieses Minimalziel erreicht wird. Ob dies gelingt, ist jedoch fraglich, da erhebliche Teile der Bedürftigen ihnen zustehende Leistungen nicht in Anspruch nehmen. Die Statistiken über Leistungsempfänger/innen geben also nur "die halbe Wahrheit" über Bedürftigkeit in Deutschland wieder. Vor der Hartz-IV-Reform kamen auf drei Grundsicherungsempfänger/innen mindestens zwei, eher drei weitere Berechtigte, die von ihrem Anspruch keinen Gebrauch machten. Dabei waren einige gesellschaftliche Gruppen von verdeckter Armut besonders stark betroffen: alleinstehende Frauen, Paarhaushalte mit erwerbstätigem Haushaltsvorstand und Altenhaushalte. Die Einführung des Arbeitslosengeldes (Alg) II im Zuge der Hartz-IV-Reform konnte zwar die verdeckte Armut von früheren Arbeitslosenhilfebeziehenden vermindern. Doch andere Bedürftige, insbesondere Erwerbstätigenhaushalte mit geringem Einkommen, nehmen offensichtlich ihnen zustehende Leistungen nach wie vor häufig nicht in Anspruch. Die Gründe für das Phänomen der verdeckten Armut sind vielfältig. Wesentliche Ursachen sind offenbar, dass viele Bedürftige die relevanten gesetzlichen Regelungen nicht kennen oder mit der Inanspruchnahme staatlicher Hilfeleistungen Stigmatisierungsängste verbinden. Dem könnte entgegengewirkt werden, indem offensiv über bestehende Ansprüche informiert und ein vorurteils- und diskriminierungsfreies Klima im Umgang mit Hilfebedürftigen gefördert wird.
Das Working Paper bietet die zusammenfassende Stellungnahme von Prof. Volker Wieland zum Ankaufprogramm der Europäischen Zentralbank für Anleihen des öffentlichen Sektors (Public Sector Purchase Programme, PSPP) am Bundesverfassungsgericht am 30.07.2019. Dabei liegt der Schwerpunkt auf der Frage der Einordnung des PSPP als monetäre, geldpolitische Maßnahme und der Verhältnismäßigkeit des Programms und seiner Umsetzung. Ebenfalls wird kurz auf die weiteren Fragen zur Umsetzung, insbesondere Ankündigung, Begrenzung und Abstand zum Primärmarkt für Staatsanleihen eingegangen.
Vergebene Chance
(2017)
In den letzten Jahren wurden verschiedene Modelle entwickelt, um das Ausfallrisiko von Banken unter Berücksichtigung von Portfolioeffekten zu quantifizieren. Bisher hat sich kein Ansatz als allgemein akzeptierter Standard durchsetzen können. Da die Modelle grundlegende konzeptionelle Unterschiede aufweisen und unterschiedliche empirische Inputdaten verwenden, hat die Auswahl eines Kreditrisikomodells unter Umständen einen erheblichen Einfluß auf die Kreditportfoliosteuerung der Bank. In diesem Beitrag soll deshalb geklärt werden, ob die Modelle systematisch abweichende Value-at-Risk-Werte berechnen und worin die Ursachen für auftretende Abweichungen liegen. Zunächst wird gezeigt, daß die bestehenden Kreditrisikomodelle in zwei grundlegende Klassen eingeteilt werden können: Assetwert-basierte Modelle und auf Ausfallraten basierende Modelle. Am Beispiel eines Musterportfolios von Krediten an deutsche Baufirmen werden zwei Vertreter der Modellklassen (CreditMetrics und CreditRisk+) verglichen und der Effekt der unterschiedlichen empirischen Inputparameter auf die Risikoergebnisse abgeschätzt. Die Analyse zeigt erhebliche Unterschiede zwischen den Modellen. Eine Analyse der Abweichungsursachen ergibt allerdings, daß der Grund für die großen Value-at-Risk-Unterschiede in erster Linie in den unterschiedlichen empirischen Inputdaten liegt, welche zu unterschiedlichen impliziten Korrelationsannahmen führen. Es wird gezeigt, wie die Modellparameter gewählt werden müssen, um identische Korrelationen zu erzeugen. Bei konsistenten Korrelationsannahmen stimmen die Ergebnisse beider Modelle weitgehend überein. Keywords: Credit Risk Models, credit risk correlation
Die vorliegende Analyse untersucht die Beschäftigungseffekte von Vermittlungsgutscheinen und Personal-Service-Agenturen mit Hilfe einer makroökonometrischen Evaluation. Neben einer mikroökonometrischen Evaluation, welche die Wirkungen auf individueller Ebene untersucht, kann eine makroökonometrische Analyse Aussagen über die gesamtwirtschaftlichen Effekte der Maßnahmen machen. Die strukturellen Multiplikatorwirkungen im makroökonomischen Kreislaufzusammenhang werden jedoch nicht berücksichtigt. Das ökonometrische Modell zur Analyse der beiden Maßnahmen basiert auf einer Matching-Funktion, die den Suchprozess von Firmen und von Arbeitern nach einem Beschäftigungsverhältnis abbildet. Die empirischen Analysen werden getrennt für Ost- und Westdeutschland sowie für die Strategietypen der Bundesagentur für Arbeit durchgeführt. Sie zeigen, dass die Ausgabe von Vermittlungsgutscheinen nur in „großstädtisch geprägten Bezirken vorwiegend in Westdeutschland mit hoher Arbeitslosigkeit“ (Strategietyp II) einen signifikant positiven Effekt auf den Suchprozess hat. Für die Personal-Service-Agenturen zeigen sich signifikant positive Effekte für Ost- als auch für Westdeutschland. Allerdings fehlt für eine abschließende Bewertung der Ergebnisse für die Personal- Service-Agenturen aufgrund der relativ geringen Teilnehmerzahl noch ein Vergleich mit mikroökonometrischen Analysen.
Unter Berücksichtigung der implementationsanalytischen und kausalen Analysen für die Einführungsphase der Vermittlungsgutscheine kommen die Autoren zu dem Schluss, dass die Erprobungsphase dieses arbeitsmarktpolitischen Instruments fortgesetzt werden sollte. Aus der Implementationsanalyse zeigt sich, dass der Verbreitungsgrad des Instruments auch nach 27 Monaten Erprobungsphase sehr gering ist. Die kausalen Analysen beziehen sich aufgrund begrenzter Datenverfügbarkeit lediglich auf zwei Ausgabemonate ein Jahr nach Einführung des Instruments (Mai und Juni 2003) und zeigen geringe positive Beschäftigungseffekte auf der Mikro-Ebene. Ob dadurch die Kosten der Vermittlungsgutscheine gerechtfertigt sind, lässt sich derzeit noch nicht abschließend beurteilen. Anzeichen für Mitnahmeeffekte und/oder Missbrauch existieren. Auch haben erfolgreiche Vermittlungen durch Gutscheine die Beschäftigungschancen in anderen Gruppen geschmälert. Deshalb werden verschiedene Vorschläge für eine kosteneffizientere Ausgestaltung der Vermittlungsgutscheine dargestellt und diskutiert. Insbesondere wird auf die Vorschläge des Bundeskabinetts vom 01.09.2004 eingegangen.
Versicherungsunternehmen haben bei der Auswahl ihrer Vermögensanlagen die gesetzlichen Restriktionen des Versicherungsaufsichtsgesetzes einzuhalten. Neben einer strukturierten Darstellung der zahlreichen Regulierungstatbestände werden aus Sicht der Finanzierungstheorie sowie den empirischen Verhältnissen an den Kapitalmärkten die im VAG enthaltenen Rahmenbedingungen einer kritischen Bewertung unterzogen.
Differenzierte Verdienststrukturen, wie zum Beispiel qualifikatorische Verdienstunterschiede, tragen der Heterogenität am Arbeitsmarkt und somit auch einer notwendigen Flexibilität der Löhne Rechnung. Trotz dieser bedeutenden Rolle der Lohnstruktur für den Arbeitsmarkt gibt es relativ wenig detaillierte empirische Evidenz zu deren Ausgestaltung und Dynamik für Westdeutschland. In dieser Arbeit werden auf Basis der Gehalts- und Lohnstrukturerhebungen für die Jahre 1990 und 1995 sowohl die Lohndispersion innerhalb bestimmter Erwerbsgruppen („Within-Dispersion“) als auch zwischen diesen Gruppen („Between-Dispersion“) untersucht. Im Fokus der Analysen steht dabei der Vergleich zwischen den Erwerbsgruppen der vollzeitbeschäftigten Männer, der vollzeitbeschäftigten Frauen und der teilzeitbeschäftigten Frauen nach den Qualifikationsniveaus: ohne Berufsabschluss, mit Berufsabschluss und mit Hochschulabschluss. Dichteschätzungen des Bruttotageslohns zeigen, dass vollzeiterwerbstätige Männer über die gesamte Verteilung den höchsten Verdienst aufweisen. Weiterhin deuten die Dichteschätzungen darauf hin, dass die geschlechtsspezifischen Lohndifferentiale mit zunehmendem Verdienst steigen. Der selbe Sachverhalt findet sich auch für die qualifikatorischen Verdienstunterschiede. Mit steigendem Qualifikationsniveau steigt sowohl der Lohn als auch die Dispersion. Besonders ausgeprägt ist die qualifikatorische Dispersion für Männer. Eine Analyse der Lohndifferentiale sowohl innerhalb der Erwerbsgruppen als auch zwischen Männern und Frauen weist darauf hin, dass in der Gruppe der teilzeiterwerbstätigen Frauen die größte Lohndispersion vorzufinden ist und dass diese im Zeitablauf zugenommen hat. Ferner sind die Lohndifferentiale im unteren Bereich der Verteilung kleiner als im oberen. Die weitere Disaggregation nach dem Qualifikationsniveau reduziert die Within-Dispersion, wobei die Lohnunterschiede innerhalb der Erwerbsgruppen mit steigendem Qualifikationsniveau tendenziell zunehmen. Schließlich nimmt die Within-Dispersion für vollzeitbeschäftigte Hochschulabsolventinnen im Zeitablauf zu, während sie für die Teilzeitbeschäftigten sinkt. Betrachtet man die Qualifikations- und die Verteilungsdimension, zeigt sich, dass in beiden Jahren lediglich für vollzeiterwerbstätige Frauen ohne Berufsabschluss bzw. mit Hochschulabschluss das geschlechtsspezifische Differential über die Verteilung hinweg steigt. Für alle anderen Gruppen geht es über die Verteilung hinweg zurück. Die Lohndifferentiale zwischen Männern und Frauen haben im Zeitablauf abgenommen. Das mittlere Reallohnwachstum zwischen 1990 und 1995 ist für teilzeiterwerbstätige Frauen mit 6,8% am höchsten und für vollzeiterwerbstätige Männer mit 0,7% am geringsten. Die Reduktion der geschlechtsspezifischen Unterschiede im Zeitablauf ist unter Berücksichtigung des Qualifikationsniveaus noch ausgeprägter. Dies gilt vor allem für teilzeiterwerbstätige Frauen ohne Berufsabschluss und mit Hochschulabschluss. Für teilzeiterwerbstätige Frauen und für vollzeiterwerbstätige Männer findet sich im Zeitablauf eine zunehmende Lohnspreizung, während für vollzeiterwerbstätige Frauen die Verteilung konstant bleibt. Eine Zunahme der Lohnspreizung ist dabei insbesondere bei Hochschulabsolventen festzustellen, sowohl für vollzeiterwerbstätige Männer wie Frauen. Im Unterschied dazu hat für diese Erwerbsgruppen die Lohndispersion bei den Mittelqualifizierten in den 90er Jahren abgenommen.
The implications of delegating fiscal decision making power to sub-national governments has become an area of significant interest over the past two decades, in the expectation that these reforms will lead to better and more efficient provision of public goods and services. The move towards decentralization has, however, not been homogeneously implemented on the revenue and expenditure side: decentralization has materialized more substantially on the latter than on the former, creating "vertical fiscal imbalances". These imbalances measure the extent to which sub-national governments’ expenditures are financed through their own revenues. This mismatch between own revenues and expenditures may have negative consequences for public finances performance, for example by softening the budget constraint of sub-national governments. Using a large sample of countries covering a long time period from the IMF’s Government Finance Statistics Yearbook, this paper is the first to examine the effects of vertical fiscal imbalances on fiscal performance through the accumulation of government debt. Our findings suggest that vertical fiscal imbalances are indeed relevant in explaining government debt accumulation, and call for a degree of caution when promoting fiscal decentralization.