Refine
Year of publication
- 2008 (100) (remove)
Document Type
- Working Paper (100) (remove)
Language
- English (100) (remove)
Is part of the Bibliography
- no (100)
Keywords
- USA (7)
- Bank (5)
- Geldpolitik (5)
- Lambda-Kalkül (5)
- Operationale Semantik (5)
- Programmiersprache (5)
- Haushalt (4)
- Liquidität (4)
- Aging (3)
- Deutschland (3)
In this paper, we examine the impact of mergers among German savings banks on the extent to which these savings banks engage in small business lending. The ongoing consolidation in the banking industry has sparked concerns about the continuous availability of credit to small businesses which has been further fueled by empirical studies that partly confirm a reduction in small business lending in the aftermath of mergers. However, using a proprietary data set of German savings banks we find strong evidence that in Germany merging savings banks do not significantly change the extent to which they lend to small businesses compared to prior to the merger or compared to the contemporaneous lending by non-merging banks. We investigate the merger related effects on small business lending in Germany from a bank-level perspective. Furthermore, we estimate small business lending and its continuous adjustment process simultaneously using recent General Method of Moments (GMM) techniques for panel data as proposed by Arellano and Bond (1991).
This paper discusses the so-called commercial approach to microfinance under economic and ethical aspects. It first shows how microfinance has developed from a purely welfare-oriented activity to a commercially relevant line of banking business. The background of this stunning success is the – almost universal – adoption of the so-called commercial approach to microfinance in the course of the last decade. As the author argues, this commercial approach is the only sound approach to adopt if one wanted microfinance to have any social and developmental impact, and therefore the wide-spread “moralistic” criticism of the commercial approach, which has again and again been expressed in the 1990s, is ill-placed from an economic and an ethical perspective. However, some recent events in microfinance raise doubts as to whether the commercial approach has not, in a number of cases, gone too far. The evident example for such a development is the Mexican microfinance institution Compartamos, which recently undertook a financially extremely successful IPO. As it seems, some microfinance institutions have by now become so radically commercial that all of those social and development considerations, which have traditionally motivated work in the field of microfinance, seem to have lost their importance. Thus there is a conflict between commercial and developmental aspirations. However, this conflict is not inevitable. The paper concludes by showing that, and how, a microfinance institution can try to combine using the strengths of the capital market and at the same time maintaining its developmental focus and importance.
We examine insurance markets with two types of customers: those who regret suboptimal decisions and those who don.t. In this setting, we characterize the equilibria under hidden information about the type of customers and hidden action. We show that both pooling and separating equilibria can exist. Furthermore, there exist separating equilibria that predict a positive correlation between the amount of insurance coverage and risk type, as in the standard economic models of adverse selection, but there also exist separating equilibria that predict a negative correlation between the amount of insurance coverage and risk type, i.e. advantageous selection. Since optimal choice of regretful customers depends on foregone alternatives, any equilibrium includes a contract which is o¤ered but not purchased.
When a spot market monopolist participates in a derivatives market, she has an incentive to deviate from the spot market monopoly optimum to make her derivatives market position more profitable. When contracts can only be written contingent on the spot price, a risk-averse monopolist chooses to participate in the derivatives market to hedge her risk, and she reduces expected profits by doing so. However, eliminating all risk is impossible. These results are independent of the shape of the demand function, the distribution of demand shocks, the nature of preferences or the set of derivatives contracts.
Since independence from British colonial rule, Uganda has had a turbulent political history characterised by putsches, dictatorship, contested electoral outcomes, civil wars and a military invasion. There were eight changes of government within a period of twenty-four years (from 1962-1986), five of which were violent and unconstitutional. This paper identifies factors that account for these recurrent episodes of political violence and state collapse. While colonialism bequeathed the country a negative legacy including a weak state apparatus, ethnic division, skewed development, elite polarisation and a narrow economic base, post-colonial leaders have on the whole exacerbated rather than reversed these trends. Factors such as ethnic rivalry, political exclusion, militarisation of politics, weak state institutions, and unequal access to opportunities for self-advancement help to account for the recurrent cycles of violence and state failure prior to 1986. External factors have also been important, particularly the country’s politically turbulent neighbourhood, the outcome of political instability and civil conflict in surrounding countries. Neighbourhood turbulence stemming from such factors as civil wars in Congo and Sudan has had spill-over effects in that it has allowed insurgent groups geographical space within which to operate as well as provided opportunities for the acquisition of instruments of war with which to destabilise the country. Critical to these processes have been the porosity of post-colonial borders and the inability by the Ugandan state to exercise effective control over its entire territory. By demonstrating the interplay between internal and external factors in shaping Uganda’s postcolonial experience, the paper makes an important shift away from conventional explanations that have focused disproportionately on internal processes. Lastly, the paper provides pointers to areas of further research such as the economic foundations of conflict that should ultimately strengthen our understanding of factors that combine to make state-making fail or succeed.
This paper is one of the first to analyse political influence on state-owned savings banks in a developed country with an established financial market: Germany. Combining a large dataset with financial and operating figures of all 457 German savings banks from 1994 to 2006 and information on over 1,250 local elections during this period we investigate the change in business behavior around elections. We find strong indications for political inflence: the probability that savings banks close branches, lay-off employees or engage in merger activities is significantly reduced around elections. At the same time they tend to increase their extraordinary spendings, which include support for social and cultural events in the area, on average by over 15%. Finally, we find that savings banks extend significantly more loans to their corporate and private customers in the run-up to an election. In further analyses, we show that the magnitude of political influence depends on bank specific, economical and political circumstances in the city or county: political influence seems to be facilitated by weak political majorities and profitable banks. Banks in economically weak areas seem to be less prone to political influence.
In this paper, we investigate how bank mergers affect bank revenues and present empirical evidence that mergers among banks have a substantial and persistent negative impact on merging banks’ revenues. We refer to merger related negative effects on banks’ revenues as dissynergies and suggest that they are a result of organizational diseconomies, the loss of customers and the temporary distraction of management from day-to-day operations by effecting the merger. For our analyses we draw on a proprietary data set with detailed financials of all 457 regional savings banks in Germany, which have been involved in 212 mergers between 1994 and 2006. We find that the negative impact of a merger on net operating revenues amounts to 3% of pro-forma consolidated banks’ operating profits and persists not only for the year of the merger but for up to four years post-merger. Only thereafter mergers exhibit a significantly superior performance compared to their respective pre-merger performance or the performance of their non-merging peers. The magnitude and persistence of merger related revenue dissynergies highlight their economic relevance. Previous research on post-merger performance mainly focuses on the effects from mergers on banks’ (cost) efficiency and profitability but fails to provide clear and consistent results. We are the first, to our knowledge, to examine the post-merger performance of banks’ net operating revenues and to empirically verify significant negative implications of mergers for banks’ net operating revenues. We propose that our finding of negative merger related effects on banks’ operating revenues is the reason why previous research fails to show merger related gains.
The "quiet life hypothesis (QLH)" posits that banks enjoy the advantages of market power in terms of foregone revenues or cost savings. We suggest a unified approach to measure competition and efficiency simultaneously to test this hypothesis. We estimate bank-specific Lerner indices as measures of competition and test if cost and profit efficiency are negatively related to market power in the case of German savings banks.We find that both market power and average revenues declined among these banks between 1996 and 2006. While we find clear evidence supporting the QLH, estimated effects of the QLH are small from an economical perspective.
Motivated by the recent discussion of the declining importance of deposits as banks´ major source of funding we investigate which factors determine funding costs at local banks. Using a panel data set of more than 800 German local savings and cooperative banks for the period from 1998 to 2004 we show that funding costs are not only driven by the relative share of comparatively cheap deposits of bank´s liabilities but among other factors especially by the size of the bank. In our empirical analysis we find strong and robust evidence that, ceteris paribus, smaller banks exhibit lower funding costs than larger banks suggesting that small banks are able to attract deposits more cheaply than their larger counterparts. We argue that this is the case because smaller banks interact more personally with customers, operate in customers´ geographic proximity and have longer and stronger relationships than larger banks and, hence, are able to charge higher prices for their services. Our finding of a strong influence of bank size on funding costs is also in an in- ternational context of great interest as mergers among small local banks - the key driver of bank growth - are a recent phenomenon not only in European banking that is expected to continue in the future. At the same time, net interest income remains by far the most important source of revenue for most local banks, accounting for approximately 70% of total operating revenues in the case of German local banks. The influence of size on funding costs is of strong economic relevance: our results suggest that an increase in size by 50%, for example, from EUR 500 million in total assets to EUR 750 million (exemplary for M&A transactions among local banks) increases funding costs, ceteris paribus, by approximately 18 basis points which relates to approx. 7% of banks´ average net interest margin.
We study the relation between cognitive abilities and stockholding using the recent Survey of Health, Ageing and Retirement in Europe (SHARE), which has detailed data on wealth and portfolio composition of individuals aged 50+ in 11 European countries and three indicators of cognitive abilities: mathematical, verbal fluency, and recall skills. We find that the propensity to invest in stocks is strongly associated with cognitive abilities, for both direct stock market participation and indirect participation through mutual funds and retirement accounts. Since the decision to invest in less information-intensive assets (such as bonds) is less strongly related to cognitive abilities, we conclude that the association between cognitive abilities and stockholding is driven by information constraints, rather than by features of preferences or psychological traits.
We investigate whether information sharing among banks has affected credit market performance in the transition countries of Eastern Europe and the former Soviet Union, using a large sample of firm-level data. Our estimates show that information sharing is associated with improved availability and lower cost of credit to firms. This correlation is stronger for opaque firms than transparent ones and stronger in countries with weak legal environments than in those with strong legal environments. In cross-sectional estimates, we control for variation in country-level aggregate variables that may affect credit, by examining the differential impact of information sharing across firm types. In panel estimates, we also control for the presence of unobserved heterogeneity at the firm level, as well as for changes in macroeconomic variables and the legal environment.
The single most important policy-induced innovation in the international financial system since the collapse of the Bretton-Woods regime is the institution of the European Monetary Union. This paper provides an account of how the process of financial integration has promoted financial development in the euro area. It starts by defining financial integration and how to measure it, analyzes the barriers that can prevent it and the effects of their removal on financial markets, and assesses whether the euro area has actually become more integrated. It then explores to which extent these changes in financial markets have influenced the performance of the euro-area economy, that is, its growth and investment, as well as its ability to adjust to shocks and to allow risk-sharing. The paper concludes analyzing further steps that are required to consolidate financial integration and enhance the future stability of financial markets.
Various concurrency primitives have been added to sequential programming languages, in order to turn them concurrent. Prominent examples are concurrent buffers for Haskell, channels in Concurrent ML, joins in JoCaml, and handled futures in Alice ML. Even though one might conjecture that all these primitives provide the same expressiveness, proving this equivalence is an open challenge in the area of program semantics. In this paper, we establish a first instance of this conjecture. We show that concurrent buffers can be encoded in the lambda calculus with futures underlying Alice ML. Our correctness proof results from a systematic method, based on observational semantics with respect to may and must convergence.
A data set of annual values of area equipped for irrigation for all 236 countries in the world during the time period 1900 - 2003 was generated. The basis for this data product was information available through various online data bases and from other published materials. The complete time series were then constructed around the reported data applying six statistical methods. The methods are discussed in terms of reliability and data uncertainties. The total area equipped for irrigation in the world in 1900 was 53.2 million hectares. Irrigation was mainly practiced in all the arid regions of the globe and in paddy rice areas of South and East Asia. In some temperate countries in Western Europe irrigation was practiced widely on pastures and meadows. The time series suggest a modest rate of increase of irrigated areas in the first half of the 20th century followed by a more dynamic development in the second half. The turn of the century is characterized by an overall consolidating trend resulting at a total of 285.8 million hectares in 2003. The major contributing countries have changed little throughout the century. This data product is regarded as a preliminary result toward an ongoing effort to develop a detailed data set and map of areas equipped for irrigation in the world over the 20th century using sub-national statistics and historical irrigation maps.
We argue for incorporating the financial economics of market microstructure into the financial econometrics of asset return volatility estimation. In particular, we use market microstructure theory to derive the cross-correlation function between latent returns and market microstructure noise, which feature prominently in the recent volatility literature. The cross-correlation at zero displacement is typically negative, and cross-correlations at nonzero displacements are positive and decay geometrically. If market makers are sufficiently risk averse, however, the cross-correlation pattern is inverted. Our results are useful for assessing the validity of the frequently-assumed independence of latent price and microstructure noise, for explaining observed cross-correlation patterns, for predicting as-yet undiscovered patterns, and for making informed conjectures as to improved volatility estimation methods.
The future of securitization
(2008)
Securitization is a financial innovation that experiences a boom-bust cycle, as many other innovations before. This paper analyzes possible reasons for the breakdown of primary and secondary securitization markets, and argues that misaligned incentives along the value chain are the primary cause of the problems. The illiquidity of asset and interbank markets, in this view, is a market failure derived from ill-designed mechanisms of coordinating financial intermediaries and investors. Thus, illiquidity is closely related to the design of the financial chains. Our policy conclusions emphasize crisis prevention rather than crisis management, and the objective is to restore a “comprehensive incentive alignment”. The toe-hold for strengthening regulation is surprisingly small. First, we emphasize the importance of equity piece retention for the long-term quality of the underlying asset pool. As a consequence, equity piece allocation needs to be publicly known, alleviating market pricing. Second, on a micro level, accountability of managers can be improved by compensation packages aiming at long term incentives, and penalizing policies with destabilizing effects on financial markets. Third, on a macro level, increased transparency relating to effective risk transfer, risk-related management compensation, and credible measurement of rating performance stabilizes the valuation of financial assets and, hence, improves the solvency of financial intermediaries. Fourth, financial intermediaries, whose risk is opaque, may be subjected to higher capital requirements.
The impact of European integration on the German system of pharmaceutical product authorization
(2008)
The European Union has evolved since 1965 into an influential political player in the regulation of pharmaceutical safety standards. The objective of establishing a single European market for pharmaceuticals makes it necessary for member-states to adopt uniform safety standards and marketing authorization procedures. This article investigates the impact of the European integration process on the German marketing authorization system for pharmaceuticals. The analysis shows that the main focal points and objectives of European regulation of pharmaceutical safety have shifted since 1965. The initial phase saw the introduction of uniform European safety standards as a result of which Germany was obliged to undertake “catch-up” modernization. From the mid-1970s, these standards were extended and specified in greater detail. Since the mid-1990s, a process of reorientation has been under way. The formation of the European Agency for the Evaluation of Medicinal Products (EMEA) and the growing importance of the European authorization procedure, combined with intensified global competition on pharmaceutical markets, are exerting indirect pressure for EU member-states to adjust their medicines policies. Consequently, over the past few years Germany has been engaged in a competition-oriented reorganization of its pharmaceutical product authorization system the outcome of which will be to give higher priority to economic interests.
We study the responses of residential property and equity prices, inflation and economic activity to monetary policy shocks in 17 countries, using data spanning 1986-2006, using single-country VARs and panel VARs in which we distinguish between groups of countries depending on their financial systems. The effect of monetary policy on property prices is about three times as large as its impact on GDP. Using monetary policy to guard against financial instability by offsetting asset-price movements thus has sizable effects on economic activity. While the financial structure influences the impact of policy on asset prices, its importance appears limited.
This paper explores the role of trade integration—or openness—for monetary policy transmission in a medium-scale New Keynesian model. Allowing for strategic complementarities in price-setting, we highlight a new dimension of the exchange rate channel by which monetary policy directly impacts domestic inflation. Although the strength of this effect increases with economic openness, it also requires that import prices respond to exchange rate changes. In this case domestic producers find it optimal to adjust their prices to exchange rate changes which alter the domestic currency price of their foreign competitors. We pin down key parameters of the model by matching impulse responses obtained from a vector autoregression on U.S. time series relative to an aggregate of industrialized countries. While we find evidence for strong complementarities, exchange rate pass-through is limited. Openness has therefore little bearing on monetary transmission in the estimated model.
Bayesian learning provides the core concept of processing noisy information. In standard Bayesian frameworks, assessing the price impact of information requires perfect knowledge of news’ precision. In practice, however, precision is rarely dis- closed. Therefore, we extend standard Bayesian learning, suggesting traders infer news’ precision from magnitudes of surprises and from external sources. We show that interactions of the different precision signals may result in highly nonlinear price responses. Empirical tests based on intra-day T-bond futures price reactions to employment releases confirm the model’s predictions and show that the effects are statistically and economically significant.
The popular Nelson-Siegel (1987) yield curve is routinely fit to cross sections of intra-country bond yields, and Diebold and Li (2006) have recently proposed a dynamized version. In this paper we extend Diebold-Li to a global context, modeling a potentially large set of country yield curves in a framework that allows for both global and country-specific factors. In an empirical analysis of term structures of government bond yields for the Germany, Japan, the U.K. and the U.S., we find that global yield factors do indeed exist and are economically important, generally explaining significant fractions of country yield curve dynamics, with interesting differences across countries.
Measuring financial asset return and volatilty spillovers, with application to global equity markets
(2008)
We provide a simple and intuitive measure of interdependence of asset returns and/or volatilities. In particular, we formulate and examine precise and separate measures of return spillovers and volatility spillovers. Our framework facilitates study of both non-crisis and crisis episodes, including trends and bursts in spillovers, and both turn out to be empirically important. In particular, in an analysis of nineteen global equity markets from the early 1990s to the present, we find striking evidence of divergent behavior in the dynamics of return spillovers vs. volatility spillovers: Return spillovers display a gently increasing trend but no bursts, whereas volatility spillovers display no trend but clear bursts.
Research with Keynesian-style models has emphasized the importance of the output gap for policies aimed at controlling inflation while declaring monetary aggregates largely irrelevant. Critics, however, have argued that these models need to be modified to account for observed money growth and inflation trends, and that monetary trends may serve as a useful cross-check for monetary policy. We identify an important source of monetary trends in form of persistent central bank misperceptions regarding potential output. Simulations with historical output gap estimates indicate that such misperceptions may induce persistent errors in monetary policy and sustained trends in money growth and inflation. If interest rate prescriptions derived from Keynesian-style models are augmented with a cross-check against money-based estimates of trend inflation, inflation control is improved substantially.
Increasingly, individuals are in charge of their own financial security and are confronted with ever more complex financial instruments. However, there is evidence that many individuals are not well-equipped to make sound saving decisions. This paper demonstrates widespread financial illiteracy among the U.S. population, particularly among specific demographic groups. Those with low education, women, African-Americans, and Hispanics display particularly low levels of literacy. Financial literacy impacts financial decision-making. Failure to plan for retirement, lack of participation in the stock market, and poor borrowing behavior can all be linked to ignorance of basic financial concepts. While financial education programs can result in improved saving behavior and financial decision-making, much can be done to improve these programs’ effectiveness.
Traditionally, aggregate liquidity shocks are modelled as exogenous events. Extending our previous work (Cao & Illing, 2007), this paper analyses the adequate policy response to endogenous systemic liquidity risk. We analyse the feedback between lender of last resort policy and incentives of private banks, determining the aggregate amount of liquidity available. We show that imposing minimum liquidity standards for banks ex ante are a crucial requirement for sensible lender of last resort policy. In addition, we analyse the impact of equity requirements and narrow banking, in the sense that banks are required to hold sufficient liquid funds so as to pay out in all contingencies. We show that such a policy is strictly inferior to imposing minimum liquidity standards ex ante combined with lender of last resort policy.
Modern macroeconomics empirically addresses economy-wide incentives behind economic actions by using insights from the way a single representative household would behave. This analytical approach requires that incentives of the poor and the rich are strictly aligned. In empirical analysis a challenging complication is that consumer and income data are typically available at the household level, and individuals living in multimember households have the potential to share goods within the household. The analytical approach of modern macroeconomics would require that intra-household sharing is also strictly aligned across the rich and the poor. Here we have designed a survey method that allows the testing of this stringent property of intra-household sharing and find that it holds: once expenditures for basic needs are subtracted from disposable household income, household-size economies implied by the remainder household incomes are the same for the rich and the poor.
After the pioneering German “Aktiengesetz” of 1965 and the Brazilian “Lei das Sociedades Anónimas” of 1976, Portugal has become the third country in the world to enact a specific regulation on groups of companies. The Code of Commercial Companies (“Código das Sociedades Comerciais”, abbreviately hereinafter CSC), enacted in 1986, contains a unitary set of rules regulating the relationships between companies, in general, and the groups of companies, in particular (arts. 481° to 508°-E CSC). With this set of rules, the Portuguese legislator has dealt with one of the major topics of modern Company Law. While this branch of law is traditionally conceived as the law of the individual company, modern economic reality is characterized by the massive emergence of large-scale enterprise networks, where parts of a whole business are allocated and insulated in several legally independent companies submitted to an unified economic direction. As Tom HADDEN put it: “Company lawyers still write and talk as if the single independent company, with its shareholders, directors and employees, was the norm. In reality, the individual company ceased to be the most significant form of organization in the 1920s and 1930s. The commercial world is now dominated both nationally and internationally by complex groups of companies”. This trend, which is now observable in any of the largest economies in the world, holds also true for small markets such as Portugal. Although Portuguese economy is still dominated by small and medium-sized enterprises, the organizational structure of the group has always been extremely common. During the 70s, it was estimated that the seven largest groups of companies owned about 50% of the equity capital of all domestic enterprises and were alone responsible for 3/4 of the internal national product. Such a trend has continued and even highlighted in the next decades, surviving to different political and economic scenarios: during the 80s, due to the process of state nationalization of these groups, an enormous public group with more than one thousand controlled companies has been created (“IPE - Instituto de Participações do Estado”); and during the 90s until today, thanks to the reprivatisation movement and the opening of our national market, we assisted to the re-emergence of some large private groups, composed of several hundred subsidiaries each, some of which are listed in foreign stock exchange markets (e.g., in the banking sector, “BCP – Banco Comercial Português”, in the industrial area, “SONAE”, and in the media and communication area, “Portugal-Telecom”).
Reform of the securities class action is once again the subject of national debate. The impetus for this debate is the reports of three different groups – The Committee on Capital Market Regulation, The Commission on the Regulation of U.S. Capital Markets In the 21st Century, and McKinsey & Company. Each of the reports focuses on a single theme: how the contemporary regulatory culture places U.S. capital markets at a competitive disadvantage to foreign markets. While multiple regulatory forces are targeted by each report’s call for reform, each of the reports singles out securities class actions as one of the prime villains that place U.S. capital markets at a competitive disadvantage. The reports’ recommendations range from insignificant changes to drastic curtailments of private class actions. Surprisingly, these current-day cries echo calls for reform heeded by Congress in the not too distant past. Major reform of the securities class action occurred with the Private Securities Litigation Reform Act of 1995.5 Among the PSLRA’s contributions is the introduction of procedures by which the court chooses from among competing petitioners a lead plaintiff for the class. The statute commands that the petitioner with the largest financial loss suffered as a consequence of the defendant’s alleged misrepresentation is presumed to be the most adequate plaintiff. Thus, the lead plaintiff provision supplants the traditional “first to file” rule for selecting the suit’s plaintiff with a mechanism that seeks to harness to the plaintiff’s economic self interest to the suits’ prosecution. Also, by eliminating the race to be the first to file, the lead plaintiff provision seeks to avoid “hair trigger” filings by overly eager plaintiffs’ counsel which Congress believed too frequently gave rise to incomplete and insubstantially pled causes of action. The PSLRA also introduced for securities class actions a heightened pleading requirement8 as well as a bar to the plaintiff obtaining any discovery prior to the district court disposing of the defendants’ motions to dismiss. By introducing the requirement that allegations involving fraud must be plead not only with particularity, but also that the pled facts must establish a “strong inference” of fraud, the PSLRA cast aside, albeit only for securities actions, the much lower notice pleading requirement that has been a fixture of American civil procedure for decades. Substantive changes to the law were also introduced by the PSLRA. With few exceptions, joint and several liability was replaced by proportionate liability so that a particular defendant’s liability is capped by that defendant’s relative degree of fault. Similarly, contribution rights among co-violators are also based on proportionate fault of each defendant. Three years after the PSLRA, Congress returned to the topic again by enacting the Securities Litigation Uniform Standards Act;13 this provision was prompted by aggressive efforts of plaintiff lawyers to bypass the limitations, most notably the bar to discovery and higher pleading requirement, of the PSLRA by bringing suit in state court. Post-SLUSA, securities fraud class actions are exclusively the domain of the federal court. In this paper, we examine the impact of the PSLRA and more particularly the impact the type of lead plaintiff on the size of settlements in securities fraud class actions. We thus provide insight into whether the type of plaintiff that heads the class action impacts the overall outcome of the case. Furthermore, we explore possible indicia that may explain why some suits settle for extremely small sums – small relative to the “provable losses” suffered by the class, small relative to the asset size of the defendantcompany, and small relative to other settlements in our sample. This evidence bears heavily on the debate over “strike suits.” Part I of this paper sets forth the contemporary debate surrounding the need for further reforms of securities class actions. In this section, we set forth the insights advanced in three prominent reports focused on the competitiveness of U.S. capital markets. In Part II we first provide descriptive statistics of our extensive data set, and thenuse multivariate regression analysis to explore the underlying relationships. In Part III, we closely examine small settlements for clues to whether they reflect evidence of strike suits. We conclude in Part IV with a set of policy recommendations based on our analysis of the data. Our goals in this paper are more modest than the Committee Report, the Chamber Report and the McKinsey Report, each of which called for wide-ranging reforms: we focus on how the PSLRA changed securities fraud settlements so as to determine whether the reforms it introduced accomplished at least some of the Act’s important goals. If the PSLRA was successful, and we think it was, then one must be somewhat skeptical of the need for further cutbacks in private securities class action so soon after the Act was passed.
The market reaction to legal shocks and their antidotes : lessons from the sovereign debt market
(2008)
This Article examines the market reaction to a series of legal events concerning the judicial interpretation of the pari passu clause in sovereign debt instruments. More generally, the Article provides insights into the reactions of investors (predominantly financial institutions), issuers (sovereigns), and those who draft bond covenants (lawyers), to unanticipated changes in the judicial interpretation of certain covenant terms.
How do fiscal and technology shocks affect real exchange rates? : New evidence for the United States
(2008)
Using vector autoregressions on U.S. time series relative to an aggregate of industrialized countries, this paper provides new evidence on the dynamic effects of government spending and technology shocks on the real exchange rate and the terms of trade. To achieve identification, we derive robust restrictions on the sign of several impulse responses from a two-country general equilibrium model. We find that both the real exchange rate and the terms of trade – whose responses are left unrestricted – depreciate in response to expansionary government spending shocks and appreciate in response to positive technology shocks.
Motivated by the prominent role of electronic limit order book (LOB) markets in today’s stock market environment, this paper provides the basis for understanding, reconstructing and adopting Hollifield, Miller, Sandas, and Slive’s (2006) (henceforth HMSS) methodology for estimating the gains from trade to the Xetra LOB market at the Frankfurt Stock Exchange (FSE) in order to evaluate its performance in this respect. Therefore this paper looks deeply into HMSS’s base model and provides a structured recipe for the planned implementation with Xetra LOB data. The contribution of this paper lies in the modification of HMSS’s methodology with respect to the particularities of the Xetra trading system that are not yet considered in HMSS’s base model. The necessary modifications, as expressed in terms of empirical caveats, are substantial to derive unbiased market efficiency measures for Xetra in the end.
We explore the pattern of elderly homeownership using microeconomic surveys of 15 OECD countries, merging 60 national household surveys on about 300,000 individuals. In all countries the survey is repeated over time, permitting construction of an international dataset of repeated cross-sectional data. We find that ownership rates decline considerably after age 60 in all countries. However, a large part of the decline depends on cohort effects. Adjusting for them, we find that ownership rates start falling after age 70 and reach a percentage point per year decline after age 75. We find that differences across country ownership trajectories are correlated with indicators measuring the degree of market regulations.
This paper introduces adaptive learning and endogenous indexation in the New-Keynesian Phillips curve and studies disinflation under inflation targeting policies. The analysis is motivated by the disinflation performance of many inflation-targeting countries, in particular the gradual Chilean disinflation with temporary annual targets. At the start of the disinflation episode price-setting firms’ expect inflation to be highly persistent and opt for backward-looking indexation. As the central bank acts to bring inflation under control, price-setting firms revise their estimates of the degree of persistence. Such adaptive learning lowers the cost of disinflation. This reduction can be exploited by a gradual approach to disinflation. Firms that choose the rate for indexation also re-assess the likelihood that announced inflation targets determine steady-state inflation and adjust indexation of contracts accordingly. A strategy of announcing and pursuing short-term targets for inflation is found to influence the likelihood that firms switch from backward-looking indexation to the central bank’s targets. As firms abandon backward-looking indexation the costs of disinflation decline further. We show that an inflation targeting strategy that employs temporary targets can benefit from lower disinflation costs due to the reduction in backward-looking indexation.
Monetary policy analysts often rely on rules-of-thumb, such as the Taylor rule, to describe historical monetary policy decisions and to compare current policy to historical norms. Analysis along these lines also permits evaluation of episodes where policy may have deviated from a simple rule and examination of the reasons behind such deviations. One interesting question is whether such rules-of-thumb should draw on policymakers "forecasts of key variables such as inflation and unemployment or on observed outcomes. Importantly, deviations of the policy from the prescriptions of a Taylor rule that relies on outcomes may be due to systematic responses to information captured in policymakers" own projections. We investigate this proposition in the context of FOMC policy decisions over the past 20 years using publicly available FOMC projections from the biannual monetary policy reports to the Congress (Humphrey-Hawkins reports). Our results indicate that FOMC decisions can indeed be predominantly explained in terms of the FOMC´s own projections rather than observed outcomes. Thus, a forecast-based rule-of-thumb better characterizes FOMC decision-making. We also confirm that many of the apparent deviations of the federal funds rate from an outcome-based Taylor-style rule may be considered systematic responses to information contained in FOMC projections.
Risk transfer with CDOs
(2008)
Modern bank management comprises both classical lending business and transfer of asset risk to capital markets through securitization. Sound knowledge of the risks involved in securitization transactions is a prerequisite for solid risk management. This paper aims to resolve a part of the opaqueness surrounding credit-risk allocation to tranches that represent claims of different seniority on a reference portfolio. In particular, this paper analyzes the allocation of credit risk to different tranches of a CDO transaction when the underlying asset returns are driven by a common macro factor and an idiosyncratic component. Junior and senior tranches are found to be nearly orthogonal, motivating a search for the whereabout of systematic risk in CDO transactions. We propose a metric for capturing the allocation of systematic risk to tranches. First, in contrast to a widely-held claim, we show that (extreme) tail risk in standard CDO transactions is held by all tranches. While junior tranches take on all types of systematic risk, senior tranches take on almost no non-tail risk. This is in stark contrast to an untranched bond portfolio of the same rating quality, which on average suffers substantial losses for all realizations of the macro factor. Second, given tranching, a shock to the risk of the underlying asset portfolio (e.g. a rise in asset correlation or in mean portfolio loss) has the strongest impact, in relative terms, on the exposure of senior tranche CDO-investors. Our findings can be used to explain major stylized facts observed in credit markets.
We show that the use of correlations for modeling dependencies may lead to counterintuitive behavior of risk measures, such as Value-at-Risk (VaR) and Expected Short- fall (ES), when the risk of very rare events is assessed via Monte-Carlo techniques. The phenomenon is demonstrated for mixture models adapted from credit risk analysis as well as for common Poisson-shock models used in reliability theory. An obvious implication of this finding pertains to the analysis of operational risk. The alleged incentive suggested by the New Basel Capital Accord (Basel II), amely decreasing minimum capital requirements by allowing for less than perfect correlation, may not necessarily be attainable.
The paper proposes a panel cointegration analysis of the joint development of government expenditures and economic growth in 23 OECD countries. The empirical evidence provides indication of a structural positive correlation between public spending and per-capita GDP which is consistent with the so-called Wagner´s law. A long-run elasticity larger than one suggests a more than proportional increase of government expenditures with respect to economic activity. In addition, according to the spirit of the law, we found that the correlation is usually higher in countries with lower per-capita GDP, suggesting that the catching-up period is characterized by a stronger development of government activities with respect to economies in a more advanced state of development.
Risk transfer with CDOs
(2008)
Modern bank management comprises both classical lending business and transfer of asset risk to capital markets through securitization. Sound knowledge of the risks involved in securitization transactions is a prerequisite for solid risk management. This paper aims to resolve a part of the opaqueness surrounding credit-risk allocation to tranches that represent claims of different seniority on a reference portfolio. In particular, this paper analyzes the allocation of credit risk to different tranches of a CDO transaction when the underlying asset returns are driven by a common macro factor and an idiosyncratic component. Junior and senior tranches are found to be nearly orthogonal, motivating a search for the where about of systematic risk in CDO transactions. We propose a metric for capturing the allocation of systematic risk to tranches. First, in contrast to a widely-held claim, we show that (extreme) tail risk in standard CDO transactions is held by all tranches. While junior tranches take on all types of systematic risk, senior tranches take on almost no non-tail risk. This is in stark contrast to an untranched bond portfolio of the same rating quality, which on average suffers substantial losses for all realizations of the macro factor. Second, given tranching, a shock to the risk of the underlying asset portfolio (e.g. a rise in asset correlation or in mean portfolio loss) has the strongest impact, in relative terms, on the exposure of senior tranche CDO-investors. Our findings can be used to explain major stylized facts observed in credit markets.
A new global crop water model was developed to compute blue (irrigation) water requirements and crop evapotranspiration from green (precipitation) water at a spatial resolution of 5 arc minutes by 5 arc minutes for 26 different crop classes. The model is based on soil water balances performed for each crop and each grid cell. For the first time a new global data set was applied consisting of monthly growing areas of irrigated crops and related cropping calendars. Crop water use was computed for irrigated land and the period 1998 – 2002. In this documentation report the data sets used as model input and methods used in the model calculations are described, followed by a presentation of the first results for blue and green water use at the global scale, for countries and specific crops. Additionally the simulated seasonal distribution of water use on irrigated land is presented. The computed model results are compared to census based statistical information on irrigation water use and to results of another crop water model developed at FAO.
A data set of monthly growing areas of 26 irrigated crops (MGAG-I) and related crop calendars (CC-I) was compiled for 402 spatial entities. The selection of the crops consisted of all major food crops including regionally important ones (wheat, rice, maize, barley, rye, millet, sorghum, soybeans, sunflower, potatoes, cassava, sugar cane, sugar beets, oil palm, rapeseed/canola, groundnuts/peanuts, pulses, citrus, date palm, grapes/vine, cocoa, coffee), major water-consuming crops (cotton), and unspecified other crops (other perennial crops, other annual crops, managed grassland). The data set refers to the time period 1998-2002 and has a spatial resolution of 5 arc minutes by 5 arc minutes which is 8 km by 8 km at the equator. This is the first time that a data set of cell-specific irrigated growing areas of irrigated crops with this spatial resolution was created. The data set is consistent to the irrigated area and water use statistics of the AQUASTAT programme of the Food and Agriculture Organization of the United Nations (FAO) (http://www.fao.org/ag/agl/aglw/aquastat/main/index.stm) and the Global Map of Irrigation Areas (GMIA) (http://www.fao.org/ag/agl/aglw/aquastat/irrigationmap/index.stm). At the cell-level it was tried to maximise consistency to the cropland extent and cropland harvested area from the Department of Geography and Earth System Science Program of the McGill University at Montreal, Quebec, Canada and the Center for Sustainability and the Global Environment (SAGE) of the University of Wisconsin at Madison, USA (http://www.geog.mcgill.ca/~nramankutty/ Datasets/Datasets.html and http://geomatics.geog.mcgill.ca/~navin/pub/Data/175crops2000/). The consistency between the grid product and the input data was quantified. MGAG-I and CC-I are fully consistent to each other on entity level. For input data other than CC-I, the consistency of MGAG-I on cell level was calculated. The consistency of MGAG-I with respect to the area equipped for irrigation (AEI) of GMIA and to the cropland extent of SAGE was characterised by the sum of the cell-specific maximum difference between the MGAG-I monthly total irrigated area and the reference area when the latter was exceeded in the grid cell. The consistency of the harvested area contained in MGAG-I with respect to SAGE harvested area was characterised by the crop-specific sum of the cell-specific difference between MGAG-I harvested area and the SAGE harvested area when the latter was exceeded in the grid cell. In all three cases, the sums are the excess areas that should not have been distributed under the assumption that the input data were correct. Globally, this cell-level excess of MGAG-I as compared to AEI is 331,304 ha or only about 0.12 % of the global AEI of 278.9 Mha found in the original grid. The respective cell-level excess of MGAG-I as compared to the SAGE cropland extent is 32.2 Mha, corresponding to about 2.2 % of the total cropland area. The respective cell-level excess of MGAG-I as compared to the SAGE harvested area is 27 % of the irrigated harvested area, or 11.5 % of the AEI. In a further step that will be published later also rainfed areas were compiled in order to form the Global data set of monthly irrigated and rainfed crop areas around the year 2000 (MIRCA2000). The data set can be used for global and continental-scale studies on food security and water use. In the future, it will be improved, e.g. with a better spatial resolution of crop calendars and an improved crop distribution algorithm. The MIRCA2000 data set, its full documentation together with future updates will be freely available through the following long-term internet site: http://www.geo.uni-frankfurt.de/ipg/ag/dl/forschung/MIRCA/index.html. The research presented here was funded by the German Research Foundation (Deutsche Forschungsgemeinschaft, DFG) within the framework of the research project entitled "Consistent assessment of global green, blue and virtual water fluxes in the context of food production: regional stresses and worldwide teleconnections". The authors thank Navin Ramankutty and Chad Monfreda for making available the current SAGE datasets on cropland extent (Ramankutty et al., 2008) and harvested area (Monfreda et al., 2008) prior to their publication.
The introduction of a common currency as well as the harmonization of rules and regulations in Europe has significantly reduced distance in all its guises. With reduced costs of overcoming space, this emphasizes centripetal forces and it should foster consolidation of financial activity. In a national context, as a rule, this led to the emergence of one financial center. Hence, Europeanization of financial and monetary affairs could foretell the relegation of some European financial hubs such as Frankfurt and Paris to third-rank status. Frankfurt’s financial history is interesting insofar as it has lost (in the 1870s) and regained (mainly in the 1980s) its preeminent place in the German context. Because Europe is still characterized by local pockets of information-sensitive assets as well as a demand for variety the national analogy probably does not hold. There is room in Europe for a number of financial hubs of an international dimension, including Frankfurt.
In this paper we consider the dynamics of spot and futures prices in the presence of arbitrage. We propose a partially linear error correction model where the adjustment coefficient is allowed to depend non-linearly on the lagged price difference. We estimate our model using data on the DAX index and the DAX futures contract. We find that the adjustment is indeed nonlinear. The linear alternative is rejected. The speed of price adjustment is increasing almost monotonically with the magnitude of the price difference.
This study develops a novel 2-step hedonic approach, which is used to construct a price index for German paintings. This approach enables the researcher to use every single auction record, instead of only those auction records that belong to a sub-sample of selected artists. This results in a substantially larger sample available for research and it lowers the selection bias that is inherent in the traditional hedonic and repeat sales methodologies. Using a unique sample of 61,135 auction records for German artworks created by 5,115 different artists over the period 1985 to 2007, we find that the geometric annual return on German art is just 3.8 percent, with a standard deviation of 17.87 percent. Although our results indicate that art underperforms the market portfolio and is not proportionally rewarded for downside risk, under some circumstances art should be included in an optimal portfolio for diversification purposes.
While companies have emerged as very proactive donors in the wake of recent major disasters like Hurricane Katrina, it remains unclear whether that corporate generosity generates benefits to firms themselves. The literature on strategic philanthropy suggests that such philanthropic behavior may be valuable because it can generate direct and indirect benefits to the firm, yet it is not known whether investors interpret donations in this way. We develop hypotheses linking the strategic character of donations to positive abnormal returns. Using event study methodology, we investigate stock market reactions to corporate donation announcements by 108 US firms made in response to Hurricane Katrina. We then use regression analysis to examine if our hypothesized predictors are associated with positive abnormal returns. Our results show that overall, corporate donations were linked to neither positive nor negative abnormal returns. We do, however, see that a number of factors moderate the relationship between donation announcements and abnormal stock returns. Implications for theory and practice are discussed.
We estimate the degree of 'stickiness' in aggregate consumption growth (sometimes interpreted as reflecting consumption habits) for thirteen advanced economies. We find that, after controlling for measurement error, consumption growth has a high degree of autocorrelation, with a stickiness parameter of about 0.7 on average across countries. The sticky-consumption-growth model outperforms the random walk model of Hall (1978), and typically fits the data better than the popular Campbell and Mankiw (1989) model. In several countries, the sticky-consumption-growth and Campbell-Mankiw models work about equally well.
Generally, information provision and certifcation have been identified as the major economic functions of rating agencies. This paper analyzes whether the “watchlist” (rating review) instrument has extended the agencies' role towards a monitoring position, as proposed by Boot, Milbourn, and Schmeits (2006). Using a data set of Moody's rating history between 1982 and 2004, we find that the overall information content of rating action has indeed increased since the introduction of the watchlist procedure. Our findings suggest that rating reviews help to establish implicit monitoring contracts between agencies and borrowers and as such enable a finer partition of rating information, thereby contributing to a higher information quality.
On 27 and 28 September 2007, a commission formed on the initiative of the authors held its first meeting in Aarhus, Denmark to deliberate on its goal of drafting a "European Model Company Law Act" (EMCLA). This project, outlined in the following pages, aims neither to force a mandatory harmonization of national company law nor to create a further, European corporate form. The goal is rather to draft model rules for a corporation that national legislatures would be free to adopt in whole or in part. Thus, the project is thought as an alternative and supplement to the existing EU instruments for the convergence of company law. The present EU instruments, their prerequisites and limits will be discussed in more detail in Part II, below. Part III will examine the US experience with such "model acts" in the area of company law. Part IV will then conclude by discussing several topics concerning the content of an EMCLA, introducing the members of the EMCLA Working Group, and explaining the Group's preliminary working plan.
This paper identifies some common errors that occur in comparative law, offers some guidelines to help avoid such errors, and provides a framework for entering into studies of the company laws of three major jurisdictions. The first section illustrates why a conscious approach to comparative company law is useful. Part I discusses some of the problems that can arise in comparative law and offers a few points of caution that can be useful for practical, theoretical and legislative comparative law. Part II discusses some relatively famous examples of comparative analysis gone astray in order to demonstrate the utility of heeding the outlined points of caution. The second section offers a framework for approaching comparative company law. Part III provides an example of using functional definition to demarcate the topic "company law", offering an "effects" test to determine whether a given provision of law should be considered as functionally part of the rules that govern the core characteristics of companies. It does this by presenting the relevant company law statutes and related topical laws of Germany, the United Kingdom and the United States, using Delaware as a proxy for the 50 states. On the basis of this definition, Part IV analyzes the system of legal functions that comprises "company law" in the United States and the European Union. It selects as the predominant factor for consideration the jurisdictions, sub-jurisdictions and rule-making entities that have legislative or rule-making competence in the relevant territorial unit, analyzes the extent of their power, presents the type of law (rules) they enact (issue), and discusses the concrete manner in which the laws and rules of the jurisdictions and sub-jurisdictions can legally interact. Part V looks at the way these jurisdictions do interact on the temporal axis of history, that is, their actual influence on each other, which in the relevant jurisdictions currently takes the form of regulatory competition and legislative harmonization. The method of the approach outlined in this paper borrows much from system theory. The analysis attempts to be detailed without losing track of the overall jurisdictional framework in the countries studied.