Refine
Year of publication
- 2008 (100) (remove)
Document Type
- Working Paper (100) (remove)
Language
- English (100) (remove)
Is part of the Bibliography
- no (100)
Keywords
- USA (7)
- Bank (5)
- Geldpolitik (5)
- Lambda-Kalkül (5)
- Operationale Semantik (5)
- Programmiersprache (5)
- Haushalt (4)
- Liquidität (4)
- Aging (3)
- Deutschland (3)
We investigate methods and tools for analyzing translations between programming languages with respect to observational semantics. The behavior of programs is observed in terms of may- and mustconvergence in arbitrary contexts, and adequacy of translations, i.e., the reflection of program equivalence, is taken to be the fundamental correctness condition. For compositional translations we propose a notion of convergence equivalence as a means for proving adequacy. This technique avoids explicit reasoning about contexts, and is able to deal with the subtle role of typing in implementations of language extensions.
We investigate methods and tools for analysing translations between programming languages with respect to observational semantics. The behaviour of programs is observed in terms of may- and mustconvergence in arbitrary contexts, and adequacy of translations, i.e., the reflection of program equivalence, is taken to be the fundamental correctness condition. For compositional translations we propose a notion of convergence equivalence as a means for proving adequacy. This technique avoids explicit reasoning about contexts, and is able to deal with the subtle role of typing in implementations of language extensions.
We investigate methods and tools for analysing translations between programming languages with respect to observational semantics. The behaviour of programs is observed in terms of may- and mustconvergence in arbitrary contexts, and adequacy of translations, i.e., the reflection of program equivalence, is taken to be the fundamental correctness condition. For compositional translations we propose a notion of convergence equivalence as a means for proving adequacy. This technique avoids explicit reasoning about contexts, and is able to deal with the subtle role of typing in implementations of language extensions.
This paper proves several generic variants of context lemmas and thus contributes to improving the tools for observational semantics of deterministic and non-deterministic higher-order calculi that use a small-step reduction semantics. The generic (sharing) context lemmas are provided for may- as well as two variants of must-convergence, which hold in a broad class of extended process- and extended lambda calculi, if the calculi satisfy certain natural conditions. As a guide-line, the proofs of the context lemmas are valid in call-by-need calculi, in callby-value calculi if substitution is restricted to variable-by-variable and in process calculi like variants of the π-calculus. For calculi employing beta-reduction using a call-by-name or call-by-value strategy or similar reduction rules, some iu-variants of ciu-theorems are obtained from our context lemmas. Our results reestablish several context lemmas already proved in the literature, and also provide some new context lemmas as well as some new variants of the ciu-theorem. To make the results widely applicable, we use a higher-order abstract syntax that allows untyped calculi as well as certain simple typing schemes. The approach may lead to a unifying view of higher-order calculi, reduction, and observational equality.
We present a higher-order call-by-need lambda calculus enriched with constructors, case-expressions, recursive letrec-expressions, a seq-operator for sequential evaluation and a non-deterministic operator amb that is locally bottom-avoiding. We use a small-step operational semantics in form of a single-step rewriting system that defines a (nondeterministic) normal order reduction. This strategy can be made fair by adding resources for bookkeeping. As equational theory we use contextual equivalence, i.e. terms are equal if plugged into any program context their termination behaviour is the same, where we use a combination of may- as well as must-convergence, which is appropriate for non-deterministic computations. We show that we can drop the fairness condition for equational reasoning, since the valid equations w.r.t. normal order reduction are the same as for fair normal order reduction. We evolve different proof tools for proving correctness of program transformations, in particular, a context lemma for may- as well as mustconvergence is proved, which restricts the number of contexts that need to be examined for proving contextual equivalence. In combination with so-called complete sets of commuting and forking diagrams we show that all the deterministic reduction rules and also some additional transformations preserve contextual equivalence.We also prove a standardisation theorem for fair normal order reduction. The structure of the ordering <=c a is also analysed: Ω is not a least element, and <=c already implies contextual equivalence w.r.t. may-convergence.
This paper discusses the effect of capital regulation on the risk taking behavior of commercial banks. We first theoretically show that capital regulation works differently in different market structures of banking sectors. In lowly concentrated markets, capital regulation is effective in mitigating risk taking behavior because banks' franchise values are low and banks have incentives to pursue risky strategies in order to increase their franchise values. If franchise values are high, on the other hand, the effect of capital regulation on bank risk taking is ambiguous as banks lack those incentives. We then test the model predictions on a cross-country sample including 421 commercial banks from 61 countries. We find that capital regulation is effective in mitigating risk taking only in markets with a low degree of concentration. The results remain robust after accounting for financial sector development, legal system effciency, and for other country and bank-specific characteristics. Keywords: Banks, market structure, risk shifting, franchise value, capital regulation
CONTENTS Preamble 1. Concept and Drivers of Globalization 1.0 A Brief Historical Perspective 1.1 Concept of Globalization 1.2 Economic Globalization 1.3 Drivers of Economic Globalization 2. Globalization and Markets 2.1 The Free Market System 2.2 Markets and the Solution of Economic Problems 2.3 African Markets and “Getting the Prices Right”. 2.4 Implications of the Imperfect Market System 2.5 Government’s Inevitable Role 2.6 The International Environment/Markets 3. Globalization and Trade Liberalisation 3.1 The Experience of the Developing Countries 3.2 Nigeria’s Experience with Trade Liberalisation 4. Global Economic Integration and Sub-Saharan Africa 4.1 Global Economic Integration 4.2 Africa’s Integration with the World Economy 4.3 The Benefits of Economic Globalization and Sub-Saharan Africa 4.4 Why has Africa Lagged? 5. Nigeria and the Global Economy 5.1 Openness of the Economy and Integration with the World Economy 5.2 Globalization and Nigeria’s Trade 5.3 Globalization and Foreign Capital Flows to Nigeria 5.4 Foreign Capital Flows and Debt Accumulation 5.5 Globalization, Growth and Development 6. Appropriate Policy Responses and Lessons 7. Concluding Remarks 8. Appreciation 9. Annex 10. References
This paper discusses the implications of transnational media production and diasporic networks for the cultural politics of migrant minorities. How are fields of cultural politics transformed if Hirschmann’s famous options ‘exit’ and ‘voice’ are no longer constituting mutually exclusive responses to dissent within a nation-state, but modes of action that can combine and build upon each other in the context of migration and diasporic media activism? Two case studies are discussed in more detail, relating to Alevi amateur television production in Germany and to a Kurdish satellite television station that reaches out to a diaspora across Europe and the Middle East. Keywords: migrant media, transnationalism, Alevis, Kurds, Turkey, Germany
Hong Kong’s Linked Exchange Rate System (LERS) has been in operation for twenty-five years during which time many other fixed exchange rate systems have succumbed to shocks and/or speculative attacks. This fact alone suggests that the LERS is a robust system which enjoys a large measure of credibility in financial markets. This paper intends to investigate whether this is indeed the case, and whether it has been the case throughout its 25-year history. In particular we will use the tools of modern finance to extract information from financial asset prices about market expectations that are related to the credibility of the LERS. The main focus is on how market participants ‘judged’ the various changes made to the LERS, such as the ‘seven technical measures’ introduced in September 1998 and the ‘three refinements’ made in May 2005. These changes have been characterizes as making the system less discretionary over time, and we hypothesize that they have also made it more credible as revealed in the prices of exchange rate related asset prices. We also investigate the relationship between interest rates and exchange rates in the current system in light of modern models of target-zone exchange rate systems. We will examine whether the intramarginal intervention in November 2007 changed the dynamic properties of the exchange rate as suggested by such models.
Ensuring financial stability : financial structure and the impact of monetary policy on asset prices
(2008)
This paper studies the responses of residential property and equity prices, inflation and economic activity to monetary policy shocks in 17 countries, using data spanning 1986-2006. We estimate VARs for individual economies and panel VARs in which we distinguish between groups of countries on the basis of the characteristics of their financial systems. The results suggest that using monetary policy to offset asset price movements in order to guard against financial instability may have large effects on economic activity. Furthermore, while financial structure influences the impact of policy on asset prices, its importance appears limited. Keywords: asset prices, monetary policy, panel VAR. JEL Number: C23, E52
Recently, the Bank of Japan outlined a “two perspectives” approach to the conduct of monetary policy that focuses on risks to price stability over different time horizons. Interpreting this as pertaining to different frequency bands, we use band spectrum regression to study the determination of inflation in Japan. We find that inflation is related to money growth and real output growth at low frequencies and the output gap at higher frequencies. Moreover, this relationship reflects Granger causality from money growth and the output gap to inflation in the relevant frequency bands. Keywords: spectral regression, frequency domain, Phillips curve, quantity theory. JEL Numbers: C22, E3, E5
We study the effect of randomness in the adversarial queueing model. All proofs of instability for deterministic queueing strategies exploit a finespun strategy of insertions by an adversary. If the local queueing decisions in the network are subject to randomness, it is far from obvious, that an adversary can still trick the network into instability. We show that uniform queueing is unstable even against an oblivious adversary. Consequently, randomizing the queueing decisions made to operate a network is not in itself a suitable fix for poor network performances due to packet pileups.
Purim and parodies
(2008)
We investigate, using the 2002 US Health and Retirement Study, the factors influencing individuals’ insecurity and expectations about terrorism, and study the effects these last have on households’ portfolio choices and spending patterns. We find that females, the religiously devout, those equipped with a better memory, the less educated, and those living close to where the events of September 2001 took place worry a lot about their safety. In addition, fear of terrorism discourages households from investing in stocks, mostly through the high levels of insecurity felt by females. Insecurity due to terrorism also makes single men less likely to own a business. Finally, we find evidence of expenditure shifting away from recreational activities that can potentially leave one exposed to a terrorist attack and towards goods that might help one cope with the consequences of terrorism materially (increased use of car and spending on the house) or psychologically (spending on personal care products by females in couples).
We document significant and robust empirical relationships in cross-country panel data between government size or social expenditure on the one hand, and trade and financial development indicators on the other. Across countries, deeper economic integration is associated with more intense government redistribution, but more developed financial markets weaken that relationship. Over time, controlling for country-specific effects, public social expenditure appears to be eroded by globalization trends where financial market development can more easily substitute for it.
The paper provides novel insights on the effect of a firm’s risk management objective on the optimal design of risk transfer instruments. I analyze the interrelation between the structure of the optimal insurance contract and the firm’s objective to minimize the required equity it has to hold to accommodate losses in the presence of multiple risks and moral hazard. In contrast to the case of risk aversion and moral hazard, the optimal insurance contract involves a joint deductible on aggregate losses in the present setting.
This paper analyzes liquidity in an order driven market. We only investigate the best limits in the limit order book, but also take into account the book behind these inside prices. When subsequent prices are close to the best ones and depth at them is substantial, larger orders can be executed without an extensive price impact and without deterring liquidity. We develop and estimate several econometric models, based on depth and prices in the book, as well as on the slopes of the limit order book. The dynamics of different dimensions of liquidity are analyzed: prices, depth at and beyond the best prices, as well as resiliency, i.e. how fast the different liquidity measures recover after a liquidity shock. Our results show a somewhat less favorable image of liquidity than often found in the literature. After a liquidity shock (in the spread or depth or in the book beyond the best limits), several dimension of liquidity deteriorate at the same time. Not only does the inside spread increase, and depth at the best prices decrease, also the difference between subsequent bid and ask prices may become larger and depth provided at them decreases. The impacts are both econometrically and economically significant. Also, our findings point to an interaction between different measures of liquidity, between liquidity at the best prices and beyond in the book, and between ask and bid side of the market.
Previous evidence suggests that less liquid stocks entail higher average returns. Using NYSE data, we present evidence that both the sensitivity of returns to liquidity and liquidity premia have significantly declined over the past four decades to levels that we cannot statistically distinguish from zero. Furthermore, the profitability of trading strategies based on buying illiquid stocks and selling illiquid stocks has declined over the past four decades, rendering such strategies virtually unprofitable. Our results are robust to several conventional liquidity measures related to volume. When using liquidity measure that is not related to volume, we find just weak evidence of a liquidity premium even in the early periods of our sample. The gradual introduction and proliferation of index funds and exchange traded funds is a possible explanation for these results.
This paper addresses and resolves the issue of microstructure noise when measuring the relative importance of home and U.S. market in the price discovery process of Canadian interlisted stocks. In order to avoid large bounds for information shares, previous studies applying the Cholesky decomposition within the Hasbrouck (1995) framework had to rely on high frequency data. However, due to the considerable amount of microstructure noise inherent in return data at very high frequencies, these estimators are distorted. We offer a modified approach that identifies unique information shares based on distributional assumptions and thereby enables us to control for microstructure noise. Our results indicate that the role of the U.S. market in the price discovery process of Canadian interlisted stocks has been underestimated so far. Moreover, we suggest that rather than stock specific factors, market characteristics determine information shares.
Innovative automated execution strategies like Algorithmic Trading gain significant market share on electronic market venues worldwide, although their impact on market outcome has not been investigated in depth yet. In order to assess the impact of such concepts, e.g. effects on the price formation or the volatility of prices, a simulation environment is presented that provides stylized implementations of algorithmic trading behavior and allows for modeling latency. As simulations allow for reproducing exactly the same basic situation, an assessment of the impact of algorithmic trading models can be conducted by comparing different simulation runs including and excluding a trader constituting an algorithmic trading model in its trading behavior. By this means the impact of Algorithmic Trading on different characteristics of market outcome can be assessed. The results indicate that large volumes to execute by the algorithmic trader have an increasing impact on market prices. On the other hand, lower latency appears to lower market volatility.
Macro announcements change the equilibrium riskfree rate. We find that treasury prices reflect part of the impact instantaneously, but intermediaries rely on their customer order flow in the 15 minutes after the announcement to discover the full impact. We show that this customer flow informativeness is strongest at times when analyst forecasts of macro variables are highly dispersed. We study 30 year treasury futures to identify the customer flow. We further show that intermediaries appear to benefit from privately recognizing informed customer flow, as, in the cross-section, their own-account trade profitability correlates with access to customer orders, controlling for volatility, competition, and the announcement surprise. These results suggest that intermediaries learn about equilibrium riskfree rates through customer orders.
We report evidence that the presence of hidden liquidity is associated with greater liquidity in the order books, greater trading volume, and smaller price impact. Limit and market order submission behavior changes when hidden liquidity is present consistent with at least some traders being able to detect hidden liquidity. We estimate a model of liquidity provision that allows us to measure variations in the marginal and total payoffs from liquidity provision in states with and without hidden liquidity. Our estimates of the expected surplus to providers of visible and hidden liquidity are positive and typically of the order of one-half to one basis points per trade. The positive liquidity provider surpluses combined with the increased trading volume when hidden liquidity is present are both consistent with liquidity externalities.
This paper considers a trading game in which sequentially arriving liquidity traders either opt for a market order or for a limit order. One class of traders is considered to have an extended trading horizon, implying their impatience is linked to their trading orientation. More specifically, sellers are considered to have a trading horizon of two periods, whereas buyers only have a single-period trading scope (the extended buyer-horizon case is completely symmetric). Clearly, as the life span of their submitted limit orders is longer, this setting implies sellers are granted a natural advantage in supplying liquidity. This benefit is hampered, however, by the direct competition arising between consecutively arriving sellers. Closed-form characterizations for the order submission strategies are obtained when solving for the equilibrium of this dynamic game. These allow to examine how these forces affect traders´ order placement decisions. Further, the analysis yields insight into the dynamic process of price formation and into the market clearing process of a non-intermediated, order driven market.
Central counterparties (CCPs) have increasingly become a cornerstone of financial markets infrastructure. We present a model where trades are time-critical, liquidity is limited and there is limited enforcement of trades. We show a CCP novating trades implements efficient trading behaviour. It is optimal for the CCP to face default losses to achieve the efficient level of trade. To cover these losses, the CCP optimally uses margin calls, and, as the default problem becomes more severe, also requires default funds and then imposes position limits.
n the last few years, many of the world’s largest financial exchanges have converted from mutual, not-for-profit organizations to publicly-traded, for-profit firms. In most cases, these exchanges have substantial responsibilities with respect to enforcing various regulations that protect investors from dishonest agents. We examine how the incentives to enforce such regulations change as an exchange converts from mutual to for-profit status. In contrast to oft-stated concerns, we find that, in many circumstances, an exchange that maximizes shareholder (rather than member) income has a greater incentive to aggressively enforce these types of regulations.
The execution, clearing, and settlement of financial transactions are all subject to substantial scale and scope economies which make each of these complementary functions a natural monopoly. Integration of trade, execution, and settlement in an exchange improves efficiency by economizing on transactions costs. When scope economies in clearing are more extensive than those in execution, integration is more costly, and efficient organization involves a trade-off of scope economies and transactions costs. A properly organized clearing cooperative can eliminate double marginalization problems and exploit scope economies, but can result in opportunism and underinvestment. Moreover, a clearing cooperative may exercise market power. Vertical integration and tying can foreclose entry, but foreclosure can be efficient because market power rents attract excessive entry. Integration of trading and post-trade services is the modal form of organization in financial markets, which is consistent with the hypothesis that transactional efficiencies explain organizational arrangements in these markets.
Central counterparties
(2008)
Central counterparties (CCPs) have increasingly become a cornerstone of financial markets infrastructure. We present a model where trades are time-critical, liquidity is limited and there is limited enforcement of trades. We show a CCP novating trades implements efficient trading behaviour. It is optimal for the CCP to face default losses to achieve the efficient level of trade. To cover these losses, the CCP optimally uses margin calls, and, as the default problem becomes more severe, also requires default funds and then imposes position limits.
Algorithmic trading has sharply increased over the past decade. Equity market liquidity has improved as well. Are the two trends related? For a recent five-year panel of New York Stock Exchange (NYSE) stocks, we use a normalized measure of electronic message traffic (order submissions, cancellations, and executions) as a proxy for algorithmic trading, and we trace the associations between liquidity and message traffic. Based on within-stock variation, we find that algorithmic trading and liquidity are positively related. To sort out causality, we use the start of autoquoting on the NYSE as an exogenous instrument for algorithmic trading. Previously, specialists were responsible for manually disseminating the inside quote. As stocks were phased in gradually during early 2003, the manual quote was replaced by a new automated quote whenever there was a change to the NYSE limit order book. This market structure change provides quicker feedback to traders and algorithms and results in more message traffic. For large-cap stocks in particular, quoted and effective spreads narrow under autoquote and adverse selection declines, indicating that algorithmic trading does causally improve liquidity.
We find and describe four futures markets where the bid-ask spread is bid down to the fixed price tick size practically all the time, and which match counterparties using a pro-rata rule. These four markets´ offered depths at the quotes on average exceed mean market order size by two orders of magnitude, and their order cancellation rates (the probability of any given offered lot being cancelled) are significantly over 96 per cent. We develop a simple theoretical model to ex- plain these facts, where strategic complementarities in the choice of limit order size cause traders to risk overtrading by submitting over-sized limit orders, most of which they expect to cancel.
We consider a multi-period rational expectations model in which risk-averse investors differ in their information on past transaction prices (the ticker). Some investors (insiders) observe prices in real-time whereas other investors (outsiders) observe prices with a delay. As prices are informative about the asset payoff, insiders get a strictly larger expected utility than outsiders. Yet, information acquisition by one investor exerts a negative externality on other investors. Thus, investors’ average welfare is maximal when access to price information is rationed. We show that a market for price information can implement the fraction of insiders that maximizes investors’ average welfare. This market features a high price to curb excessive acquisition of ticker information. We also show that informational efficiency is greater when the dissemination of ticker information is broader and more timely.
We show on an abstract level that contextual equivalence in non-deterministic program calculi defined by may- and must-convergence is maximal in the following sense. Using also all the test predicates generated by the Boolean, forall- and existential closure of may- and must-convergence does not change the contextual equivalence. The situation is different if may- and total must-convergence is used, where an expression totally must-converges if all reductions are finite and terminate with a value: There is an infinite sequence of test-predicates generated by the Boolean, forall- and existential closure of may- and total must-convergence, which also leads to an infinite sequence of different contextual equalities.
In this paper, we examine the impact of mergers among German savings banks on the extent to which these savings banks engage in small business lending. The ongoing consolidation in the banking industry has sparked concerns about the continuous availability of credit to small businesses which has been further fueled by empirical studies that partly confirm a reduction in small business lending in the aftermath of mergers. However, using a proprietary data set of German savings banks we find strong evidence that in Germany merging savings banks do not significantly change the extent to which they lend to small businesses compared to prior to the merger or compared to the contemporaneous lending by non-merging banks. We investigate the merger related effects on small business lending in Germany from a bank-level perspective. Furthermore, we estimate small business lending and its continuous adjustment process simultaneously using recent General Method of Moments (GMM) techniques for panel data as proposed by Arellano and Bond (1991).
This paper discusses the so-called commercial approach to microfinance under economic and ethical aspects. It first shows how microfinance has developed from a purely welfare-oriented activity to a commercially relevant line of banking business. The background of this stunning success is the – almost universal – adoption of the so-called commercial approach to microfinance in the course of the last decade. As the author argues, this commercial approach is the only sound approach to adopt if one wanted microfinance to have any social and developmental impact, and therefore the wide-spread “moralistic” criticism of the commercial approach, which has again and again been expressed in the 1990s, is ill-placed from an economic and an ethical perspective. However, some recent events in microfinance raise doubts as to whether the commercial approach has not, in a number of cases, gone too far. The evident example for such a development is the Mexican microfinance institution Compartamos, which recently undertook a financially extremely successful IPO. As it seems, some microfinance institutions have by now become so radically commercial that all of those social and development considerations, which have traditionally motivated work in the field of microfinance, seem to have lost their importance. Thus there is a conflict between commercial and developmental aspirations. However, this conflict is not inevitable. The paper concludes by showing that, and how, a microfinance institution can try to combine using the strengths of the capital market and at the same time maintaining its developmental focus and importance.
We examine insurance markets with two types of customers: those who regret suboptimal decisions and those who don.t. In this setting, we characterize the equilibria under hidden information about the type of customers and hidden action. We show that both pooling and separating equilibria can exist. Furthermore, there exist separating equilibria that predict a positive correlation between the amount of insurance coverage and risk type, as in the standard economic models of adverse selection, but there also exist separating equilibria that predict a negative correlation between the amount of insurance coverage and risk type, i.e. advantageous selection. Since optimal choice of regretful customers depends on foregone alternatives, any equilibrium includes a contract which is o¤ered but not purchased.
When a spot market monopolist participates in a derivatives market, she has an incentive to deviate from the spot market monopoly optimum to make her derivatives market position more profitable. When contracts can only be written contingent on the spot price, a risk-averse monopolist chooses to participate in the derivatives market to hedge her risk, and she reduces expected profits by doing so. However, eliminating all risk is impossible. These results are independent of the shape of the demand function, the distribution of demand shocks, the nature of preferences or the set of derivatives contracts.
Since independence from British colonial rule, Uganda has had a turbulent political history characterised by putsches, dictatorship, contested electoral outcomes, civil wars and a military invasion. There were eight changes of government within a period of twenty-four years (from 1962-1986), five of which were violent and unconstitutional. This paper identifies factors that account for these recurrent episodes of political violence and state collapse. While colonialism bequeathed the country a negative legacy including a weak state apparatus, ethnic division, skewed development, elite polarisation and a narrow economic base, post-colonial leaders have on the whole exacerbated rather than reversed these trends. Factors such as ethnic rivalry, political exclusion, militarisation of politics, weak state institutions, and unequal access to opportunities for self-advancement help to account for the recurrent cycles of violence and state failure prior to 1986. External factors have also been important, particularly the country’s politically turbulent neighbourhood, the outcome of political instability and civil conflict in surrounding countries. Neighbourhood turbulence stemming from such factors as civil wars in Congo and Sudan has had spill-over effects in that it has allowed insurgent groups geographical space within which to operate as well as provided opportunities for the acquisition of instruments of war with which to destabilise the country. Critical to these processes have been the porosity of post-colonial borders and the inability by the Ugandan state to exercise effective control over its entire territory. By demonstrating the interplay between internal and external factors in shaping Uganda’s postcolonial experience, the paper makes an important shift away from conventional explanations that have focused disproportionately on internal processes. Lastly, the paper provides pointers to areas of further research such as the economic foundations of conflict that should ultimately strengthen our understanding of factors that combine to make state-making fail or succeed.
This paper is one of the first to analyse political influence on state-owned savings banks in a developed country with an established financial market: Germany. Combining a large dataset with financial and operating figures of all 457 German savings banks from 1994 to 2006 and information on over 1,250 local elections during this period we investigate the change in business behavior around elections. We find strong indications for political inflence: the probability that savings banks close branches, lay-off employees or engage in merger activities is significantly reduced around elections. At the same time they tend to increase their extraordinary spendings, which include support for social and cultural events in the area, on average by over 15%. Finally, we find that savings banks extend significantly more loans to their corporate and private customers in the run-up to an election. In further analyses, we show that the magnitude of political influence depends on bank specific, economical and political circumstances in the city or county: political influence seems to be facilitated by weak political majorities and profitable banks. Banks in economically weak areas seem to be less prone to political influence.
In this paper, we investigate how bank mergers affect bank revenues and present empirical evidence that mergers among banks have a substantial and persistent negative impact on merging banks’ revenues. We refer to merger related negative effects on banks’ revenues as dissynergies and suggest that they are a result of organizational diseconomies, the loss of customers and the temporary distraction of management from day-to-day operations by effecting the merger. For our analyses we draw on a proprietary data set with detailed financials of all 457 regional savings banks in Germany, which have been involved in 212 mergers between 1994 and 2006. We find that the negative impact of a merger on net operating revenues amounts to 3% of pro-forma consolidated banks’ operating profits and persists not only for the year of the merger but for up to four years post-merger. Only thereafter mergers exhibit a significantly superior performance compared to their respective pre-merger performance or the performance of their non-merging peers. The magnitude and persistence of merger related revenue dissynergies highlight their economic relevance. Previous research on post-merger performance mainly focuses on the effects from mergers on banks’ (cost) efficiency and profitability but fails to provide clear and consistent results. We are the first, to our knowledge, to examine the post-merger performance of banks’ net operating revenues and to empirically verify significant negative implications of mergers for banks’ net operating revenues. We propose that our finding of negative merger related effects on banks’ operating revenues is the reason why previous research fails to show merger related gains.
The "quiet life hypothesis (QLH)" posits that banks enjoy the advantages of market power in terms of foregone revenues or cost savings. We suggest a unified approach to measure competition and efficiency simultaneously to test this hypothesis. We estimate bank-specific Lerner indices as measures of competition and test if cost and profit efficiency are negatively related to market power in the case of German savings banks.We find that both market power and average revenues declined among these banks between 1996 and 2006. While we find clear evidence supporting the QLH, estimated effects of the QLH are small from an economical perspective.
Motivated by the recent discussion of the declining importance of deposits as banks´ major source of funding we investigate which factors determine funding costs at local banks. Using a panel data set of more than 800 German local savings and cooperative banks for the period from 1998 to 2004 we show that funding costs are not only driven by the relative share of comparatively cheap deposits of bank´s liabilities but among other factors especially by the size of the bank. In our empirical analysis we find strong and robust evidence that, ceteris paribus, smaller banks exhibit lower funding costs than larger banks suggesting that small banks are able to attract deposits more cheaply than their larger counterparts. We argue that this is the case because smaller banks interact more personally with customers, operate in customers´ geographic proximity and have longer and stronger relationships than larger banks and, hence, are able to charge higher prices for their services. Our finding of a strong influence of bank size on funding costs is also in an in- ternational context of great interest as mergers among small local banks - the key driver of bank growth - are a recent phenomenon not only in European banking that is expected to continue in the future. At the same time, net interest income remains by far the most important source of revenue for most local banks, accounting for approximately 70% of total operating revenues in the case of German local banks. The influence of size on funding costs is of strong economic relevance: our results suggest that an increase in size by 50%, for example, from EUR 500 million in total assets to EUR 750 million (exemplary for M&A transactions among local banks) increases funding costs, ceteris paribus, by approximately 18 basis points which relates to approx. 7% of banks´ average net interest margin.