Refine
Year of publication
Document Type
- Working Paper (1310)
- Part of Periodical (282)
- Article (162)
- Report (96)
- Doctoral Thesis (34)
- Conference Proceeding (14)
- Part of a Book (7)
- Book (5)
- Periodical (4)
- Preprint (4)
Language
- English (1924) (remove)
Is part of the Bibliography
- no (1924)
Keywords
- Deutschland (58)
- Financial Institutions (48)
- ECB (42)
- Capital Markets Union (37)
- Banking Union (34)
- monetary policy (34)
- Financial Markets (33)
- Banking Regulation (30)
- Household Finance (28)
- Monetary Policy (28)
Institute
- Wirtschaftswissenschaften (1924) (remove)
I have assessed changes in the monetary policy stance in the euro area since its inception by applying a Bayesian time-varying parameter framework in conjunction with the Hamiltonian Monte Carlo algorithm. I find that the estimated policy response has varied considerably over time. Most of the results suggest that the response weakened after the onset of the financial crisis and while quantitative measures were still in place, although there are also indications that the weakening of the response to the expected inflation gap may have been less pronounced. I also find that the policy response has become more forceful over the course of the recent sharp rise in inflation. Furthermore, it is essential to model the stochastic volatility relating to deviations from the policy rule as it materially influences the results.
This paper presents and compares Bernoulli iterative approaches for solving linear DSGE models. The methods are compared using nearly 100 different models from the Macroeconomic Model Data Base (MMB) and different parameterizations of the monetary policy rule in the medium-scale New Keynesian model of Smets and Wouters (2007) iteratively. I find that Bernoulli methods compare favorably in solving DSGE models to the QZ, providing similar accuracy as measured by the forward error of the solution at a comparable computation burden. The method can guarantee convergence to a particular, e.g., unique stable, solution and can be combined with other iterative methods, such as the Newton method, lending themselves especially to refining solutions.
Can right‐wing terrorism increase support for far‐right populist parties and if so, why? Exploiting quasi‐random variation between successful and failed attacks across German municipalities, we find that successful attacks lead to significant increases in the vote share for the right‐wing, populist Alternative für Deutschland (AfD) party in state elections. Investigating channels, we find that successful attacks lead to differential increases in turnout which are mainly captured by the AfD. Using the German SOEP, a longitudinal panel of individuals, we investigate terror’s impact on individual political attitudes. We first document that people residing in municipalities that experience successful or failed attacks are indistinguishable. We then show that successful terror leads individuals to prefer the AfD, adopt more populist attitudes and report significantly greater political participation at the local level. Terror also leads voters to migrate away from (some) mainstream parties to the AfD. We also find differential media reporting: successful attacks receive more media coverage among local and regional publishers, coverage which makes significantly more use of words related to Islam and terror. Our results hold despite the fact that most attacks are motivated by right‐wing causes and targeted against migrants. Moreover, successful attacks that receive the most media coverage have nearly double the effect on the AfD vote share in state elections and they also increase the AfD vote share in Federal elections, highlighting media salience as a driver of our overall results.
The importance of agile methods has increased in recent years, not only to manage IT projects but also to establish flexible and adaptive organisational structures, which are essential to deal with disruptive changes and build successful digital business strategies. This paper takes an industry-specific perspective by analysing the dissemination, objectives and relative popularity of agile frameworks in the German banking sector. The data provides insights into expectations and experiences associated with agile methods and indicates possible implementation hurdles and success factors. Our research provides the first comprehensive analysis of agile methods in the German banking sector. The comparison with a selected number of fintechs has revealed some differences between banks and fintechs. We found that almost all banks and fintechs apply agile methods in IT projects. However, fintechs have relatively more experience with agile methods than banks and use them more intensively. Scrum is the most relevant framework used in practice. Scaled agile frameworks are so far negligible in the German banking sector. Acceleration of projects is apparently the most important objective of deploying agile methods. In addition, agile methods can contribute to cost savings and lead to improved quality and innovation performance, though for banks it is evidently more challenging to reach their respective targets than for fintechs. Overall our findings suggest that German banks are still in a maturing process of becoming more agile and that there is room for an accelerated adoption of agile methods in general and scaled agile frameworks in particular.
Fabo, Janˇcokov ́a, Kempf, and P ́astor (2021) show that papers written by central bank researchers find quantitative easing (QE) to be more effective than papers written by academics. Weale and Wieladek (2022) show that a subset of these results lose statistical significance when OLS regressions are replaced by regressions that downweight outliers. We examine those outliers and find no reason to downweight them. Most of them represent estimates from influential central bank papers published in respectable academic journals. For example, among the five papers finding the largest peak effect of QE on output, all five are published in high-quality journals (Journal of Monetary Economics, Journal of Money, Credit and Banking, and Applied Economics Letters), and their average number of citations is well over 200. Moreover, we show that these papers have supported policy communication by the world’s leading central banks and shaped the public perception of the effectiveness of QE. New evidence based on quantile regressions further supports the results in Fabo et al. (2021).
This paper examines rent sharing in private investments in public equity (PIPEs) between newly public firms and private investors. The evidence suggests highly asymmetric rent sharing. Newly public firms earn a negative return of up to −15% in the first post-PIPE year, while investors benefit due to the ability to dictate transaction terms. The results are economically relevant because newly public firms are, at least in recent years, more likely to tap private rather than public markets for follow-on financing shortly after the initial public offering (IPO), and because the results for newly public firms contrast with those for the broad PIPE market in Lim et al. (2021). The study also contributes to the PIPE literature by offering an integrative view of competing theories of the cross-section of post-PIPE stock returns. We simultaneously test proxies for corporate governance, asymmetric information, bargaining power, and managerial entrenchment. While all explanations have univariate predictive power for the post-PIPE performance, only the proxies for corporate governance and asymmetric information are robust in ceteris-paribus tests.
Background: Nations are imposing unprecedented measures at large-scale to contain the spread of COVID-19 pandemic. Recent studies indicate that measures such as lockdowns may have slowed down the growth of COVID-19. However, in addition to substantial economic and social costs, these measures also limit the exposure to Ultraviolet-B radiation (UVB). Emerging observational evidence indicate the protective role of UVB and vitamin D in reducing the severity and mortality of COVID-19 deaths. In this observational study, we empirically outline the independent protective roles of lockdown and UVB exposure as measured by ultraviolet index (UVI), whilst also examining whether the severity of lockdown is associated with a reduction in the protective role.
Methods: We apply a log-linear fixed-effects model to a panel dataset of 162 countries over a period of 108 days (n=6049). We use the cumulative number of COVID-19 deaths as the dependent variable and isolate the mitigating influence of lockdown severity on the association between UVI and growth-rates of COVID-19 deaths from time-constant country-specific and time-varying country-specific potentially confounding factors.
Findings: After controlling for time-constant and time-varying factors, we find that a unit increase in UVI and lockdown severity are independently associated with 17% [-1.8 percentage points] and 77% [-7.9 percentage points] decline in COVID-19 deaths growth rate, indicating their respective protective roles. However, the widely utilized and least severe lockdown (recommendation to not leave the house) already fully mitigates the protective role of UVI by 95% [1.8 percentage points] indicating its downside.
Interpretation: We find that lockdown severity and UVI are independently associated with a slowdown in the daily growth rates of cumulative COVID-19 deaths. However, we find consistent evidence that increase in lockdown severity is associated with a significant reduction in the protective role of UVI in reducing COVID-19 deaths. Our results suggest that lockdowns in conjunction with adequate exposure to UVB radiation might have provided even more substantial health benefits, than lockdowns alone. For example, we estimate that there would be 21% fewer deaths on average with sufficient UVB exposure while people were recommended not to leave their house. Therefore, our study outlines the importance of considering UVB exposure, especially while implementing lockdowns and may support policy decision making in countries imposing such measures.
Competing Interest Statement: RKM is a PhD researcher at Goethe University, Frankfurt. He also is an employee of a multinational chemical company involved in vitamin D business and holds the shares of the company. This study is intended to contribute to the ongoing COVID-19 crisis and is not sponsored by his company. All other authors declare no competing interests. The views expressed in the paper are those of the authors and do not represent that of any organization. No other relationships or activities that could appear to have influenced the submitted work.
We use census data to show that structural transformation reflects a fundamental reallocation of labour from goods to services, instead of a relabelling that occurs when goods-producing firms outsource their in-house service production. The novelty of our approach is that it categorizes labour by occupations, which are invariant to outsourcing. We find that the reallocation of labour from goods-producing to service-producing occupations is a robust feature in censuses from around the world and different time periods. To understand the underlying forces, we propose a tractable model in which uneven occupation-specific technological change generates structural transformation of occupation employment.
We propose a novel approach to the study of international trade based on a theory of country integration that embodies a broad systemic viewpoint on the relationship between trade and growth. Our model leads to an indicator of country openness that measures a country's level of integration through the full architecture of its connections in the trade network. We apply our methodology to a sample of 204 countries and find a sizable and significant positive relationship between our integration measure and a country's growth rate, while that of the traditional measures of outward orientation is only minor and statistically insignificant.
The SVB case is a wake-up call for Europe’s regulators as it demonstrates the destructive power of a bank-run: it undermines the role of loss absorbing capital, elbowing governments to bailout affected banks. Many types of bank management weaknesses, like excessive duration risk, may raise concerns of bank losses – but to serve as a run-trigger, there needs to be a large enough group of bank depositors that fails to be fully covered by a deposit insurance scheme. Latent run-risk is the root cause of inefficient liquidations, and we argue that a run on SVB assets could have been avoided altogether by a more thoughtful deposit insurance scheme, sharply distinguishing between loss absorbing capital (equity plus bail-in debt) and other liabilities which are deemed not to be bail-inable, namely demand deposits. These evidence-based insights have direct implications for Europe’s banking regulation, suggesting a minimum and a maximum for a banks’ loss absorption capacity.
Flows of funds run by banks or by firms that belong to the same financial group as a bank are less volatile and less sensitive to bad past performance. This enables bank-affiliated funds to better weather distress and to hold lower precautionary cash buffers in comparison with their unaffiliated peers. Banks provide liquidity support to distressed affiliated funds by buying shares of those funds that are experiencing large outflows. This, in turn, diminishes the severity of strategic complementarities in investors’ redemptions. Liquidity support and other benefits of bank affiliation are conditional on the financial health of the parent company. Distress in the banking system spills over to the mutual fund sector via ownership links. Our research high-lights substantial dependencies between the banking system and the asset management industry, and identifies an important channel via which financial stability risks depend on the organisational structure of the financial sector.
This paper defends The Transformation of Values into Prices on the Basis of Random Systems, published in EIER, by answering to the Comments made in the same journal by Professors Mori, Morioka and Yamazaki. The clarifications mainly concern the justification of the randomness assumptions, the conditions needed to obtain the equality of total profit with total surplus value in the simplified one-industry system and the invariance of the results to changes in the units of measurement.
We contribute to the debate about the future of capital markets and corporate finance, which has ensued against the background of a significant boom in private markets and a corresponding decline in the number of firms and the amount of capital raised in public markets in the US and Europe.
Our research sheds light on the fluctuating significance of public and private markets for corporate finance over time, and challenges the conventional view of a linear progression from one market to the other. We argue instead that a more complex pattern of interaction between public and private markets emerges, after taking a long-term perspective and examining historical developments more closely.
We claim that there is a dynamic divide between these markets, and identify certain factors that determine the degree to which investors, capital, and companies gravitate more towards one market than the other. However, in response to the status quo, other factors will gain momentum and favor the respective other market, leading to a new (unstable) equilibrium. Hence, we observe the oscillating domains of public and private markets over time. While these oscillations imply ‘competition’ between these markets, we unravel the complementarities between them, which also militate against a secular trend towards one market. Finally, we examine the role of regulation in this dynamic divide as well as some policy implications arising from our findings.
Art-related non-fungible tokens (NFTs) took the digital art space by storm in 2021, generating massive amounts of volume and attracting a large number of users to a previously obscure part of blockchain technology. Still, very little is known about the attributes that influence the price of these digital assets. This paper attempts to evaluate the level of speculation associated with art NFTs, comprehend the characteristics that confer value on them and design a profitable trading strategy based on our findings. We analyze 860,067 art NFTs that have been deployed on the Ethereum blockchain and have been involved in 317,950 sales using machine learning methods to forecast the probability of sale, the trade frequency and the average price. We find that NFTs are highly speculative assets and that their price and recurrence of sale are heavily determined by the floor and the last sale prices, independent of any fundamental value.
The discount control mechanisms that closed-end funds often choose to adopt before IPO are supposedly implemented to narrow the difference between share price and net asset value. We find evidence that non-discretionary discount control mechanisms such as mandatory continuation votes serve as costly signals of information to reveal higher fund quality to investors. Rents of the skill signaled through the announcement of such policies accrue to managers rather than investors as differences in skill are revealed through growing assets under management rather than risk- adjusted performance.
We analyze the performance of marketplace lending using loan cash flow data from the largest platform, Lending Club. We find substantial risk-adjusted performance of about 40 basis points per month for the entire loan portfolio. Other loan portfolios grouped by risk category have similar risk-adjusted performance. We show that characteristics of the local bank sector for each loan, such as concentration of deposits and the presence of national banks, are related to the performance of loans. Thus, marketplace lending has the potential to finance a growing share of the consumer credit market in the absence of a competitive response from the traditional incumbents.
Armstrong et al. (2022) review the empirical methods used in the accounting literature to draw causal inferences. They document a growing number of studies using quasi-experimental methods and provide a critical perspective on this trend as well as the use of these methods in the accounting literature. In this discussion, I complement their review by broadening the perspective. I argue for a design-based approach to accounting research that shifts attention from methods to the entire research design. I also discuss why studies that aim to draw causal inferences are important, how these studies fit into the scientific process, and why assessing the strength of the research design is important when evaluating studies and aggregating research findings.
Financial ties between drug companies and medical researchers are thought to bias results published in medical journals. To enable readers to account for such bias, most medical journals require authors to disclose potential conflicts of interest. For such policies to be effective, conflict disclosure must modify readers’ beliefs. We therefore examine whether disclosure of financial ties with industry reduces article citations, indicating a discount. A challenge to estimating this effect is selection as drug companies may seek out higher quality authors as consultants or fund their studies, generating a positive correlation between disclosed conflicts and citations. Our analysis confirms this positive association. Including observable controls for article and author quality attenuates but does not eliminate this relation. To tease out whether other researchers discount articles with conflicts, we perform three tests. First, we show that the positive association is weaker for review articles, which are more susceptible to bias. Second, we examine article recommendations to family physicians by medical experts, who choose from articles that are a priori more homogenous in quality. Here, we find a significantly negative association between disclosure and expert recommendations, consistent with discounting. Third, we conduct an analysis within author and article, exploiting journal policy changes that result in conflict disclosure by an author. We examine the effect of this disclosure on citations to a previously published article by the same author. This analysis reveals a negative citation effect. Overall, we find evidence that disclosures negatively affect citations, consistent with the notion that other researchers discount articles with disclosed conflicts.
Speculative news on corporate takeovers may hurt productivity because uncertainty and threat of job loss cause anxiety, distraction, and reduced collaboration and morale among employees and managers. Using a panel of OECD-headquartered firms, we show that firm productivity temporarily declines upon announcements of speculative takeover rumors that do not materialize. This productivity dip is more pronounced for targets and for firms in countries with weaker employee rights and less long-term orientation. Abnormal stock returns mirror these results. The evidence fosters our understanding of potential real effects of speculative financial news and the costs of takeover threats.
Sample-based longitudinal discrete choice experiments: preferences for electric vehicles over time
(2021)
Discrete choice experiments have emerged as the state-of-the-art method for measuring preferences, but they are mostly used in cross-sectional studies. In seeking to make them applicable for longitudinal studies, our study addresses two common challenges: working with different respondents and handling altering attributes. We propose a sample-based longitudinal discrete choice experiment in combination with a covariate-extended hierarchical Bayes logit estimator that allows one to test the statistical significance of changes. We showcase this method’s use in studies about preferences for electric vehicles over six years and empirically observe that preferences develop in an unpredictable, non-monotonous way. We also find that inspecting only the absolute differences in preferences between samples may result in misleading inferences. Moreover, surveying a new sample produced similar results as asking the same sample of respondents over time. Finally, we experimentally test how adding or removing an attribute affects preferences for the other attributes.
This paper examines the performance of 538 sovereign wealth fund (SWF) investments into venture capital, private equity, and real asset funds (“alternative asset funds”) from 52 countries around the world over the years 1995-2020. The data indicate SWFs are significantly slower to fully liquidate and earn lower returns from their investments, particularly from their investments in venture capital funds. The longer duration and lower performance of SWFs is more pronounced for strategic SWFs than savings SWFs. We show that venture capital fund investments are more likely to be in countries with lower quality disclosure indices. SWFs are more often in buyout funds, and in larger funds with a greater number of limited partners. SWF performance is enhanced by having different types of institutional investors in the same limited partnership. Overall, the data indicate sovereign wealth funds make large investments in alternative asset funds with a longer-term view and earn a lower financial return consistent with strategic and political SWF investment motives.
This paper examines the causes and consequences of hedge fund investments in exchange traded funds (ETFs) using U.S. data from 1998 to 2018. The data indicate that transient hedge funds and quasi-indexer hedge funds are substantially more likely to invest in ETFs. Unexpected hedge fund inflows cause a rise in ETF investments, and the economic significance of unexpected flow is more than twice as large for transient than quasi-indexer hedge funds. ETF investment is in general associated with lower hedge fund performance. But when ETF investment is accompanied by an increase in total flow and unexpected flow, the negative impact of ETF holdings on performance is mitigated. The data are consistent with the view that hedge fund ETF investment unrelated to unexpected flow is an agency cost of delegated portfolio management.
This paper characterizes the stationary equilibrium of a continuous-time neoclassical production economy with capital accumulation in which households can insure against idiosyncratic income risk through long-term insurance contracts. Insurance companies operating in perfectly competitive markets can commit to future contractual obligations, whereas households cannot. For the case in which household labor productivity takes two values, one of which is zero, and where households have logutility we provide a complete analytical characterization of the optimal consumption insurance contract, the stationary consumption distribution and the equilibrium aggregate capital stock and interest rate. Under parameter restrictions, there is a unique stationary equilibrium with partial consumption insurance and a stationary consumption distribution that takes a truncated Pareto form. The unique equilibrium interest rate (capital stock) is strictly decreasing (increasing) in income risk. The paper provides an analytically tractable alternative to the standard incomplete markets general equilibrium model developed in Aiyagari (1994) by retaining its physical structure, but substituting the assumed incomplete asset markets structure with one in which limits to consumption insurance emerge endogenously, as in Krueger and Uhlig (2006).
We analyze efficient risk-sharing arrangements when the value from deviating is determined endogenously by another risk sharing arrangement. Coalitions form to insure against idiosyncratic income risk. Self-enforcing contracts for both the original coalition and any coalition formed (joined) after deviations rely on a belief in future cooperation which we term "trust". We treat the contracting conditions of original and deviation coalitions symmetrically and show that higher trust tightens incentive constraints since it facilitates the formation of deviating coalitions. As a consequence, although trust facilitates the initial formation of coalitions, the extent of risk sharing in successfully formed coalitions is declining in the extent of trust and efficient allocations might feature resource burning or utility burning: trust is indeed a double-edged sword.
Employing the art-collection records of Burton and Emily Hall Tremaine, we consider whether early-stage art investors can be understood as venture capitalists. Because the Tremaines bought artists’ work very close to an artwork’s creation, with 69% of works in our study purchased within one year of the year when they were made, their collecting practice can best be framed as venture-capital investment in art. The Tremaines also illustrate art collecting as social-impact investment, owing to their combined strategy of art sales and museum donations for which the collectors received a tax credit under US rules. Because the Tremaines’ museum donations took place at a time that U.S. marginal tax rates from 70% to 91%, the near “donation parity” with markets, creating a parallel to ESG investment in the management of multiple forms of value.
Venture capital (VC) funds backed by large multi-fund families tend to perform substantially better due to cross-fund cash flows (CFCFs), a liquidity support mechanism provided by matching distributions and capital calls within a VC fund family. The dynamics of this mechanism coincide with the sensitivity of different stage projects owing to market liquidity conditions. We find that the early-stage funds demand relatively more intra-family CFCFs than later-stage funds during liquidity stress periods. We show that the liquidity improvement based on the timing of CFCF allocation reflects how fund families arrange internal liquidity provision and explains a large part of their outperformance.
Lack of privacy due to surveillance of personal data, which is becoming ubiquitous around the world, induces persistent conformity to the norms prevalent under the surveillance regime. We document this channel in a unique laboratory---the widespread surveillance of private citizens in East Germany. Exploiting localized variation in the intensity of surveillance before the fall of the Berlin Wall, we show that, at the present day, individuals who lived in high-surveillance counties are more likely to recall they were spied upon, display more conformist beliefs about society and individual interactions, and are hesitant about institutional and social change. Social conformity is accompanied by conformist economic choices: individuals in high-surveillance counties save more and are less likely to take out credit, consistent with norms of frugality. The lack of differences in risk aversion and binding financial constraints by exposure to surveillance helps to support a beliefs channel.
Supranational supervision
(2022)
We exploit the establishment of a supranational supervisor in Europe (the Single Supervisory Mechanism) to learn how the organizational design of supervisory institutions impacts the enforcement of financial regulation. Banks under supranational supervision are required to increase regulatory capital for exposures to the same firm compared to banks under the local supervisor. Local supervisors provide preferential treatment to larger institutes. The central supervisor removes such biases, which results in an overall standardized behavior. While the central supervisor treats banks more equally, we document a loss in information in banks’ risk models associated with central supervision. The tighter supervision of larger banks results in a shift of particularly risky lending activities to smaller banks. We document lower sales and employment for firms receiving most of their funding from banks that receive a tighter supervisory treatment. Overall, the central supervisor treats banks more equally but has less information about them than the local supervisor.
Industry concentration and markups in the US have been rising over the last 3-4 decades. However, the causes remain largely unknown. This paper uses machine learning on regulatory documents to construct a novel dataset on compliance costs to examine the effect of regulations on market power. The dataset is comprehensive and consists of all significant regulations at the 6-digit NAICS level from 1970-2018. We find that regulatory costs have increased by $1 trillion during this period. We document that an increase in regulatory costs results in lower (higher) sales, employment, markups, and profitability for small (large) firms. Regulation driven increase in concentration is associated with lower elasticity of entry with respect to Tobin's Q, lower productivity and investment after the late 1990s. We estimate that increased regulations can explain 31-37% of the rise in market power. Finally, we uncover the political economy of rulemaking. While large firms are opposed to regulations in general, they push for the passage of regulations that have an adverse impact on small firms.
Resolving financial distress where property rights are not clearly defined: the case of China
(2022)
We use data on financially distressed Chinese companies in order to study a debt market where property rights are crudely defined and poorly enforced. To help with identification we use an event where a business-friendly province published new guidelines regarding the administration and enforcement of assets pledged as collateral. Although by no means a comprehensive reform of bankruptcy law or property rights, by instructing courts to enforce existing, albeit rudimentary, contractual rights the new guidelines virtually eliminated creditors runs and produced a sharp increase in the survival rate of financially-distressed companies. These changes illustrate how piecemeal reforms of property rights and their enforcement may have a significant impact on economic outcomes. Our analysis and results challenge the view that a fully fledged system of private property is a precondition for economic development.
This study examines the recent literature on the expectations, beliefs and perceptions of investors who incorporate Environmental, Social, Governance (ESG) considerations in investment decisions with the aim to generate superior performance and also make a societal impact. Through the lens of equilibrium models of agents with heterogeneous tastes for ESG investments, green assets are expected to generate lower returns in the long run than their non- ESG counterparts. However, at the short run, ESG investment can outperform non-ESG investment through various channels. Empirically, results of ESG outperformance are mixed. We find consensus in the literature that some investors have ESG preference and that their actions can generate positive social impact. The shift towards more sustainable policies in firms is motivated by the increased market values and the lower cost of capital of green firms driven by investors’ choices.
This paper provides a review of the development of the non-fungible tokens (NFTs) market, with a particular focus on its pricing determinants, its current applications and future opportunities. We investigate the current state of the NFT markets and highlight the perception and expectations of investors towards these products. We summarize and compare the financial and econometric models that have been used in the literature for the pricing of non-fungible tokens with a special focus on their predictive performance. Our intention is to design a framework that can help understanding the price formation of NFTs. We further aim to shed light on the value creating determinants of NFTs in order to better understand the investors’ behavior on the blockchain.
Biased auctioneers
(2022)
We construct a neural network algorithm that generates price predictions for art at auction, relying on both visual and non-visual object characteristics. We find that higher automated valuations relative to auction house pre-sale estimates are associated with substantially higher price-to-estimate ratios and lower buy-in rates, pointing to estimates’ informational inefficiency. The relative contribution of machine learning is higher for artists with less dispersed and lower average prices. Furthermore, we show that auctioneers’ prediction errors are persistent both at the artist and at the auction house level, and hence directly predictable themselves using information on past errors.
This paper examines how the implementation of a new dark order - Midpoint Extended Life Order on NASDAQ - impacts financial markets stability in terms of occurrences of mini-flash crashes in individual securities. We use high-frequency order book data and apply panel regression analysis to estimate the effect of M-ELO trading on market stability and liquidity provision. The results suggest a predominance of a speed bump effect of M-ELO rather than a darkness effect. We find that the introduction of M-ELO increases market stability by reducing the average number of mini-flash crashes, but its impact on market quality is mixed.
The right to ask questions and voice their opinions at annual general meetings (AGMs) represents one of the few avenues for shareholders to communicate directly and publicly with the firm’s management. Examining AGM transcripts of U.S. companies between 2007 and 2021, we find that shareholders actively express their concerns about environmental, social and governance (ESG) issues in accordance with their specific relationship with the company. Further, they are also demonstrably more vocal about ESG issues at AGMs of firms with poor sustainability performance. What is more, we show that this soft engagement translates into a more negative tone which, in turn, results in lower approval rates for management proposals. Shareholders' soft engagement at AGMs is hence an effective way to "walk the talk".
The issuance of sustainability-linked loans (SLLs) has grown exponentially in recent years. Using a scoring methodology, we examine the underlying key performance indicators of a large sample of SLLs and analyze whether their design creates effective incentives for improving corporate sustainability performance. We demonstrate that the majority of loans fails to meet key requirements that would make them credible instruments for generating effective sustainability incentives. These findings call into question the actual sustainability impact that may be achieved through the issuance of ESG-linked debt.
Consumers purchase energy in many forms. Sometimes energy goods are consumed directly, for instance, in the form of gasoline used to operate a vehicle, electricity to light a home, or natural gas to heat a home. At other times, the cost of energy is embodied in the prices of goods and services that consumers buy, say when purchasing an airline ticket or when buying online garden furniture made from plastic to be delivered by mail. Previous research has focused on quantifying the pass-through of the price of crude oil or the price of motor gasoline to U.S. inflation. Neither approach accounts for the fact that percent changes in refined product prices need not be proportionate to the percent change in the price of oil, that not all energy is derived from oil, and that the correlation of price shocks across energy markets is far from one. This paper develops a vector autoregressive model that quantifies the joint impact of shocks to several energy prices on headline and core CPI inflation. Our analysis confirms that focusing on gasoline price shocks alone will underestimate the inflationary pressures emanating from the energy sector, but not enough to overturn the conclusion that much of the observed increase in headline inflation in 2021 and 2022 reflected non-energy price shocks.
We propose a new instrument for estimating the price elasticity of gasoline demand that exploits systematic differences across U.S. states in the pass-through of oil price shocks to retail gasoline prices. These differences, which are primarily driven by variation in the cost of producing and distributing gasoline, create cross-sectional dispersion in gasoline price growth in response to an aggregate oil price shock. We find that the elasticity was stable near -0.3 until the end of 2014, but subsequently rose to about -0.2. Our estimates inform the recent debate about gasoline-tax holidays and policies to reduce carbon emissions.
This paper investigates retirees’ optimal purchases of fixed and variable longevity income annuities using their defined contribution (DC) plan assets and given their expected Social Security benefits. As an alternative, we also evaluate using plan assets to boost Social Security benefits through delayed claiming. We determine that including deferred income annuities in DC accounts is welfare enhancing for all sex/education groups examined. We also show that providing access to well-designed variable deferred annuities with some equity exposure further enhances retiree wellbeing, compared to having access only to fixed annuities. Nevertheless, for the least educated, delaying claiming Social Security is preferred, whereas the most educated benefit more from using accumulated DC plan assets to purchase deferred annuities.
We investigate consumption patterns in Europe with supervised machine learning methods and reveal differences in age and wealth impact across countries. Using data from the third wave (2017) of the Eurosystem’s Household Finance and Consumption Survey (HFCS), we assess how age and (liquid) wealth affect the marginal propensity to consume (MPC) in the Netherlands, Germany, France, and Italy. Our regression analysis takes the specification by Christelis et al. (2019) as a starting point. Decision trees are used to suggest alternative variable splits to create categorical variables for customized regression specifications. The results suggest an impact of differing wealth distributions and retirement systems across the studied Eurozone members and are relevant to European policy makers due to joint Eurozone monetary policy and increasing supranational fiscal authority of the EU. The analysis is further substantiated by a supervised machine learning analysis using a random forest and XGBoost algorithm.
Mamma mia! Revealing hidden heterogeneity by PCA-biplot : MPC puzzle for Italy's elderly poor
(2023)
I investigate consumption patterns in Italy and use a PCA-biplot to discover a consumption puzzle for the elderly poor. Data from the third wave (2017) of the Eurosystem’s Household Finance and Consumption Survey (HFCS) indicate that Italian poor old-aged households boast lower levels of the marginal propensity to consume (MPC) than suggested by the dominant consumption models. A customized regression analysis exhibits group differences with richer peers to be only half as large as prescribed by a traditional linear regression model. This analysis has benefited from a visualization technique for high-dimensional matrices related to the unsupervised machine learning literature. I demonstrate that PCA-biplots are a useful tool to reveal hidden relations and to help researchers to formulate simple research questions. The method is presented in detail and suggestions on incorporating it in the econometric modeling pipeline are given.
Fund companies regularly send shareholder letters to their investors. We use textual analysis to investigate whether these letters’ writing style influences fund flows and whether it predicts performance and investment styles. Fund investors react to the tone and content of shareholder letters: A less negative tone leads to higher net flows. Thus, fund companies can use shareholder letters as a tactical instrument to influence flows. However, at the same time, a dishonest communication that is not consistent with the fund’s actual performance decreases flows. A positive writing style predicts higher idiosyncratic risk as well as more style bets, while there is no consistent predictive power for future performance.
Optimal monetary policy studies typically rely on a single structural model and identification of model-specific rules that minimize the unconditional volatilities of inflation and real activity. In their proposed approach, the authors take a large set of structural models and look for the model-robust rules that minimize the volatilities at those frequencies that policymakers are most interested in stabilizing. Compared to the status quo approach, their results suggest that policymakers should be more restrained in their inflation responses when their aim is to stabilize inflation and output growth at specific frequencies. Additional caution is called for due to model uncertainty.
Who should hold bail-inable debt and how can regulators police holding restrictions effectively?
(2023)
This paper analyses the demand-side prerequisites for the efficient application of the bail-in tool in bank resolution, scrutinises whether the European bank crisis management and deposit insurance (CMDI) framework is apt to establish them, and proposes amendments to remedy identified shortcomings.
The first applications of the new European CMDI framework, particularly in Italy, have shown that a bail-in of debt holders is especially problematic if they are households or other types of retail investors. Such debt holders may be unable to bear losses, and the social implications of bailing them in may create incentives for decision makers to refrain from involving them in bank resolution. In turn, however, if investors can expect resolution authorities (RAs) to behave inconsistently over time and bail-out bank capital and debt holders despite earlier vows to involve them in bank rescues, the pricing and monitoring incentives that the crisis management framework seeks to invigorate would vanish. As a result, market discipline would be suboptimal and moral hazard would persist. Therefore, the policy objectives of the CMDI framework will only be achieved if critical bail-in capital is not held by retail investors without sufficient loss-bearing capacity. Currently, neither the CMDI framework nor capital market regulation suffice to assure that this precondition is met. Therefore, some amendments are necessary. In particular, debt instruments that are most likely to absorb losses in resolution should have a high minimum denomination and banks should not be allowed to self-place such securities.
The loan impairment rules recently introduced by IFRS 9 require banks to estimate their future credit losses by using forward-looking information. We use supervisory loan-level data from Germany to investigate how banks apply their reporting discretion and adjust their lending upon the announcement of the new rules. Our identification strategy exploits a cut-off for the level of provisions at the investment grade threshold based on banks’ internal rating of a borrower. We find that banks required to adopt the new rules assign better internal ratings to exactly the same borrowers compared to banks that do not apply IFRS 9 around this cut-off. This pattern is consistent with a strategic use of the increased reporting discretion that is inherent to rules requiring forward-looking loss estimation. At the same time, banks also reduce their lending exposure to exactly those borrowers at the highest risk of experiencing a rating downgrade below the cutoff. These loans would be associated with additional provisions in future periods, both in the intensive and extensive margin. The lending change thus mitigates some of the negative effects of increased reporting opportunism on banks’ crisis resilience. However, when these firms with internal ratings around the investment grade cut-off obtain less external funding through banks, the introduction of IFRS 9 will likely also be associated with real economic effects
Using German and US brokerage data we find that investors are more likely to sell speculative stocks trading at a gain. Investors’ gain realizations are monotonically increasing in a stock’s speculativeness. This translates into a high disposition effect for speculative and a much lower disposition effect for non-speculative stocks. Our findings hold across asset classes (stocks, passive, and active funds) and explain cross-sectional differences in investor selling behavior which previous literature attributed primarily to investor demographics. Our results are robust to rank or attention effects and can be linked to realization utility and rolling mental account.
Recent empirical evidence shows that most international prices are sticky in dollars. This paper studies the policy implications of this fact in the context of an open economy model, allowing for an arbitrary structure of asset markets, general preferences and technologies, time- or state-dependent price setting, and a rich set of shocks. We show that although monetary policy is less efficient and cannot implement the flexible-price allocation, inflation targeting remains robustly optimal in non-U.S. economies. The implementation of this non-cooperative policy results in a "global monetary cycle" with other countries importing the monetary stance of the U.S. The capital controls cannot unilaterally improve the allocation and are useful only when coordinated across countries. Thanks to the dominance of the dollar, the U.S. can extract rents in international goods and asset markets and enjoy a higher welfare than other economies. Although international cooperation benefits other countries by improving global demand for dollar-invoiced goods, it is not in the self-interest of the U.S. and may be hard to sustain.
Output gap revisions can be large even after many years. Real-time reliability tests might therefore be sensitive to the choice of the final output gap vintage that the real-time estimates are compared to. This is the case for the Federal Reserve’s output gap. When accounting for revisions in response to the global financial crisis in the final output gap, the improvement in real-time reliability since the mid-1990s is much smaller than found by Edge and Rudd (Review of Economics and Statistics, 2016, 98(4), 785-791). The negative bias of real-time estimates from the 1980s has disappeared, but the size of revisions continues to be as large as the output gap itself.
The authors systematically analyse how the realtime reliability assessment is affected through varying the final output gap vintage. They find that the largest changes are caused by output gap revisions after recessions. Economists revise their models in response to such events, leading to economically important revisions not only for the most recent years, but reaching back up to two decades. This might improve the understanding of past business cycle dynamics, but decreases the reliability of real-time output gaps ex post.
We have designed and implemented an experimental module in the 2014 Health and Retirement Study to measure older persons' willingness to defer claiming of Social Security benefits. Under the current system’ status quo where delaying claiming boosts eventual benefits, we show that 46% of the respondents would delay claiming and work longer. If respondents were instead offered an actuarially fair lump sum payment instead of higher lifelong benefits, about 56% indicate they would delay claiming. Without a work requirement, the average amount needed to induce delayed claiming is only $60,400, while when part-time work is stipulated, the amount is slightly higher, $66,700. This small difference implies a low utility value of leisure foregone, of under 20% of average household income.
The great financial crisis and the euro area crisis led to a substantial reform of financial safety nets across Europe and – critically – to the introduction of supranational elements. Specifically, a supranational supervisor was established for the euro area, with discrete arrangements for supervisory competences and tasks depending on the systemic relevance of supervised credit institutions. A resolution mechanism was created to allow the frictionless resolution of large financial institutions. This resolution mechanism has been now complemented with a funding instrument.
While much more progress has been achieved than most observers could imagine 12 years ago, the banking union remains unfinished with important gaps and deficiencies. The experience over the past years, especially in the area of crisis management and resolution, has provided impetus for reform discussions, as reflected most lately in the Eurogroup statement of 16 June 2022.
This Policy Insight looks primarily at the current and the desired state of the banking union project. The key underlying question, and the focus here, is the level of ambition and how it is matched with effective legal and regulatory tools. Specifically, two questions will structure the discussions:
What would be a reasonable definition and rationale for a ‘complete’ banking union? And what legal reforms would be required to achieve it?
Banking union is a case of a new remit of EU-level policy that so far has been established on the basis of long pre-existing treaty stipulations, namely, Article 127(6) TFEU (for banking supervision) and Article 114 TFEU (for crisis management and deposit insurance). Could its completion be similarly carried out through secondary law? Or would a more comprehensive overhaul of the legal architecture be required to ensure legal certainty and legitimacy?
Using the negotiation process of the Basel Committee on Banking Supervision (BCBS), this paper studies the way regulators form their positions on regulatory issues in the process of international standard-setting and the consequences on the resultant harmonized framework. Leveraging on leaked voting records and corroborating them using machine learning techniques on publicly available speeches, we construct a unique dataset containing the positions of banks and national regulators on the regulatory initiatives of Basel II and III. We document that the probability of a regulator opposing a specific initiative increases by 30% if their domestic national champion opposes the new rule, particularly when the proposed rule disproportionately affects them. We find the effect is driven by regulators who had prior experience of working in large banks – lending support to the private-interest theories of regulation. Meanwhile smaller banks, even when they collectively have a higher share in the domestic market, do not have any impact on regulators’ stand – providing little support to public-interest theories of regulation. Finally, we show this decision-making process manifests into significant watering down of proposed rules, thereby limiting the potential gains from harmonization of international financial regulation.
The modern tontine : an innovative instrument for longevity risk management in an aging society
(2020)
We investigate whether a historical pension concept, the tontine, yields enough innovative potential to extend and improve the prevailing privately funded pension solutions in a modern way. The tontine basically generates an age-increasing cash flow, which can help to match the increasing financing needs at old ages. In contrast to traditional pension products, however, the tontine generates volatile cash flows, which means that the insurance character of the tontine cannot be guaranteed in every situation. By employing Multi Cumulative Prospect Theory (MCPT) we answer the question to what extent tontines can be a complement to or a substitute for traditional annuities. We find that it is only optimal to invest in tontines for a certain range of initial wealth. In addition, we investigate in how far the tontine size, the volatility of individual liquidity needs and expected mortality rates contribute to the demand for tontines.
The financial sector plays an important role in financing the green transformation. Various regulatory initiatives in the EU aim to improve transparency in relation to the sustainability of financial products and the sustainability of economic activities of non-financial and financial undertakings. For credit institutions, the Green Asset Ratio (GAR) has been established by the European regulatory authorities as a KPI for measuring the proportion of Taxonomy-aligned on-balance-sheet exposure in relation to the total assets. The breakdown of the total GAR by type of counterparty, environmental objective and type of asset provides in-depth information about the sustainability profile of a credit institution. This information, which has not been available to date, may also initiate discussions between management and shareholders or other stakeholders regarding the future sustainability strategy of credit institutions. This paper provides an overview of the regulatory background and the method of calculating the GAR along different dimensions. Finally, the potential benefits and limitations of the GAR are discussed.
Highly interconnected global supply chains make countries vulnerable to supply chain disruptions. The authors estimate the macroeconomic effects of global supply chain shocks for the euro area. Their empirical model combines business cycle variables with data from international container trade.
Using a novel identification scheme, they augment conventional sign restrictions on the impulse responses by narrative information about three episodes: the Tohoku earthquake in 2011, the Suez Canal obstruction in 2021, and the Shanghai backlog in 2022. They show that a global supply chain shock causes a drop in euro area real economic activity and a strong increase in consumer prices. Over a horizon of one year, the global supply chain shock explains about 30% of inflation dynamics. They also use regional data on supply chain pressure to isolate shocks originating in China.
Their results show that supply chain disruptions originating in China are an important driver for unexpected movements in industrial production, while disruptions originating outside China are an especially important driver for the dynamics of consumer prices.
The author proposes a Differential-Independence Mixture Ensemble (DIME) sampler for the Bayesian estimation of macroeconomic models.It allows sampling from particularly challenging, high-dimensional black-box posterior distributions which may also be computationally expensive to evaluate. DIME is a “Swiss Army knife”, combining the advantages of a broad class of gradient-free global multi-start optimizers with the properties of a Monte Carlo Markov chain (MCMC). This includes fast burn-in and convergence absent any prior numerical optimization or initial guesses, good performance for multimodal distributions, a large number of chains (the “ensemble”) running in parallel, an endogenous proposal density generated from the state of the full ensemble, which respects the bounds of the prior distribution. The author shows that the number of parallel chains scales well with the number of necessary ensemble iterations.
DIME is used to estimate the medium-scale heterogeneous agent New Keynesian (“HANK”) model with liquid and illiquid assets, thereby for the first time allowing to also include the households’ preference parameters. The results mildly point towards a less accentuated role of household heterogeneity for the empirical macroeconomic dynamics.
The authors estimate perceptions about the Fed's monetary policy rule from panel data on professional forecasts of interest rates and macroeconomic conditions. The perceived dependence of the federal funds rate on economic conditions is time-varying and cyclical: high during tightening episodes but low during easings. Forecasters update their perceptions about the policy rule in response to monetary policy actions, measured by high-frequency interest rate surprises, suggesting that forecasters have imperfect information about the rule. The perceived rule impacts asset prices crucial for monetary policy transmission, driving how interest rates respond to macroeconomic news and explaining term premia in long-term interest rates.
We employ a proprietary transaction-level dataset in Germany to examine how capital requirements affect the liquidity of corporate bonds. Using the 2011 European Banking Authority capital exercise that mandated certain banks to increase regulatory capital, we find that affected banks reduce their inventory holdings, pre-arrange more trades, and have smaller average trade size. While non-bank affiliated dealers increase their market-making activity, they are unable to bridge this gap - aggregate liquidity declines. Our results are stronger for banks with a higher capital shortfall, for non-investment grade bonds, and for bonds where the affected banks were the dominant market-maker.
We develop a two-sector incomplete markets integrated assessment model to analyze the effectiveness of green quantitative easing (QE) in complementing fiscal policies for climate change mitigation. We model green QE through an outstanding stock of private assets held by a monetary authority and its portfolio allocation between a clean and a dirty sector of production. Green QE leads to a partial crowding out of private capital in the green sector and to a modest reduction of the global temperature by 0.04 degrees of Celsius until 2100. A moderate global carbon tax of 50 USD per tonne of carbon is 4 times more effective.
Many people do not understand the concepts of life expectancy and longevity risk, potentially leading them to under-save for retirement or to not purchase longevity insurance, which in turn could reduce wellbeing at older ages. We investigate alternative ways to increase the salience of both concepts, allowing us to assess whether these change peoples’ perceptions and financial decision making. Using randomly-assigned vignettes providing subjects with information about either life expectancy or longevity, we show that merely prompting people to think about financial decisions changes their perceptions regarding subjective survival probabilities. Moreover, this information also boosts respondents’ interest in saving and demand for longevity insurance. In particular, longevity information influences both subjective survival probabilities and financial decisions, while life expectancy information influences only annuity choices. We provide some evidence that many people are simply unaware of longevity risk.
When the COVID-19 crisis struck, banks using internal-rating based (IRB) models quickly recognized the increase in risk and reduced lending more than banks using a standardized approach. This effect is not driven by borrowers’ quality or by banks in countries with credit booms before the pandemic. The higher risk sensitivity of IRB models does not always result in lower credit provision when risk intensifies. Certain features of the IRB models – the use of a downturn Loss Given Default parameter – can increase banks’ resilience and preserve their intermediation capacity also during downturns. Affected borrowers were not able to fully insulate and decreased corporate investments.
Previous studies document a relationship between gambling activity at the aggregate level and investments in securities with lottery-like features. We combine data on individual gambling consumption with portfolio holdings and trading records to examine whether gambling and trading act as substitutes or complements. We find that gamblers are more likely than the average investor to hold lottery stocks, but significantly less likely than active traders who do not gamble. Our results suggest that gambling behavior across domains is less relevant compared to other portfolio characteristics that predict investing in high-risk and high-skew securities, and that gambling on and off the stock market act as substitutes to satisfy the same need, e.g., sensation seeking.
Crowdfunding platforms offer project initiators the opportunity to acquire funds from the Internet crowd and, therefore, have become a valuable alternative to traditional sources of funding. However, some processes on crowdfunding platforms cause undesirable external effects that influence the funding success of projects. In this context, we focus on the phenomenon of project overfunding. Massively overfunded projects have been discussed to overshadow other crowdfunding projects which in turn receive less funding. We propose a funding redistribution mechanism to internalize these overfunding externalities and to improve overall funding results. To evaluate this concept, we develop and deploy an agent-based model (ABM). This ABM is based on a multi-attribute decision-making approach and is suitable to simulate the dynamic funding processes on a crowdfunding platform. Our evaluation provides evidence that possible modifications of the crowdfunding mechanisms bear the chance to optimize funding results and to alleviate existing flaws.
Colocation services offered by stock exchanges enable market participants to achieve execution costs for large orders that are substantially lower and less sensitive to transacting against high-frequency traders. However, these benefits manifest only for orders executed on the colocated brokers' own behalf, whereas customers' order execution costs are substantially higher. Analyses of individual order executions indicate that customer orders originating from colocated brokers are less actively monitored and achieve inferior execution quality. This suggests that brokers do not make effective use of their technology, possibly due to agency frictions or poor algorithm selection and parameter choice by customers.
The leading premium
(2022)
In this paper, we consider conditional measures of lead-lag relationships between aggregate growth and industry-level cash-flow growth in the US. Our results show that firms in leading industries pay an average annualized return 3.6\% higher than that of firms in lagging industries. Using both time series and cross sectional tests, we estimate an annual pure timing premium ranging from 1.2% to 1.7%. This finding can be rationalized in a model in which (a) agents price growth news shocks, and (b) leading industries provide valuable resolution of uncertainty about the growth prospects of lagging industries.
Advances in Machine Learning (ML) led organizations to increasingly implement predictive decision aids intended to improve employees’ decision-making performance. While such systems improve organizational efficiency in many contexts, they might be a double-edged sword when there is the danger of a system discontinuance. Following cognitive theories, the provision of ML-based predictions can adversely affect the development of decision-making skills that come to light when people lose access to the system. The purpose of this study is to put this assertion to the test. Using a novel experiment specifically tailored to deal with organizational obstacles and endogeneity concerns, we show that the initial provision of ML decision aids can latently prevent the development of decision-making skills which later becomes apparent when the system gets discontinued. We also find that the degree to which individuals 'blindly' trust observed predictions determines the ultimate performance drop in the post-discontinuance phase. Our results suggest that making it clear to people that ML decision aids are imperfect can have its benefits especially if there is a reasonable danger of (temporary) system discontinuances.
Search costs for lenders when evaluating potential borrowers are driven by the quality of the underwriting model and by access to data. Both have undergone radical change over the last years, due to the advent of big data and machine learning. For some, this holds the promise of inclusion and better access to finance. Invisible prime applicants perform better under AI than under traditional metrics. Broader data and more refined models help to detect them without triggering prohibitive costs. However, not all applicants profit to the same extent. Historic training data shape algorithms, biases distort results, and data as well as model quality are not always assured. Against this background, an intense debate over algorithmic discrimination has developed. This paper takes a first step towards developing principles of fair lending in the age of AI. It submits that there are fundamental difficulties in fitting algorithmic discrimination into the traditional regime of anti-discrimination laws. Received doctrine with its focus on causation is in many cases ill-equipped to deal with algorithmic decision-making under both, disparate treatment, and disparate impact doctrine. The paper concludes with a suggestion to reorient the discussion and with the attempt to outline contours of fair lending law in the age of AI.
Many nations incentivize retirement saving by letting workers defer taxes on pension contributions, imposing them when retirees withdraw their funds. Using a dynamic life cycle model, we show how ‘Rothification’ – that is, taxing 401(k) contributions rather than payouts – alters saving, investment, consumption, and Social Security claiming patterns. We find that taxing pension contributions instead of withdrawals leads to delayed retirement, somewhat lower lifetime tax payments, and relatively small reductions in consumption. Indeed, the two tax regimes generate quite similar relative inequality metrics: the relative consumption inequality ratio under TEE is only four percent higher than in the EET case. Moreover, results indicate that the Gini measures are also strikingly similar under the EET and the TEE regimes for lifetime consumption, cash on hand, and 401(k) assets, differing by only 1-4 percent. While tax payments are higher early in life under the TEE regime, they are slightly lower in the long run. Moreover, higher EET tax payments are also accompanied by higher volatility. We therefore find few reasons for policymakers to favor either tax approach on egalitarian or revenue-enhancing grounds.
We analyze how market fragmentation affects market quality of SME and other less actively traded stocks. Compared to large stocks, they are less likely to be traded on multiple venues and show, if at all, low levels of fragmentation. Concerning the impact of fragmentation on market quality, we find evidence for a hockey stick effect: Fragmentation has no effect for infrequently traded stocks, a negative effect on liquidity of slightly more active stocks, and increasing benefits for liquidity of large and actively traded stocks. Consequently, being traded on multiple venues is not necessarily harmful for SME stock market quality.
The authors propose a new method to forecast macroeconomic variables that combines two existing approaches to mixed-frequency data in DSGE models. The first existing approach estimates the DSGE model in a quarterly frequency and uses higher frequency auxiliary data only for forecasting. The second method transforms a quarterly state space into a monthly frequency. Their algorithm combines the advantages of these two existing approaches.They compare the new method with the existing methods using simulated data and real-world data. With simulated data, the new method outperforms all other methods, including forecasts from the standard quarterly model. With real world data, incorporating auxiliary variables as in their method substantially decreases forecasting errors for recessions, but casting the model in a monthly frequency delivers better forecasts in normal times.
Correction to: Computational Economics https://doi.org/10.1007/s10614-020-10061-x
The original publication has been updated. In the original publication of this article, under the Introduction heading section, the corrections to the second paragraph’s inline equation were not incorporated. The author’s additional corrections have also been incorporated. The publisher apologizes for the error made during production.
Search costs for lenders when evaluating potential borrowers are driven by the quality of the underwriting model and by access to data. Both have undergone radical change over the last years, due to the advent of big data and machine learning. For some, this holds the promise of inclusion and better access to finance. Invisible prime applicants perform better under AI than under traditional metrics. Broader data and more refined models help to detect them without triggering prohibitive costs. However, not all applicants profit to the same extent. Historic training data shape algorithms, biases distort results, and data as well as model quality are not always assured. Against this background, an intense debate over algorithmic discrimination has developed. This paper takes a first step towards developing principles of fair lending in the age of AI. It submits that there are fundamental difficulties in fitting algorithmic discrimination into the traditional regime of anti-discrimination laws. Received doctrine with its focus on causation is in many cases ill-equipped to deal with algorithmic decision-making under both, disparate treatment, and disparate impact doctrine. The paper concludes with a suggestion to reorient the discussion and with the attempt to outline contours of fair lending law in the age of AI.
We investigate the impact of uneven transparency regulation across countries and industries on the location of economic activity. Using two distinct sources of regulatory variation—the varying extent of financial-reporting requirements and the staggered introduction of electronic business registers in Europe—, we consistently document that direct exposure to transparency regulation is negatively associated with the focal industry’s economic activity in terms of inputs (e.g., employment) and outputs (e.g., production). By contrast, we find that indirect exposure to supplier and customer industries’ transparency regulation is positively associated with the focal industry’s economic activity. Our evidence suggests uneven transparency regulation can reallocate economic activity from regulated toward unregulated countries and industries, distorting the location of economic activity.
To ensure the credibility of market discipline induced by bail-in, neither retail investors nor peer banks should appear prominently among the investor base of banks’ loss absorbing capital. Empirical evidence on bank-level data provided by the German Federal Financial Supervisory Authority raises a few red flags. Our list of policy recommendations encompasses disclosure policy, data sharing among supervisors, information transparency on holdings of bail-inable debt for all stakeholders, threshold values, and a well-defined upper limit for any bail-in activity. This document was provided by the Economic Governance Support Unit at the request of the ECON Committee.
European banks have substantial investments in assets that are
measured without directly observable market prices (mark-to-
model). Financial disclosures of these value estimates lack
standardization and are hard to compare across banks. These
comparability concerns are concentrated in large European
banks that extensively rely on level 3 estimates with the most
unobservable inputs. Although the relevant balance sheet
positions only represent a small fraction of these large banks’
total assets (2.9%), their value equals a significant fraction of core
equity tier 1 (48.9%). Incorrect valuations thus have a potential to
impact financial stability. 85% of these bank assets are under
direct ECB supervision. Prudential regulation requires value
adjustments that are apt to shield capital against valuation risk.
Yet, stringent enforcement is critical for achieving this objective.
This document was provided by the Economic Governance
Support Unit at the request of the ECON Committee.
Linear rational-expectations models (LREMs) are conventionally "forwardly" estimated as follows. Structural coefficients are restricted by economic restrictions in terms of deep parameters. For given deep parameters, structural equations are solved for "rational-expectations solution" (RES) equations that determine endogenous variables. For given vector autoregressive (VAR) equations that determine exogenous variables, RES equations reduce to reduced-form VAR equations for endogenous variables with exogenous variables (VARX). The combined endogenous-VARX and exogenous-VAR equations comprise the reduced-form overall VAR (OVAR) equations of all variables in a LREM. The sequence of specified, solved, and combined equations defines a mapping from deep parameters to OVAR coefficients that is used to forwardly estimate a LREM in terms of deep parameters. Forwardly-estimated deep parameters determine forwardly-estimated RES equations that Lucas (1976) advocated for making policy predictions in his critique of policy predictions made with reduced-form equations.
Sims (1980) called economic identifying restrictions on deep parameters of forwardly-estimated LREMs "incredible", because he considered in-sample fits of forwardly-estimated OVAR equations inadequate and out-of-sample policy predictions of forwardly-estimated RES equations inaccurate. Sims (1980, 1986) instead advocated directly estimating OVAR equations restricted by statistical shrinkage restrictions and directly using the directly-estimated OVAR equations to make policy predictions. However, if assumed or predicted out-of-sample policy variables in directly-made policy predictions differ significantly from in-sample values, then, the out-of-sample policy predictions won't satisfy Lucas's critique.
If directly-estimated OVAR equations are reduced-form equations of underlying RES and LREM-structural equations, then, identification 2 derived in the paper can linearly "inversely" estimate the underlying RES equations from the directly-estimated OVAR equations and the inversely-estimated RES equations can be used to make policy predictions that satisfy Lucas's critique. If Sims considered directly-estimated OVAR equations to fit in-sample data adequately (credibly) and their inversely-estimated RES equations to make accurate (credible) out-of-sample policy predictions, then, he should consider the inversely-estimated RES equations to be credible. Thus, inversely-estimated RES equations by identification 2 can reconcile Lucas's advocacy for making policy predictions with RES equations and Sims's advocacy for directly estimating OVAR equations.
The paper also derives identification 1 of structural coefficients from RES coefficients that contributes mainly by showing that directly estimated reduced-form OVAR equations can have underlying LREM-structural equations.
Short sale bans may improve market quality during crises: new evidence from the 2020 Covid crash
(2022)
In theory, banning short selling stabilizes stock prices but undermines pricing efficiency and has ambiguous impacts on market liquidity. Empirical studies find mixed and conflicting results. This paper leverages cross-country policy variation during the 2020 Covid crisis to assess differential impacts of bans on stock liquidity, prices, and volatility. Results suggest that bans improved liquidity and stabilized prices for illiquid stocks but temporarily diminished liquidity for highly liquid stocks.The findings support theories in which short sale bans may improve liquidity by selectively filtering out informed— potentially predatory—traders. Thus, policies that target the most illiquid stocks may deliver better overall market quality than uniform short sale bans imposed on all stocks.
With open banking, consumers take greater control over their own financial data and share it at their discretion. Using a rich set of loan application data from the largest German FinTech lender in consumer credit, this paper studies what characterizes borrowers who share data and assesses its impact on loan application outcomes. I show that riskier borrowers share data more readily, which subsequently leads to an increase in the probability of loan approval and a reduction in interest rates. The effects hold across all credit risk profiles but are the most pronounced for borrowers with lower credit scores (a higher increase in loan approval rate) and higher credit scores (a larger reduction in interest rate). I also find that standard variables used in credit scoring explain substantially less variation in loan application outcomes when customers share data. Overall, these findings suggest that open banking improves financial inclusion, and also provide policy implications for regulators engaged in the adoption or extension of open banking policies.
With free delivery of products virtually being a standard in E-commerce, product returns pose a major challenge for online retailers and society. For retailers, product returns involve significant transportation, labor, disposal, and administrative costs. From a societal perspective, product returns contribute to greenhouse gas emissions and packaging disposal and are often a waste of natural resources. Therefore, reducing product returns has become a key challenge. This paper develops and validates a novel smart green nudging approach to tackle the problem of product returns during customers’ online shopping processes. We combine a green nudge with a novel data enrichment strategy and a modern causal machine learning method. We first run a large-scale randomized field experiment in the online shop of a German fashion retailer to test the efficacy of a novel green nudge. Subsequently, we fuse the data from about 50,000 customers with publicly-available aggregate data to create what we call enriched digital footprints and train a causal machine learning system capable of optimizing the administration of the green nudge. We report two main findings: First, our field study shows that the large-scale deployment of a simple, low-cost green nudge can significantly reduce product returns while increasing retailer profits. Second, we show how a causal machine learning system trained on the enriched digital footprint can amplify the effectiveness of the green nudge by “smartly” administering it only to certain types of customers. Overall, this paper demonstrates how combining a low-cost marketing instrument, a privacy-preserving data enrichment strategy, and a causal machine learning method can create a win-win situation from both an environmental and economic perspective by simultaneously reducing product returns and increasing retailers’ profits.
Financial literacy affects wealth accumulation, and pension planning plays a key role in this relationship. In a large field experiment, we employ a digital pension aggregation tool to confront a treatment group with a simplified overview of their current pension claims across all pillars of the pension system. We combine survey and administrative bank data to measure the effects on actual saving behavior. Access to the tool decreases pension uncertainty for treated individuals. Average savings increase - especially for the financially less literate. We conclude that simplification of pension information can potentially reduce disparities in pension planning and savings behavior.
This paper utilizes a comprehensive worker-firm panel for the Netherlands to quantifythe impact of ICT capital-skill complementarity on the finance wage premium after the Global Financial Crisis. We apply additive worker and firm fixed-effect models to account for unobserved worker- and firm-heterogeneity and show that firm fixed-effects correct for a downward bias in the estimated finance wage premium. Our results indicate a sizable finance wage premium for both fixed- and full-hourly wages. The complementarity between ICT capital spending and the share of high skill workers at the firm-level reduces the full-wage premium considerably and the fixed-wage premium almost entirely.
India has recorded 142,186 deaths over 36 administrative regions placing India third in the world after the US and Brazil for COVID-19 deaths as of 12 December 2020. Studies indicate that south-west monsoon season plays a role in the dynamics of contagious diseases, which tend to peak post-monsoon season. Recent studies show that vitamin D and its primary source Ultraviolet-B (UVB) radiation may play a protective role in mitigating COVID-19 deaths. However, the combined roles of the monsoon season and UVB radiation in COVID-19 in India remain still unclear. In this observational study, we empirically study the respective roles of monsoon season and UVB radiation, whilst further exploring, whether the monsoon season negatively impacts the protective role of UVB radiation in COVID-19 deaths in India. We use a log-linear Mundlak model to a panel dataset of 36 administrative regions in India from 14 March 2020–19 November 2020 (n = 6751). We use the cumulative COVID-19 deaths as the dependent variable. We isolate the association of monsoon season and UVB radiation as measured by Ultraviolet Index (UVI) from other confounding time-constant and time-varying region-specific factors. After controlling for various confounding factors, we observe that a unit increase in UVI and the monsoon season are separately associated with 1.2 percentage points and 7.5 percentage points decline in growth rates of COVID-19 deaths in the long run. These associations translate into substantial relative changes. For example, a permanent unit increase of UVI is associated with a decrease of growth rates of COVID-19 deaths by 33% (= − 1.2 percentage points) However, the monsoon season, mitigates the protective role of UVI by 77% (0.92 percentage points). Our results indicate a protective role of UVB radiation in mitigating COVID-19 deaths in India. Furthermore, we find evidence that the monsoon season is associated with a significant reduction in the protective role of UVB radiation. Our study outlines the roles of the monsoon season and UVB radiation in COVID-19 in India and supports health-related policy decision making in India.
Shares of open-end real estate funds are typically traded directly between the investor and the fund management company. However, we provide empirical evidence for the growth of secondary market activities, i.e., the trading of shares on stock exchanges. We find high trading levels in situations where the fund management company suspends the issue or redemption of shares. Shares trade at a discount when the fund management company suspends the redemption, whereas shares trade at a premium when the fund management company suspends the issue. We also find evidence that secondary market trading activity is increasing since German regulation introduced a minimum holding period and a mandatory notice period for open-end real estate funds.
Consider two independent random walks. By chance, there will be spells of association between them where the two processes move in the same direction, or in opposite direction. We compute the probabilities of the length of the longest spell of such random association for a given sample size, and discuss measures like mean and mode of the exact distributions. We observe that long spells (relative to small sample sizes) of random association occur frequently, which explains why nonsense correlation between short independent random walks is the rule rather than the exception. The exact figures are compared with approximations. Our finite sample analysis as well as the approximations rely on two older results popularized by Révész (Stat Pap 31:95–101, 1990, Statistical Papers). Moreover, we consider spells of association between correlated random walks. Approximate probabilities are compared with finite sample Monte Carlo results.