Refine
Year of publication
- 2021 (10) (remove)
Document Type
- Article (8)
- Working Paper (2)
Has Fulltext
- yes (10)
Is part of the Bibliography
- no (10)
Keywords
Der Einsatz von Künstliche Intelligenz (KI) – Technologien eröffnet viele Chancen, birgt aber auch viele Risiken – insbesondere in der Finanzbranche. Dieses Whitepaper gibt einen Überblick über den aktuellen Stand der Anwendung und Regulierung von KI-Technologien in der Finanzbranche, und diskutiert Chancen und Risiken von KI. KI findet in der Finanzbranche zahlreiche Anwendungsgebiete. Dazu gehören Chatbots, intelligente Assistenten für Kunden, automatischer Hochfrequenzhandel, automatisierte Betrugserkennung, Überwachung der Compliance, Gesichtserkennungssoftware zur Kundenidentifikation u. v. m. Auch Finanzaufsichtsbehörden setzen zunehmend KI-Anwendungen ein, um große und komplexe Datenmengen (Big Data) automatisiert und skalierbar auf Muster zu untersuchen und ihren Aufsichtspflichten nachzukommen.
Die Regulierung von KI in der Finanzbranche ist ein Balanceakt. Auf der einen Seite gibt es eine Notwendigkeit Flexibilität zu gewährleisten, um Innovationen nicht einzudämmen und im internationalen Wettbewerb nicht abgehängt zu werden. Strenge Auflagen können in diesem Zusammenhang als Barriere für die erfolgreiche Weiter-)Entwicklung von KI-Applikationen in der Finanzbranche wirken. Auf der anderen Seite müssen Persönlichkeitsrechte geschützt und Entscheidungsprozesse nachvollziehbar bleiben. Die fehlende Erklärbarkeit und Interpretierbarkeit von KI-Modellen entsteht in erster Linie durch Intransparenz bei einem Großteil heutiger KI-Anwendungen, bei welchen zwar die Natur der Ein- und Ausgaben beobachtbar und verständlich ist, nicht jedoch die genauen Verarbeitungsschritte dazwischen (Blackbox Prinzip).
Dieses Spannungsfeld zeigt sich auch im aktuellen regulatorischen Ansatz verschiedener Behörden. So werden einerseits die positiven Seiten von KI betont, wie Effizienz- und Effektivitätsgewinne sowie Rentabilitäts- und Qualitätssteigerungen (Bundesregierung, 2019) oder neue Methoden der Gefahrenanalyse in der Finanzmarktregulierung (BaFin, 2018a). Andererseits, wird darauf verwiesen, dass durch KI getroffene Entscheidungen immer von Menschen verantwortet werden müssen (EU Art. 22 DSGVO) und demokratische Rahmenbedingungen des Rechtsstaats zu wahren seien (FinTechRat, 2017).
Für die Zukunft sehen wir die Notwendigkeit internationale Regularien prinzipienbasiert, vereinheitlicht und technologieneutral weiterzuentwickeln, ohne dabei die Entwicklung neuer KIbasierter Geschäftsmodelle zu bremsen. Im globalen Wettstreit sollte Europa bei der Regulierung des KI-Einsatzes eine Vorreiterrolle einnehmen und damit seine demokratischen Werte der digitalen Freiheit, Selbstbestimmung und das Recht auf Information weltweit exportieren. Förderprogramme sollten einen stärkeren Fokus auf die Entwicklung nachhaltiger und verantwortungsvoller KI in Banken legen. Dazu zählt insbesondere die (Weiter-)Entwicklung breit einsetzbarer Methoden, die es erlauben, menschen-interpretierbare Erklärungen für erzeugte Ausgaben bereitzustellen und Problemen wie dem Blackbox Prinzip entgegenzuwirken.
Aus Sicht der Unternehmen in der Finanzbranche könnte eine Kooperation mit BigTech-Unternehmen sinnvoll sein, um gemeinsam das Potential der Technologie bestmöglich ausschöpfen zu können. Nützlich wäre auch ein gemeinsames semantisches Metadatenmodell zur Beschreibung der in der Finanzbranche anfallenden Daten. In Zukunft könnten künstliche Intelligenzen Daten aus sozialen Netzwerken berücksichtigen oder Smart Contracts aushandeln. Eine der größten Herausforderungen der Zukunft wird das Anwerben geeigneten Personals darstellen.
Sample-based longitudinal discrete choice experiments: preferences for electric vehicles over time
(2021)
Discrete choice experiments have emerged as the state-of-the-art method for measuring preferences, but they are mostly used in cross-sectional studies. In seeking to make them applicable for longitudinal studies, our study addresses two common challenges: working with different respondents and handling altering attributes. We propose a sample-based longitudinal discrete choice experiment in combination with a covariate-extended hierarchical Bayes logit estimator that allows one to test the statistical significance of changes. We showcase this method’s use in studies about preferences for electric vehicles over six years and empirically observe that preferences develop in an unpredictable, non-monotonous way. We also find that inspecting only the absolute differences in preferences between samples may result in misleading inferences. Moreover, surveying a new sample produced similar results as asking the same sample of respondents over time. Finally, we experimentally test how adding or removing an attribute affects preferences for the other attributes.
COVID-19 HAS AGAIN TIGHTENED ITS GRIP AROUND THE WORLD AND THE HEALTH SYSTEM. THIS ARTICLE GIVES AN INTRODUCTION TO EXPLAINABLE INTERACTIVE MACHINE LEARNING AND PROVIDES INSIGHTS ON HOW THIS METHOD MAY NOT ONLY HELP IN ENGINEERING MORE POWERFUL AI SYSTEMS, BUT ALSO HOW IT MAY HELP TO EASE THE BURDEN OF VIRAL STRAINS ON THE HEALTHCARE SYSTEM.
The mobile games business is an ever-increasing sub-sector of the entertainment industry. Due to its high profitability but also high risk and competitive atmosphere, game publishers need to develop strategies that allow them to release new products at a high rate, but without compromising the already short lifespan of the firms' existing games. Successful game publishers must enlarge their user base by continually releasing new and entertaining games, while simultaneously motivating the current user base of existing games to remain active for more extended periods. Since the core-component reuse strategy has proven successful in other software products, this study investigates the advantages and drawbacks of this strategy in mobile games. Drawing on the widely accepted Product Life Cycle concept, the study investigates whether the introduction of a new mobile game built with core-components of an existing mobile game curtails the incumbent's product life cycle. Based on real and granular data on the gaming activity of a popular mobile game, the authors find that by promoting multi-homing (i.e., by smartly interlinking the incumbent and new product with each other so that users start consuming both games in parallel), the core-component reuse strategy can prolong the lifespan of the incumbent game.
Device-to-device (D2D) communication is an innovative solution for improving wireless network performance to efficiently handle the ever-increasing mobile data traffic. Communication takes place directly between two devices that are in each other’s transmission range. So far, research has focused on the technical challenges of implementing this technology and assumes a user’s general willingness to participate as forwarder in this technology. However, this simplifying assumption is not realistic, as willingness to participate in D2D communication can vary depending on the user. In this work, we consider the scenario that a user can act as a forwarder for a receiver who is not directly or insufficiently reached by the base station and accordingly has no or poor Internet connection. We take a user-centric approach and investigate the willingness to provide an Internet connection as a forwarder. We are the first to investigate user preferences for D2D communication using a choice-based conjoint analysis. Our results, based on a representative sample of potential users (N=181), show that the social relationship between the potential forwarder and the receiver has the greatest impact on the potential forwarder’s decision to provide an Internet connection to the receiver, accepting sacrifices in terms of additional battery consumption and reduced own service performance. In a detailed segment analysis, we observe significant preference differences depending on smartphone usage behavior and user age. Taking the corresponding preferences into account when matching forwarders and receivers can further increase technology adoption.
This paper explores the interplay of feature-based explainable AI (XAI) tech- niques, information processing, and human beliefs. Using a novel experimental protocol, we study the impact of providing users with explanations about how an AI system weighs inputted information to produce individual predictions (LIME) on users’ weighting of information and beliefs about the task-relevance of information. On the one hand, we find that feature-based explanations cause users to alter their mental weighting of available information according to observed explanations. On the other hand, explanations lead to asymmetric belief adjustments that we inter- pret as a manifestation of the confirmation bias. Trust in the prediction accuracy plays an important moderating role for XAI-enabled belief adjustments. Our results show that feature-based XAI does not only superficially influence decisions but re- ally change internal cognitive processes, bearing the potential to manipulate human beliefs and reinforce stereotypes. Hence, the current regulatory efforts that aim at enhancing algorithmic transparency may benefit from going hand in hand with measures ensuring the exclusion of sensitive personal information in XAI systems. Overall, our findings put assertions that XAI is the silver bullet solving all of AI systems’ (black box) problems into perspective.
Having a gatekeeper position in a collaborative network offers firms great potential to gain competitive advantages. However, it is not well understood what kind of collaborations are associated with such a position. Conceptually grounded in social network theory, this study draws on the resource-based view and the relational factors view to investigate which types of collaboration characterize firms that are in a gatekeeper position, which ultimately could improve firm performance in subsequent periods. The empirical analysis utilizes a unique longitudinal data set to examine dynamic network formation. We used a data crawling approach to reconstruct collaboration networks among the 500 largest companies in Germany over nine years and matched these networks with performance data. The results indicate that firms in gatekeeper positions often engage in medium-intensity collaborations and less likely weak-intensity collaborations. Strong-intensity collaborations are not related to the likelihood of being a gatekeeper. Our study further reveals that a firm's knowledge base is an important moderator and that this knowledge base can increase the benefits of having a gatekeeper position in terms of firm performance.
Artificial Intelligence (AI) and Machine Learning (ML) are currently hot topics in industry and business practice, while management-oriented research disciplines seem reluctant to adopt these sophisticated data analytics methods as research instruments. Even the Information Systems (IS) discipline with its close connections to Computer Science seems to be conservative when conducting empirical research endeavors. To assess the magnitude of the problem and to understand its causes, we conducted a bibliographic review on publications in high-level IS journals. We reviewed 1,838 articles that matched corresponding keyword-queries in journals from the AIS senior scholar basket, Electronic Markets and Decision Support Systems (Ranked B). In addition, we conducted a survey among IS researchers (N = 110). Based on the findings from our sample we evaluate different potential causes that could explain why ML methods are rather underrepresented in top-tier journals and discuss how the IS discipline could successfully incorporate ML methods in research undertakings.