AIS Logo
Living knowledge for digital leadership
All AI Governance & Ethics Digital Transformation & Innovation Supply Chain & Operations AI Adoption & Implementation Platform Ecosystems & Strategy SME & Entrepreneurship Cybersecurity & Risk AI Applications & Technologies Digital Health & Well-being Digital Work & Collaboration Education & Training
Structural Estimation of Auction Data through Equilibrium Learning and Optimal Transport
International Conference on Wirtschaftsinformatik (2025)

Structural Estimation of Auction Data through Equilibrium Learning and Optimal Transport

Markus Ewert and Martin Bichler
This study proposes a new method for analyzing auction data to understand bidders' private valuations. It extends an existing framework by reformulating the estimation challenge as an optimal transport problem, which avoids the statistical limitations of traditional techniques. This novel approach uses a proxy equilibrium model to analytically evaluate bid distributions, leading to more accurate and robust estimations.

Problem Designing profitable auctions, such as setting an optimal reserve price, requires knowing how much bidders are truly willing to pay, but this information is hidden. Existing methods to estimate these valuations from observed bids often suffer from statistical biases and inaccuracies, especially with limited data, leading to poor auction design and lost revenue for sellers.

Outcome - The proposed optimal transport-based estimator consistently outperforms established kernel-based techniques, showing significantly lower error in estimating true bidder valuations.
- The new method is more robust, providing accurate estimates even in scenarios with high variance in bidding behavior where traditional methods fail.
- In practical tests, reserve prices set using the new method's estimates led to significant revenue gains for the auctioneer, while prices derived from older methods resulted in zero revenue.
Structural Estimation, Auctions, Equilibrium Learning, Optimal Transport, Econometrics
A Case Study on Large Vehicles Scheduling for Railway Infrastructure Maintenance: Modelling and Sensitivity Analysis
International Conference on Wirtschaftsinformatik (2025)

A Case Study on Large Vehicles Scheduling for Railway Infrastructure Maintenance: Modelling and Sensitivity Analysis

Jannes Glaubitz, Thomas Wolff, Henry Gräser, Philipp Sommerfeldt, Julian Reisch, David Rößler-von Saß, and Natalia Kliewer
This study presents an optimization-driven approach to scheduling large vehicles for preventive railway infrastructure maintenance, using real-world data from Deutsche Bahn. It employs a greedy heuristic and a Mixed Integer Programming (MIP) model to evaluate key factors influencing scheduling efficiency. The goal is to provide actionable insights for strategic decision-making and improve operational management.

Problem Railway infrastructure maintenance is a critical operational task that often causes significant disruptions, delays, and capacity restrictions for both passenger and freight services. These disruptions reduce the overall efficiency and attractiveness of the railway system. The study addresses the challenge of optimizing maintenance schedules to maximize completed work while minimizing interference with regular train operations.

Outcome - The primary bottleneck in maintenance scheduling is the limited availability and reusability of pre-defined work windows ('containers'), not the number of maintenance vehicles.
- Increasing scheduling flexibility by allowing work containers to be booked multiple times dramatically improves maintenance completion rates, from 84.7% to 98.2%.
- Simply adding more vehicles to the fleet provides only marginal improvements, as scheduling efficiency is the limiting factor.
- Increasing the operational radius for vehicles from depots and moderately extending shift lengths can further improve maintenance coverage.
- The analysis suggests that large, predefined maintenance containers are often inefficient and should be split into smaller sections to improve flexibility and resource utilization.
Railway Track Maintenance Planning, Maintenance Track Possession Problem, Operations Research, Mixed Integer Programming, Vehicle Scheduling, Sensitivity Analysis, Optimization
The Role of Generative AI in P2P Rental Platforms: Investigating the Effects of Timing and Interactivity on User Reliance in Content (Co-)Creation Processes
International Conference on Wirtschaftsinformatik (2025)

The Role of Generative AI in P2P Rental Platforms: Investigating the Effects of Timing and Interactivity on User Reliance in Content (Co-)Creation Processes

Niko Spatscheck, Myriam Schaschek, Christoph Tomitza, and Axel Winkelmann
This study investigates how Generative AI can best assist users on peer-to-peer (P2P) rental platforms like Airbnb in writing property listings. Through an experiment with 244 participants, the researchers tested how the timing of when AI suggestions are offered and the level of interactivity (automatic vs. user-prompted) influence how much a user relies on the AI.

Problem While Generative AI offers a powerful way to help property hosts create compelling listings, platforms don't know the most effective way to implement these tools. It's unclear if AI assistance is more impactful at the beginning or end of the writing process, or if users prefer to actively ask for help versus receiving it automatically. This study addresses this knowledge gap to provide guidance for designing better AI co-writing assistants.

Outcome - Offering AI suggestions earlier in the writing process significantly increases how much users rely on them.
- Allowing users to actively prompt the AI for assistance leads to a slightly higher reliance compared to receiving suggestions automatically.
- Higher cognitive load (mental effort) reduces a user's reliance on AI-generated suggestions.
- For businesses like Airbnb, these findings suggest that AI writing tools should be designed to engage users at the very beginning of the content creation process to maximize their adoption and impact.
Human-genAI collaboration, Co-writing, P2P rental platforms, Reliance, Generative AI, Cognitive Load
A Framework for Context-Specific Theorizing on Trust and Reliance in Collaborative Human-AI Decision-Making Environments
International Conference on Wirtschaftsinformatik (2025)

A Framework for Context-Specific Theorizing on Trust and Reliance in Collaborative Human-AI Decision-Making Environments

Niko Spatscheck
This study analyzes 59 empirical research papers to understand why findings on human trust in AI have been inconsistent. It synthesizes this research into a single framework that identifies the key factors influencing how people decide to trust and rely on AI systems for decision-making. The goal is to provide a more unified and context-aware understanding of the complex relationship between humans and AI.

Problem Effective collaboration between humans and AI is often hindered because people either trust AI too much (overreliance) or too little (underreliance), leading to poor outcomes. Existing research offers conflicting explanations for this behavior, creating a knowledge gap for developers and organizations. This study addresses the problem that prior research has largely ignored the specific context—such as the user's expertise, the AI's design, and the nature of the task—which is crucial for explaining these inconsistencies.

Outcome - The study created a comprehensive framework that categorizes the factors influencing trust and reliance on AI into three main groups: human-related (e.g., user expertise, cognitive biases), AI-related (e.g., performance, explainability), and decision-related (e.g., risk, complexity).
- It concludes that trust is not static but is dynamically shaped by the interaction of these various contextual factors.
- This framework provides a practical tool for researchers and businesses to better predict how users will interact with AI and to design systems that foster appropriate levels of trust, leading to better collaborative performance.
AI Systems, Trust, Reliance, Collaborative Decision-Making, Human-AI Collaboration, Contextual Factors, Conceptual Framework
Unveiling Location-Specific Price Drivers: A Two-Stage Cluster Analysis for Interpretable House Price Predictions
International Conference on Wirtschaftsinformatik (2025)

Unveiling Location-Specific Price Drivers: A Two-Stage Cluster Analysis for Interpretable House Price Predictions

Paul Gümmer, Julian Rosenberger, Mathias Kraus, Patrick Zschech, and Nico Hambauer
This study proposes a novel machine learning approach for house price prediction using a two-stage clustering method on 43,309 German property listings from 2023. The method first groups properties by location and then refines these groups with additional property features, subsequently applying interpretable models like linear regression (LR) or generalized additive models (GAM) to each cluster. This balances predictive accuracy with the ability to understand the model's decision-making process.

Problem Predicting house prices is difficult because of significant variations in local markets. Current methods often use either highly complex 'black-box' models that are accurate but hard to interpret, or overly simplistic models that are interpretable but fail to capture the nuances of different market segments. This creates a trade-off between accuracy and transparency, making it difficult for real estate professionals to get reliable and understandable property valuations.

Outcome - The two-stage clustering approach significantly improved prediction accuracy compared to models without clustering.
- The mean absolute error was reduced by 36% for the Generalized Additive Model (GAM/EBM) and 58% for the Linear Regression (LR) model.
- The method provides deeper, cluster-specific insights into how different features, like construction year and living space, affect property prices in different local markets.
- By segmenting the market, the model reveals that price drivers vary significantly across geographical locations and property types, enhancing market transparency for buyers, sellers, and analysts.
House Pricing, Cluster Analysis, Interpretable Machine Learning, Location-Specific Predictions
The PV Solution Guide: A Prototype for a Decision Support System for Photovoltaic Systems
International Conference on Wirtschaftsinformatik (2025)

The PV Solution Guide: A Prototype for a Decision Support System for Photovoltaic Systems

Chantale Lauer, Maximilian Lenner, Jan Piontek, and Christian Murlowski
This study presents the conceptual design of the 'PV Solution Guide,' a user-centric prototype for a decision support system for homeowners considering photovoltaic (PV) systems. The prototype uses a conversational agent and 3D modeling to adapt guidance to specific house types and the user's level of expertise. An initial evaluation compared the prototype's usability and trustworthiness against an established tool.

Problem Current online tools and guides for homeowners interested in PV systems are often too rigid, failing to accommodate unique home designs or varying levels of user knowledge. Information is frequently scattered, incomplete, or biased, leading to consumer frustration, distrust, and decision paralysis, which ultimately hinders the adoption of renewable energy.

Outcome - The study developed the 'PV Solution Guide,' a prototype decision support system designed to be more adaptive and user-friendly than existing tools.
- In a comparative evaluation, the prototype significantly outperformed the established 'Solarkataster Rheinland-Pfalz' tool in usability, with a System Usability Scale (SUS) score of 80.21 versus 56.04.
- The prototype also achieved a higher perceived trust score (82.59% vs. 76.48%), excelling in perceived benevolence and competence.
- Key features contributing to user trust and usability included transparent cost structures, personalization based on user knowledge and housing, and an interactive 3D model of the user's home.
Decision Support Systems, Photovoltaic Systems, Human-Centered Design, Qualitative Research
Designing AI-driven Meal Demand Prediction Systems
International Conference on Wirtschaftsinformatik (2025)

Designing AI-driven Meal Demand Prediction Systems

Alicia Cabrejas Leonhardt, Maximilian Kalff, Emil Kobel, and Max Bauch
This study outlines the design of an Artificial Intelligence (AI) system for predicting meal demand, with a focus on the airline catering industry. Through interviews with various stakeholders, the researchers identified key system requirements and developed nine fundamental design principles. These principles were then consolidated into a feasible system architecture to guide the development of effective forecasting tools.

Problem Inaccurate demand forecasting creates significant challenges for industries like airline catering, leading to a difficult balance between waste and customer satisfaction. Overproduction results in high costs and food waste, while underproduction causes lost sales and unhappy customers. This paper addresses the need for a more precise, data-driven approach to forecasting to improve sustainability, reduce costs, and enhance operational efficiency.

Outcome - The research identified key requirements for AI-driven demand forecasting systems based on interviews with industry experts.
- Nine core design principles were established to guide the development of these systems, focusing on aspects like data integration, sustainability, modularity, transparency, and user-centric design.
- A feasible system architecture was proposed that consolidates all nine principles, demonstrating a practical path for implementation.
- The findings provide a framework for creating advanced AI tools that can improve prediction accuracy, reduce food waste, and support better decision-making in complex operational environments.
meal demand prediction, forecasting methodology, customer choice behaviour, supervised machine learning, design science research
Analyzing German Parliamentary Speeches: A Machine Learning Approach for Topic and Sentiment Classification
International Conference on Wirtschaftsinformatik (2025)

Analyzing German Parliamentary Speeches: A Machine Learning Approach for Topic and Sentiment Classification

Lukas Pätz, Moritz Beyer, Jannik Späth, Lasse Bohlen, Patrick Zschech, Mathias Kraus, and Julian Rosenberger
This study investigates political discourse in the German parliament (the Bundestag) by applying machine learning to analyze approximately 28,000 speeches from the last five years. The researchers developed and trained two separate models to classify the topic and the sentiment (positive or negative tone) of each speech. These models were then used to identify trends in topics and sentiment across different political parties and over time.

Problem In recent years, Germany has experienced a growing public distrust in political institutions and a perceived divide between politicians and the general population. While much political discussion is analyzed from social media, understanding the formal, unfiltered debates within parliament is crucial for transparency and for assessing the dynamics of political communication. This study addresses the need for tools to systematically analyze this large volume of political speech to uncover patterns in parties' priorities and rhetorical strategies.

Outcome - Debates are dominated by three key policy areas: Economy and Finance, Social Affairs and Education, and Foreign and Security Policy, which together account for about 70% of discussions.
- A party's role as either government or opposition strongly influences its tone; parties in opposition use significantly more negative language than those in government, and this tone shifts when their role changes after an election.
- Parties on the political extremes (AfD and Die Linke) consistently use a much higher percentage of negative language compared to centrist parties.
- Parties tend to be most critical (i.e., use more negative sentiment) when discussing their own core policy areas, likely as a strategy to emphasize their priorities and the need for action.
- The developed machine learning models proved highly effective, demonstrating that this computational approach is a feasible and valuable method for large-scale analysis of political discourse.
Natural Language Processing, German Parliamentary, Discourse Analysis, Bundestag, Machine Learning, Sentiment Analysis, Topic Classification
Challenges and Mitigation Strategies for AI Startups: Leveraging Effectuation Theory in a Dynamic Environment
International Conference on Wirtschaftsinformatik (2025)

Challenges and Mitigation Strategies for AI Startups: Leveraging Effectuation Theory in a Dynamic Environment

Marleen Umminger, Alina Hafner
This study investigates the unique benefits and obstacles encountered by Artificial Intelligence (AI) startups. Through ten semi-structured interviews with founders in the DACH region, the research identifies key challenges and applies effectuation theory to explore effective strategies for navigating the uncertain and dynamic high-tech field.

Problem While investment in AI startups is surging, founders face unique challenges related to data acquisition, talent recruitment, regulatory hurdles, and intense competition. Existing literature often groups AI startups with general digital ventures, overlooking the specific difficulties stemming from AI's complexity and data dependency, which creates a need for tailored mitigation strategies.

Outcome - AI startups face core resource challenges in securing high-quality data, accessing affordable AI models, and hiring skilled technical staff like CTOs.
- To manage costs, founders often use publicly available data, form partnerships with customers for data access, and start with open-source or low-cost MVP models.
- Founders navigate competition by tailoring solutions to specific customer needs and leveraging personal networks, while regulatory uncertainty is managed by either seeking legal support or framing compliance as a competitive advantage to attract enterprise customers.
- Effectuation theory proves to be a relevant framework, as successful founders tend to leverage existing resources and networks (bird-in-hand), form strategic partnerships (crazy quilt), and adapt flexibly to unforeseen events (lemonade) rather than relying on long-term prediction.
Artificial intelligence, Entrepreneurial challenge, Effectuation theory, Qualitative research, AI startups, Mitigation strategies
BPMN4CAI: A BPMN Extension for Modeling Dynamic Conversational AI
International Conference on Wirtschaftsinformatik (2025)

BPMN4CAI: A BPMN Extension for Modeling Dynamic Conversational AI

Björn-Lennart Eger, Daniel Rose, and Barbara Dinter
This study develops and evaluates a standard-compliant extension for Business Process Model and Notation (BPMN) called BPMN4CAI. Using a Design Science Research methodology, the paper creates a framework that systematically extends existing BPMN elements to better model the dynamic and context-sensitive interactions of Conversational AI systems. The applicability of the BPMN4CAI framework is demonstrated through a case study in the insurance industry.

Problem Conversational AI systems like chatbots are increasingly integrated into business processes, but the standard modeling language, BPMN, is designed for predictable, deterministic processes. This creates a gap, as traditional BPMN cannot adequately represent the dynamic, context-aware dialogues and flexible decision-making inherent to modern AI. Businesses lack a standardized method to formally and accurately model processes involving these advanced AI agents.

Outcome - The study successfully developed BPMN4CAI, an extension to the standard BPMN, which allows for the formal modeling of Conversational AI in business processes.
- The new extension elements (e.g., Conversational Task, AI Decision Gateway, Human Escalation Event) facilitate the representation of adaptive decision-making, context management, and transparent interactions.
- A proof-of-concept demonstrated that BPMN4CAI improves model clarity and provides a semantic bridge for technical implementation compared to standard BPMN.
- The evaluation also identified limitations, noting that modeling highly dynamic, non-deterministic process paths and visualizing complex context transfers remains a challenge.
Conversational AI, BPMN, Business Process Modeling, Chatbots, Conversational Agent
Generative Al in Business Process Optimization: A Maturity Analysis of Business Applications
International Conference on Wirtschaftsinformatik (2025)

Generative Al in Business Process Optimization: A Maturity Analysis of Business Applications

Ralf Mengele
This study analyzes the current state of Generative AI (GAI) in the business world by systematically reviewing scientific literature. It identifies where GAI applications have been explored or implemented across the value chain and evaluates the maturity of these use cases. The goal is to provide managers and researchers with a clear overview of which business areas can already benefit from GAI and which require further development.

Problem While Generative AI holds enormous potential for companies, its recent emergence means it is often unclear where the technology can be most effectively applied. Businesses lack a comprehensive, systematic overview that evaluates the maturity of GAI use cases across different business processes, making it difficult to prioritize investment and adoption.

Outcome - The most mature and well-researched applications of Generative AI are in product development and in maintenance and repair within the manufacturing sector.
- The manufacturing segment as a whole exhibits the most mature GAI use cases compared to other parts of the business value chain.
- Technical domains show a higher level of GAI maturity and successful implementation than process areas dominated by interpersonal interactions, such as marketing and sales.
- GAI models like Generative Adversarial Networks (GANs) are particularly mature, proving highly effective for tasks like generating synthetic data for early damage detection in machinery.
- Research into GAI is still in its early stages for many business areas, with fields like marketing, sales, and human resources showing low implementation and maturity.
Generative AI, Business Processes, Optimization, Maturity Analysis, Literature Review, Manufacturing
Designing Scalable Enterprise Systems: Learning From Digital Startups
International Conference on Wirtschaftsinformatik (2025)

Designing Scalable Enterprise Systems: Learning From Digital Startups

Richard J. Weber, Max Blaschke, Maximilian Kalff, Noah Khalil, Emil Kobel, Oscar A. Ulbricht, Tobias Wuttke, Thomas Haskamp, and Jan vom Brocke
This study investigates how to design enterprise systems (ES) suitable for the rapidly changing needs of digital startups. Using a design science research approach involving 11 startups, the researchers identified key system requirements and developed nine design principles to create ES that are flexible, adaptable, and scalable.

Problem Traditional enterprise systems are often rigid, assuming business processes are stable and standardized. This design philosophy clashes with the needs of dynamic digital startups, which require highly adaptable systems to support continuous process evolution and rapid growth.

Outcome - The study identified core requirements for enterprise systems in startups, highlighting the need for agility, speed, and minimal overhead to support early-stage growth.
- Nine key design principles for scalable ES were developed, focusing on automation, integration, data-driven decision-making, flexibility, and user-centered design.
- A proposed ES architecture emphasizes a modular approach with a central workflow engine, enabling systems to adapt and scale with the startup.
- The research concludes that for startups, ES design must prioritize process adaptability and transparency over the rigid reliability typical of traditional systems.
Enterprise systems, Business process management, Digital entrepreneurship
Perbaikan Proses Bisnis Onboarding Pelanggan di PT SEVIMA Menggunakan Heuristic Redesign
Jurnal SISFO (2025)

Perbaikan Proses Bisnis Onboarding Pelanggan di PT SEVIMA Menggunakan Heuristic Redesign

Ribka Devina Margaretha, Mahendrawathi ER, Sugianto Halim
This study addresses challenges in PT SEVIMA's customer onboarding process, where Account Managers (AMs) were not always aligned with client needs. Using a Business Process Management (BPM) Lifecycle approach combined with heuristic principles (Resequencing, Specialize, Control Addition, and Empower), the research redesigns the existing workflow. The goal is to improve the matching of AMs to clients, thereby increasing onboarding efficiency and customer satisfaction.

Problem PT SEVIMA, an IT startup for the education sector, struggled with an inefficient customer onboarding process. The primary issue was the frequent mismatch between the assigned Account Manager's skills and the specific, technical needs of the new client, leading to implementation delays and decreased satisfaction.

Outcome - Recommends grouping Account Managers (AMs) based on specialization profiles built from post-project evaluations.
- Suggests moving the initial client needs survey to occur before an AM is assigned to ensure a better match.
- Proposes involving the technical migration team earlier in the process to align strategies from the start.
- These improvements aim to enhance onboarding efficiency, reduce rework, and ultimately increase client satisfaction.
Business Process Redesign, Customer Onboarding, Knowledge-Intensive Process, Heuristics Method, Startup, BPM Lifecycle
Successfully Organizing AI Innovation Through Collaboration with Startups
MIS Quarterly Executive (2023)

Successfully Organizing AI Innovation Through Collaboration with Startups

Jana Oehmichen, Alexander Schult, John Qi Dong
This study examines how established firms can successfully partner with Artificial Intelligence (AI) startups to foster innovation. Based on an in-depth analysis of six real-world AI implementation projects across two startups, the research identifies five key challenges and provides corresponding recommendations for navigating these collaborations effectively.

Problem Established companies often lack the specialized expertise needed to leverage AI technologies, leading them to partner with startups. However, these collaborations introduce unique difficulties, such as assessing a startup's true capabilities, identifying high-impact AI applications, aligning commercial interests, and managing organizational change, which can derail innovation efforts.

Outcome - Challenge 1: Finding the right AI startup. Firms should overcome the inscrutability of AI startups by assessing credible quality signals, such as investor backing, academic achievements of staff, and success in prior contests, rather than relying solely on product demos.
- Challenge 2: Identifying the right AI use case. Instead of focusing on data availability, companies should collaborate with startups in workshops to identify use cases with the highest potential for value creation and business impact.
- Challenge 3: Agreeing on commercial terms. To align incentives and reduce information asymmetry, contracts should include performance-based or usage-based compensation, linking the startup's payment to the value generated by the AI solution.
- Challenge 4: Considering the impact on people. Firms must manage user acceptance by carefully selecting the degree of AI autonomy, involving employees in the design process, and clarifying the startup's role to mitigate fears of job displacement.
- Challenge 5: Overcoming implementation roadblocks. Depending on the company's organizational maturity, it should either facilitate deep collaboration between the startup and all internal stakeholders or use the startup to build new systems that bypass internal roadblocks entirely.
Artificial Intelligence, AI Innovation, Corporate-startup collaboration, Open Innovation, Digital Transformation, AI Startups
How Siemens Democratized Artificial Intelligence
MIS Quarterly Executive (2023)

How Siemens Democratized Artificial Intelligence

Benjamin van Giffen, Helmuth Ludwig
This paper presents an in-depth case study on how the global technology company Siemens successfully moved artificial intelligence (AI) projects from pilot stages to full-scale, value-generating applications. The study analyzes Siemens' journey through three evolutionary stages, focusing on the concept of 'AI democratization', which involves integrating the unique skills of domain experts, data scientists, and IT professionals. The findings provide a framework for how other organizations can build the necessary capabilities to adopt and scale AI technologies effectively.

Problem Many companies invest in artificial intelligence but struggle to progress beyond small-scale prototypes and pilot projects. This failure to scale prevents them from realizing the full business value of AI. The core problem is the difficulty in making modern AI technologies broadly accessible to employees, which is necessary to identify, develop, and implement valuable applications across the organization.

Outcome - Siemens successfully scaled AI by evolving through three stages: 1) Tactical AI pilots, 2) Strategic AI enablement, and 3) AI democratization for business transformation.
- Democratizing AI, defined as the collaborative integration of domain experts, data scientists, and IT professionals, is crucial for overcoming key adoption challenges such as defining AI tasks, managing data, accepting probabilistic outcomes, and addressing 'black-box' fears.
- Key initiatives that enabled this transformation included establishing a central AI Lab to foster co-creation, an AI Academy for upskilling employees, and developing a global AI platform to support scaling.
- This approach allowed Siemens to transform manufacturing processes with predictive quality control and create innovative healthcare products like the AI-Rad Companion.
- The study concludes that democratizing AI creates value by rooting AI exploration in deep domain knowledge and reduces costs by creating scalable infrastructures and processes.
Artificial Intelligence, AI Democratization, Digital Transformation, Organizational Capability, Case Study, AI Adoption, Siemens
How Boards of Directors Govern Artificial Intelligence
MIS Quarterly Executive (2023)

How Boards of Directors Govern Artificial Intelligence

Benjamin van Giffen, Helmuth Ludwig
This study investigates how corporate boards of directors oversee and integrate Artificial Intelligence (AI) into their governance practices. Based on in-depth interviews with high-profile board members from diverse industries, the research identifies common challenges and provides examples of effective strategies for board-level AI governance.

Problem Despite the transformative impact of AI on the business landscape, the majority of corporate boards struggle to understand its implications and their role in governing it. This creates a significant gap, as boards have a fiduciary responsibility to oversee strategy, risk, and investment related to critical technologies, yet AI is often not a mainstream boardroom topic.

Outcome - Identified four key groups of board-level AI governance issues: Strategy and Firm Competitiveness, Capital Allocation, AI Risks, and Technology Competence.
- Boards should ensure AI is integrated into the company's core business strategy by evaluating its impact on the competitive landscape and making it a key topic in annual strategy meetings.
- Effective capital allocation involves encouraging AI experimentation, securing investments in foundational AI capabilities, and strategically considering external partnerships and acquisitions.
- To manage risks, boards must engage with experts, integrate AI-specific risks into Enterprise Risk Management (ERM) frameworks, and address ethical, reputational, and legal challenges.
- Enhancing technology competence requires boards to develop their own AI literacy, review board and committee composition for relevant expertise, and include AI competency in executive succession planning.
AI governance, board of directors, corporate governance, artificial intelligence, strategic management, risk management, technology competence
Evolution of the Metaverse
MIS Quarterly Executive (2023)

Evolution of the Metaverse

Mary Lacity, Jeffrey K. Mullins, Le Kuai
This paper explores the potential opportunities and risks of the emerging metaverse for business and society through an interview format with leading researchers. The study analyzes the current state of metaverse technologies, their potential business applications, and critical considerations for governance and ethical implementation for IT practitioners.

Problem Following renewed corporate interest and massive investment, the concept of the metaverse has generated significant hype, but businesses lack clarity on its definition, tangible value, and long-term impact. This creates uncertainty for leaders about how to approach the technology, differentiate it from past virtual worlds, and navigate the significant risks of surveillance, data privacy, and governance.

Outcome - The business value of the metaverse centers on providing richer, safer experiences for customers and employees, reducing costs, and meeting organizational goals through applications like immersive training, virtual collaboration, and digital twins.
- Companies face a critical choice between centralized 'Web 2' platforms, which monetize user data, and decentralized 'Web 3' models that offer users more control over their digital assets and identity.
- The metaverse can improve employee onboarding, training for dangerous tasks, and collaboration, offering a greater sense of presence than traditional videoconferencing.
- Key challenges include the lack of a single, interoperable metaverse (which is likely over a decade away), limited current capabilities of decentralized platforms, and the potential for negative consequences like addiction and surveillance.
- Businesses are encouraged to explore potential use cases, participate in creating open standards, and consider both the immense promise and potential perils before making significant investments.
Metaverse, Virtual Worlds, Augmented Reality, Web 3.0, Digital Twin, Business Strategy, Governance
How Germany Successfully Implemented Its Intergovernmental FLORA System
MIS Quarterly Executive (2025)

How Germany Successfully Implemented Its Intergovernmental FLORA System

Julia Amend, Simon Feulner, Alexander Rieger, Tamara Roth, Gilbert Fridgen, and Tobias Guggenberger
This paper presents a case study on Germany's implementation of FLORA, a blockchain-based IT system designed to manage the intergovernmental processing of asylum seekers. It analyzes how the project navigated legal and technical challenges across different government levels. Based on the findings, the study offers three key recommendations for successfully deploying similar complex, multi-agency IT systems in the public sector.

Problem Governments face significant challenges in digitalizing services that require cooperation across different administrative layers, such as federal and state agencies. Legal mandates often require these layers to maintain separate IT systems, which complicates data exchange and modernization. Germany's asylum procedure previously relied on manually sharing Excel-based lists between agencies, a process that was slow, error-prone, and created data privacy risks.

Outcome - FLORA replaced inefficient Excel-based lists with a decentralized system, enabling a more efficient and secure exchange of procedural information between federal and state agencies.
- The system created a 'single procedural source of truth,' which significantly improved the accuracy, completeness, and timeliness of information for case handlers.
- By streamlining information exchange, FLORA reduced the time required for initial stages of the asylum procedure by up to 50%.
- The blockchain-based architecture enhanced legal compliance by reducing procedural errors and providing a secure way to manage data that adheres to strict GDPR privacy requirements.
- The study recommends that governments consider decentralized IT solutions to avoid the high hidden costs of centralized systems, deploy modular solutions to break down legacy architectures, and use a Software-as-a-Service (SaaS) model to lower initial adoption barriers for agencies.
intergovernmental IT systems, digital government, blockchain, public sector innovation, case study, asylum procedure, Germany
Load More Showing 90 of 101