AIS Logo
Living knowledge for digital leadership
All AI Governance & Ethics Digital Transformation & Innovation Supply Chain & Operations AI Adoption & Implementation Platform Ecosystems & Strategy SME & Entrepreneurship Cybersecurity & Risk AI Applications & Technologies Digital Health & Well-being Digital Work & Collaboration Education & Training
Systematizing Different Types of Interfaces to Interact with Data Trusts
International Conference on Wirtschaftsinformatik (2025)

Systematizing Different Types of Interfaces to Interact with Data Trusts

David Acev, Florian Rieder, Dennis M. Riehle, and Maria A. Wimmer
This study conducts a systematic literature review to analyze the various types of interfaces used for interaction with Data Trusts, which are organizations that manage data on behalf of others. The research categorizes these interfaces into human-system (e.g., user dashboards) and system-system (e.g., APIs) interactions. The goal is to provide a clear classification and highlight existing gaps in research to support the future implementation of trustworthy Data Trusts.

Problem As the volume of data grows, there is an increasing need for trustworthy data sharing mechanisms like Data Trusts. However, for these trusts to function effectively, the interactions between data providers, users, and the trust itself must be seamless and standardized. The problem is a lack of clear understanding and systematization of the different interfaces required, which creates ambiguity and hinders the development of reliable and interoperable Data Trust ecosystems.

Outcome - The study categorizes interfaces for Data Trusts into two primary groups: Human-System Interfaces (user interfaces like GUIs, CLIs) and System-System Interfaces (technical interfaces like APIs).
- A significant gap exists in the current literature, which often lacks specific details and clear definitions for how these interfaces are implemented within Data Trusts.
- The research highlights a scarcity of standardized and interoperable technical interfaces, which is crucial for ensuring trustworthy and efficient data sharing.
- The paper concludes that developing robust, well-defined interfaces is a vital and foundational step for building functional and widely adopted Data Trusts.
Data Trust, user interface, API, interoperability, data sharing
Understanding How Freelancers in the Design Domain Collaborate with Generative Artificial Intelligence
International Conference on Wirtschaftsinformatik (2025)

Understanding How Freelancers in the Design Domain Collaborate with Generative Artificial Intelligence

Fabian Helms, Lisa Gussek, and Manuel Wiesche
This study explores how generative AI (GenAI), specifically text-to-image generation (TTIG) systems, impacts the creative work of freelance designers. Through qualitative interviews with 10 designers, the researchers conducted a thematic analysis to understand the nuances of this new form of human-AI collaboration.

Problem While the impact of GenAI on creative fields is widely discussed, there is little specific research on how it affects freelance designers. This group is uniquely vulnerable to technological disruption due to their direct market exposure and lack of institutional support, creating an urgent need to understand how these tools are changing their work processes and job security.

Outcome - The research identified four key tradeoffs freelancers face when using GenAI: creativity can be enhanced (inspiration) but also risks becoming generic (standardization).
- Efficiency is increased, but this can be undermined by 'overprecision', a form of perfectionism where too much time is spent on minor AI-driven adjustments.
- The interaction with AI is viewed dually: either as a helpful 'sparring partner' for ideas or as an unpredictable tool causing a frustrating lack of control.
- For the future of work, GenAI is seen as forcing a job transition where designers must adapt new skills, while also posing a direct threat of job loss, particularly for junior roles.
Generative Artificial Intelligence, Online Freelancing, Human-AI collaboration, Freelance designers, Text-to-image generation, Creative process
Extracting Explanatory Rationales of Activity Relationships using LLMs - A Comparative Analysis
International Conference on Wirtschaftsinformatik (2025)

Extracting Explanatory Rationales of Activity Relationships using LLMs - A Comparative Analysis

Kerstin Andree, Zahi Touqan, Leon Bein, and Luise Pufahl
This study investigates using Large Language Models (LLMs) to automatically extract and classify the reasons (explanatory rationales) behind the ordering of tasks in business processes from text. The authors compare the performance of various LLMs and four different prompting techniques (Vanilla, Few-Shot, Chain-of-Thought, and a combination) to determine the most effective approach for this automation.

Problem Understanding why business process steps occur in a specific order (due to laws, business rules, or best practices) is crucial for process improvement and redesign. However, this information is typically buried in textual documents and must be extracted manually, which is a very expensive and time-consuming task for organizations.

Outcome - Few-Shot prompting, where the model is given a few examples, significantly improves classification accuracy compared to basic prompting across almost all tested LLMs.
- The combination of Few-Shot learning and Chain-of-Thought reasoning also proved to be a highly effective approach.
- Interestingly, smaller and more cost-effective LLMs (like GPT-4o-mini) achieved performance comparable to or even better than larger models when paired with sophisticated prompting techniques.
- The findings demonstrate that LLMs can successfully automate the extraction of process knowledge, making advanced process analysis more accessible and affordable for organizations with limited resources.
Activity Relationships Classification, Large Language Models, Explanatory Rationales, Process Context, Business Process Management, Prompt Engineering
Gender Bias in LLMs for Digital Innovation: Disparities and Fairness Concerns
International Conference on Wirtschaftsinformatik (2025)

Gender Bias in LLMs for Digital Innovation: Disparities and Fairness Concerns

Sumin Kim-Andres¹ and Steffi Haag¹
This study investigates gender bias in large language models (LLMs) like ChatGPT within the context of digital innovation and entrepreneurship. Using two tasks—associating gendered terms with professions and simulating venture capital funding decisions—the researchers analyzed ChatGPT-4o's outputs to identify how societal gender biases are reflected and reinforced by AI.

Problem As businesses increasingly integrate AI tools for tasks like brainstorming, hiring, and decision-making, there's a significant risk that these systems could perpetuate harmful gender stereotypes. This can create disadvantages for female entrepreneurs and innovators, potentially widening the existing gender gap in technology and business leadership.

Outcome - ChatGPT-4o associated male-denoting terms with digital innovation and tech-related professions significantly more often than female-denoting terms.
- In simulated venture capital scenarios, the AI model exhibited 'in-group bias,' predicting that both male and female venture capitalists would be more likely to fund entrepreneurs of their own gender.
- The study confirmed that LLMs can perpetuate gender bias through implicit cues like names alone, even when no explicit gender information is provided.
- The findings highlight the risk of AI reinforcing stereotypes in professional decision-making, which can limit opportunities for underrepresented groups in business and innovation.
Gender Bias, Large Language Models, Fairness, Digital Innovation, Artificial Intelligence
Using Large Language Models for Healthcare Data Interoperability: A Data Mediation Pipeline to Integrate Heterogeneous Patient-Generated Health Data and FHIR
International Conference on Wirtschaftsinformatik (2025)

Using Large Language Models for Healthcare Data Interoperability: A Data Mediation Pipeline to Integrate Heterogeneous Patient-Generated Health Data and FHIR

Torben Ukena, Robin Wagler, and Rainer Alt
This study explores the use of Large Language Models (LLMs) to streamline the integration of diverse patient-generated health data (PGHD) from sources like wearables. The researchers propose and evaluate a data mediation pipeline that combines an LLM with a validation mechanism to automatically transform various data formats into the standardized Fast Healthcare Interoperability Resources (FHIR) format.

Problem Integrating patient-generated health data from various devices into clinical systems is a major challenge due to a lack of interoperability between different data formats and hospital information systems. This data fragmentation hinders clinicians' ability to get a complete view of a patient's health, potentially leading to misinformed decisions and obstacles to patient-centered care.

Outcome - LLMs can effectively translate heterogeneous patient-generated health data into the valid, standardized FHIR format, significantly improving healthcare data interoperability.
- Providing the LLM with a few examples (few-shot prompting) was more effective than providing it with abstract rules and guidelines (reasoning prompting).
- The inclusion of a validation and self-correction loop in the pipeline is crucial for ensuring the LLM produces accurate and standard-compliant output.
- While successful with text-based data, the LLM struggled to accurately aggregate values from complex structured data formats like JSON and CSV, leading to lower semantic accuracy in those cases.
FHIR, semantic interoperability, large language models, hospital information system, patient-generated health data
Acceptance Analysis of the Metaverse: An Investigation in the Paper- and Packaging Industry
International Conference on Wirtschaftsinformatik (2025)

Acceptance Analysis of the Metaverse: An Investigation in the Paper- and Packaging Industry

First Author¹, Second Author¹, Third Author¹,², and Fourth Author²
This study investigates employee acceptance of metaverse technologies within the traditionally conservative paper and packaging industry. Using the Technology Acceptance Model 3, the research was conducted as a living lab experiment in a leading packaging company. The methodology combined qualitative content analysis with quantitative multiple regression modelling to assess the key factors influencing adoption.

Problem While major technology companies are heavily investing in the metaverse for workplace applications, there is a significant research gap concerning employee acceptance of these immersive technologies. This is particularly relevant for traditionally non-digital industries, like paper and packaging, which are seeking to digitalize but face unique adoption barriers. This study addresses the lack of empirical data on how employees in such sectors perceive and accept metaverse tools for work and collaboration.

Outcome - Employees in the paper and packaging industry show a moderate but ambiguous acceptance of the metaverse, with an average score of 3.61 out of 5.
- The most significant factors driving acceptance are the perceived usefulness (PU) of the technology for their job and its perceived ease of use (PEU).
- Job relevance was found to be a key influencer of perceived usefulness, while an employee's confidence in their own computer skills (computer self-efficacy) was a key predictor for perceived ease of use.
- While employees recognized benefits like improved virtual collaboration, they also raised concerns about hardware limitations (e.g., headset weight, image clarity) and the technology's overall maturity compared to existing tools.
Metaverse, Technology Acceptance Model 3, Living lab, Paper and Packaging industry, Workplace
Generative AI Usage of University Students: Navigating Between Education and Business
International Conference on Wirtschaftsinformatik (2025)

Generative AI Usage of University Students: Navigating Between Education and Business

Fabian Walke, Veronika Föller
This study investigates how university students who also work professionally use Generative AI (GenAI) in both their academic and business lives. Using a grounded theory approach, the researchers interviewed eleven part-time students from a distance learning university to understand the characteristics, drivers, and challenges of their GenAI usage.

Problem While much research has explored GenAI in education or in business separately, there is a significant gap in understanding its use at the intersection of these two domains. Specifically, the unique experiences of part-time students who balance professional careers with their studies have been largely overlooked.

Outcome - GenAI significantly enhances productivity and learning for students balancing work and education, helping with tasks like writing support, idea generation, and summarizing content.
- Students express concerns about the ethical implications, reliability of AI-generated content, and the risk of academic misconduct or being falsely accused of plagiarism.
- A key practical consequence is that GenAI tools like ChatGPT are replacing traditional search engines for many information-seeking tasks due to their speed and directness.
- The study highlights a strong need for universities to provide clear guidelines, regulations, and formal training on using GenAI effectively and ethically.
- User experience is a critical factor; a positive, seamless interaction with a GenAI tool promotes continuous usage, while a poor experience diminishes willingness to use it.
Artificial Intelligence, ChatGPT, Enterprise, Part-time students, Generative AI, Higher Education
Designing for Digital Inclusion: Iterative Enhancement of a Process Guidance User Interface for Senior Citizens
International Conference on Wirtschaftsinformatik (2025)

Designing for Digital Inclusion: Iterative Enhancement of a Process Guidance User Interface for Senior Citizens

Michael Stadler, Markus Noeltner, Julia Kroenung
This study developed and tested a user interface designed to help senior citizens use online services more easily. Using a travel booking website as a case study, the researchers combined established design principles with a step-by-step visual guide and refined the design over three rounds of testing with senior participants.

Problem As more essential services like banking, shopping, and booking appointments move online, many senior citizens face significant barriers to participation due to complex and poorly designed interfaces. This digital divide can lead to both technological and social disadvantages for the growing elderly population, a problem many businesses fail to address.

Outcome - A structured, visual process guide significantly helps senior citizens navigate and complete online tasks.
- Iteratively refining the user interface based on direct feedback from seniors led to measurable improvements in performance, with users completing tasks faster in each subsequent round.
- Simple design adaptations, such as reducing complexity, using clear instructions, and ensuring high-contrast text, effectively reduce the cognitive load on older users.
- The findings confirm that designing digital services with seniors in mind is crucial for creating a more inclusive digital world and can help businesses reach a larger customer base.
Usability for Seniors, Process Guidance, Digital Accessibility, Digital Inclusion, Senior Citizens, Heuristic Evaluation, User Interface Design
The GenAI Who Knew Too Little – Revisiting Transactive Memory Systems in Human GenAI Collaboration
International Conference on Wirtschaftsinformatik (2025)

The GenAI Who Knew Too Little – Revisiting Transactive Memory Systems in Human GenAI Collaboration

Christian Meske, Tobias Hermanns, Florian Brachten
This study investigates how traditional models of team collaboration, known as Transactive Memory Systems (TMS), manifest when humans work with Generative AI. Through in-depth interviews with 14 knowledge workers, the research analyzes the unique dynamics of expertise recognition, trust, and coordination that emerge in these partnerships.

Problem While Generative AI is increasingly used as a collaborative tool, our understanding of teamwork is based on human-to-human interaction. This creates a knowledge gap, as the established theories do not account for an AI partner that operates on algorithms rather than social cues, potentially leading to inefficient and frustrating collaborations.

Outcome - Human-AI collaboration is asymmetrical: Humans learn the AI's capabilities, but the AI fails to recognize and remember human expertise beyond a single conversation.
- Trust in GenAI is ambivalent and requires verification: Users simultaneously see the AI as an expert yet doubt its reliability, forcing them to constantly verify its outputs, a step not typically taken with trusted human colleagues.
- Teamwork is hierarchical, not mutual: Humans must always take the lead and direct a passive AI that lacks initiative, creating a 'boss-employee' dynamic rather than a reciprocal partnership where both parties contribute ideas.
Generative AI, Transactive Memory Systems, Human-AI Collaboration, Knowledge Work, Trust in AI, Expertise Recognition, Coordination
Aisle be Back: State-of-the-Art Adoption of Retail Service Robots in Brick-and-Mortar Retail
International Conference on Wirtschaftsinformatik (2025)

Aisle be Back: State-of-the-Art Adoption of Retail Service Robots in Brick-and-Mortar Retail

Luisa Strelow, Michael Dominic Harr, and Reinhard Schütte
This study analyzes the current state of Retail Service Robot (RSR) adoption in physical, brick-and-mortar (B&M) stores. Using a dual research method that combines a systematic literature review with a multi-case study of major European retailers, the paper synthesizes how these robots are currently being used for various operational tasks.

Problem Brick-and-mortar retailers are facing significant challenges, including acute staff shortages and intense competition from online stores, which threaten their operational efficiency. While service robots offer a potential solution to sustain operations and transform the customer experience, a comprehensive understanding of their current adoption in retail environments is lacking.

Outcome - Retail Service Robots (RSRs) are predominantly adopted for tasks related to information exchange and goods transportation, which improves both customer service and operational efficiency.
- The potential for more advanced, human-like (anthropomorphic) interaction between robots and customers has not yet been fully utilized by retailers.
- The adoption of RSRs in the B&M retail sector is still in its infancy, with most robots being used for narrowly defined, single-purpose tasks rather than leveraging their full multi-functional potential.
- Research has focused more on customer-robot interactions than on employee-robot interactions, leaving a gap in understanding employee acceptance and collaboration.
- Many robotic systems discussed in academic literature are prototypes tested in labs, with few long-term, real-world deployments reported, especially in customer service roles.
Retail Service Robot, Brick-and-Mortar, Technology Adoption, Artificial Intelligence, Automation
Fostering Active Student Engagement in Flipped Classroom Teaching with Social Normative Feedback Research Paper
International Conference on Wirtschaftsinformatik (2025)

Fostering Active Student Engagement in Flipped Classroom Teaching with Social Normative Feedback Research Paper

Maximilian May, Konstantin Hopf, Felix Haag, Thorsten Staake, and Felix Wortmann
This study examines the effectiveness of social normative feedback in improving student engagement within a flipped classroom setting. Through a randomized controlled trial with 140 undergraduate students, researchers provided one group with emails comparing their assignment progress to their peers, while a control group received no such feedback during the main study period.

Problem The flipped classroom model requires students to be self-regulated, but many struggle with procrastination, leading to late submissions of graded assignments and underuse of voluntary learning materials. This behavior negatively affects academic performance, creating a need for scalable digital interventions that can encourage more timely and active student participation.

Outcome - The social normative feedback intervention significantly reduced late submissions of graded assignments by 8.4 percentage points (an 18.5% decrease) compared to the control group.
- Submitting assignments earlier was strongly correlated with higher correctness rates and better academic performance.
- The feedback intervention helped mitigate the decline in assignment quality that was observed in later course modules for the control group.
- The intervention did not have a significant effect on students' engagement with optional, voluntary assignments during the semester.
Flipped Classroom, Social Normative Feedback, Self Regulated Learning, Digital Interventions, Student Engagement, Higher Education
The Value of Blockchain-Verified Micro-Credentials in Hiring Decisions
International Conference on Wirtschaftsinformatik (2025)

The Value of Blockchain-Verified Micro-Credentials in Hiring Decisions

Lyuba Stafyeyeva
This study investigates how blockchain verification and the type of credential-issuing institution (university vs. learning academy) influence employer perceptions of a job applicant's trustworthiness, expertise, and salary expectations. Using an experimental design with 200 participants, the research evaluated how different credential formats affected hiring assessments.

Problem Verifying academic credentials is often slow, expensive, and prone to fraud, undermining trust in the system. While new micro-credentials (MCs) offer an alternative, their credibility is often unclear to employers, and it is unknown if technologies like blockchain can effectively solve this trust issue in real-world hiring scenarios.

Outcome - Blockchain verification did not significantly increase employers' perceptions of an applicant's trustworthiness or expertise.
- Employers showed no significant preference for credentials issued by traditional universities over those from alternative learning academies, suggesting a shift toward competency-based hiring.
- Applicants with blockchain-verified credentials were offered lower minimum starting salaries, indicating that while verification may reduce hiring risk for employers, it does not increase the candidate's perceived value.
- The results suggest that institutional prestige is becoming less important than verifiable skills in the hiring process.
micro-credentials, blockchain, trust, verification, employer decision-making
Evaluating Consumer Decision-Making Trade-Offs in Smart Service Systems in the Smart Home Domain
International Conference on Wirtschaftsinformatik (2025)

Evaluating Consumer Decision-Making Trade-Offs in Smart Service Systems in the Smart Home Domain

Björn Konopka and Manuel Wiesche
This study investigates the trade-offs consumers make when purchasing smart home devices. Using a choice-based conjoint analysis, the research evaluates the relative importance of eight attributes related to performance (e.g., reliability), privacy (e.g., data storage), and market factors (e.g., price and provider).

Problem While smart home technology is increasingly popular, there is limited understanding of how consumers weigh different factors, particularly how they balance privacy concerns against product performance and cost. This study addresses this gap by quantifying which features consumers prioritize when making purchasing decisions for smart home systems.

Outcome - Reliability and the device provider are the most influential factors in consumer decision-making, significantly outweighing other attributes.
- Price and privacy-related attributes (such as data collection scope, purpose, and user controls) play a comparatively lesser role.
- Consumers strongly prefer products that are reliable and made by a trusted (in this case, domestic) provider.
- The findings indicate that consumers are willing to trade off privacy concerns for tangible benefits in performance and trust in the manufacturer.
Smart Service Systems, Smart Home, Conjoint, Consumer Preferences, Privacy
LLMs for Intelligent Automation - Insights from a Systematic Literature Review
International Conference on Wirtschaftsinformatik (2025)

LLMs for Intelligent Automation - Insights from a Systematic Literature Review

David Sonnabend, Mahei Manhai Li and Christoph Peters
This study conducts a systematic literature review to examine how Large Language Models (LLMs) can enhance Intelligent Automation (IA). The research aims to overcome the limitations of traditional Robotic Process Automation (RPA), such as handling unstructured data and workflow changes, by systematically investigating the integration of LLMs.

Problem Traditional Robotic Process Automation (RPA) struggles with complex tasks involving unstructured data and dynamic workflows. While Large Language Models (LLMs) show promise in addressing these issues, there has been no systematic investigation into how they can specifically advance the field of Intelligent Automation (IA), creating a significant research gap.

Outcome - LLMs are primarily used to process complex inputs, such as unstructured text, within automation workflows.
- They are leveraged to generate automation workflows directly from natural language commands, simplifying the creation process.
- LLMs are also used to guide goal-oriented Graphical User Interface (GUI) navigation, making automation more adaptable to interface changes.
- A key research gap was identified in the lack of systems that combine these different capabilities and enable continuous learning at runtime.
Large Language Models (LLMs), Intelligent Process Automation (IPA), Intelligent Automation (IA), Cognitive Automation (CA), Tool Learning, Systematic Literature Review, Robotic Process Automation (RPA)
Label Error Detection in Defect Classification using Area Under the Margin (AUM) Ranking on Tabular Data
International Conference on Wirtschaftsinformatik (2025)

Label Error Detection in Defect Classification using Area Under the Margin (AUM) Ranking on Tabular Data

Pavlos Rath-Manakidis, Kathrin Nauth, Henry Huick, Miriam Fee Unger, Felix Hoenig, Jens Poeppelbuss, and Laurenz Wiskott
This study introduces an efficient method using Area Under the Margin (AUM) ranking with gradient-boosted decision trees to detect labeling errors in tabular data. The approach is designed to improve data quality for machine learning models used in industrial quality control, specifically for flat steel defect classification. The method's effectiveness is validated on both public and real-world industrial datasets, demonstrating it can identify problematic labels in a single training run.

Problem Automated surface inspection systems in manufacturing rely on machine learning models trained on large datasets. The performance of these models is highly dependent on the quality of the data labels, but errors frequently occur due to annotator mistakes or ambiguous defect definitions. Existing methods for finding these label errors are often computationally expensive and not optimized for the tabular data formats common in industrial applications.

Outcome - The proposed AUM method is as effective as more complex, computationally expensive techniques for detecting label errors but requires only a single model training run.
- The method successfully identifies both synthetically created and real-world label errors in industrial datasets related to steel defect classification.
- Integrating this method into quality control workflows significantly reduces the manual effort required to find and correct mislabeled data, improving the overall quality of training datasets and subsequent model performance.
- In a real-world test, the method flagged suspicious samples for expert review, where 42% were confirmed to be labeling errors.
Label Error Detection, Automated Surface Inspection System (ASIS), Machine Learning, Gradient Boosting, Data-centric AI
Measuring AI Literacy of Future Knowledge Workers: A Mediated Model of AI Experience and AI Knowledge
International Conference on Wirtschaftsinformatik (2025)

Measuring AI Literacy of Future Knowledge Workers: A Mediated Model of AI Experience and AI Knowledge

Sarah Hönigsberg, Sabrine Mallek, Laura Watkowski, and Pauline Weritz
This study investigates how future professionals develop AI literacy, which is the ability to effectively use and understand AI tools. Using a survey of 352 business school students, the researchers examined how hands-on experience with AI (both using and designing it) and theoretical knowledge about AI work together to build overall proficiency. The research proposes a new model showing that knowledge acts as a critical bridge between simply using AI and truly understanding it.

Problem As AI becomes a standard tool in professional settings, simply knowing how to use it isn't enough; professionals need a deeper understanding, or "AI literacy," to use it effectively and responsibly. The study addresses the problem that current frameworks for teaching AI skills often overlook the specific needs of knowledge workers and don't clarify how hands-on experience translates into true competence. This gap makes it difficult for companies and universities to design effective training programs to prepare the future workforce.

Outcome - Hands-on experience with AI is crucial, but it doesn't directly create AI proficiency; instead, it serves to build a foundation of AI knowledge.
- This structured AI knowledge is the critical bridge that turns practical experience into true AI literacy, allowing individuals to critique and apply AI insights effectively.
- Experience in designing or configuring AI systems has a significantly stronger positive impact on developing AI literacy than just using AI tools.
- The findings suggest that education and corporate training should combine practical, hands-on projects with structured learning about how AI works to build a truly AI-literate workforce.
knowledge worker, Al literacy, digital intelligence, digital literacy, AI knowledge
Typing Less, Saying More? – The Effects of Using Generative AI in Online Consumer Review Writing
International Conference on Wirtschaftsinformatik (2025)

Typing Less, Saying More? – The Effects of Using Generative AI in Online Consumer Review Writing

Maximilian Habla
This study investigates how using Generative AI (GenAI) impacts the quality and informativeness of online consumer reviews. Through a scenario-based online experiment, the research compares reviews written with and without GenAI assistance, analyzing factors like the writer's cognitive load and the resulting review's detail, complexity, and sentiment.

Problem Writing detailed, informative online reviews is a mentally demanding task for consumers, which often results in less helpful content for others making purchasing decisions. While platforms use templates to help, these still require significant effort from the reviewer. This study addresses the gap in understanding whether new GenAI tools can make it easier for people to write better, more useful reviews.

Outcome - Using GenAI significantly reduces the perceived cognitive load (mental effort) for people writing reviews.
- Reviews written with the help of GenAI are more informative, covering a greater number and a wider diversity of product aspects and topics.
- GenAI-assisted reviews tend to exhibit higher linguistic complexity and express a more positive sentiment, even when the star rating given by the user is the same.
- Contrary to the initial hypothesis, the reduction in cognitive load did not directly account for the increase in review informativeness, suggesting other mechanisms are at play.
Online Reviews, Informativeness, GenAI, Cognitive Load Theory, Linguistic Complexity, Sentiment Analysis
Unveiling the Influence of Personality, Identity, and Organizational Culture on Generative AI Adoption in the Workplace
International Conference on Wirtschaftsinformatik (2025)

Unveiling the Influence of Personality, Identity, and Organizational Culture on Generative AI Adoption in the Workplace

Dugaxhin Xhigoli
This qualitative study examines how an employee's personality, professional identity, and company culture influence their engagement with generative AI (GenAI). Through 23 expert interviews, the research explores the underlying factors that shape different AI adoption behaviors, from transparent integration to strategic concealment.

Problem As companies rapidly adopt generative AI, they encounter a wide range of employee responses, yet there is limited understanding of what drives this variation. This study addresses the research gap by investigating why employees differ in their AI usage, specifically focusing on how individual psychology and the organizational environment interact to shape these behaviors.

Outcome - The study identified four key dimensions influencing GenAI adoption: Personality-driven usage behavior, AI-driven changes to professional identity, organizational culture factors, and the organizational risks of unmanaged AI use.
- Four distinct employee archetypes were identified: 'Innovative Pioneers' who openly use and identify with AI, 'Hidden Users' who identify with AI but conceal its use for competitive advantage, 'Transparent Users' who openly use AI as a tool, and 'Critical Skeptics' who remain cautious and avoid it.
- Personality traits, particularly those from the 'Dark Triad' like narcissism, and competitive work environments significantly drive the strategic concealment of AI use.
- A company's culture is critical; open, innovative cultures foster ethical and transparent AI adoption, whereas rigid, hierarchical cultures encourage concealment and the rise of risky 'Shadow AI'.
Generative AI, Personality Traits, AI Identity, Organizational Culture, AI Adoption
Load More Showing 72 of 101