Baruch College: Not all AI is Created Equal – A Meta-Analysis Revealing Drivers of AI Resistance Across Markets, Methods, and Time
The meta-analysis reveals that while consumers generally show a slight aversion to AI (Cohen’s d = -0.21), resistance is context-dependent—stronger for embodied forms like robots and high-risk domains—and evolves over time, with negative evaluations decreasing, especially in settings with greater ecological validity.
Baruch College: Not all AI is Created Equal – A Meta-Analysis Revealing Drivers of AI Resistance Across Markets, Methods, and Time
Summary of Read" class="text-blue-600 hover:text-blue-800" target="_blank" rel="noopener noreferrer">https://www.sciencedirect.com/science/article/pii/S0167811625000114'>Read Full Report
Presents a meta-analysis of two decades of studies examining consumer resistance to artificial intelligence (AI). The authors synthesize findings from hundreds of studies with over 76,000 participants, revealing that AI aversion is context-dependent and varies based on the AI's label, application domain, and perceived characteristics.
Interestingly, the study finds that negative consumer responses have decreased over time, particularly for cognitive evaluations of AI. Furthermore, the meta-analysis indicates that research design choices influence observed AI resistance, with studies using more ecologically valid methods showing less aversion.
-
Consumers exhibit an overall small but statistically significant aversion to AI (average Cohen’s d = -0.21). This means that, on average, people tend to respond more negatively to outputs or decisions labeled as coming from AI compared to those labeled as coming from humans.
-
Consumer aversion to AI is strongly context-dependent, varying significantly by the AI label and the application domain. Embodied forms of AI, such as robots, elicit the most negative responses (d = -0.83) compared to AI assistants or mere algorithms. Furthermore, domains involving higher stakes and risks, like transportation and public safety, trigger more negative responses than domains focused on productivity and performance, such as business and management.
-
Consumer responses to AI are not static and have evolved over time, generally becoming less negative, particularly for cognitive evaluations (e.g., performance or competence judgements). While initial excitement around generative AI in 2021 led to a near null-effect in cognitive evaluations, affective and behavioral responses still remain significantly negative overall.
-
The characteristics ascribed to AI significantly influence consumer responses. Negative responses are stronger when AI is described as having high autonomy (d = -0.28), inferior performance (d = -0.53), lacking human-like cues (anthropomorphism) (d = -0.23), and not recognizing the user's uniqueness (d = -0.24). Conversely, limiting AI autonomy, highlighting superior performance, incorporating anthropomorphic cues, and emphasizing uniqueness recognition can alleviate AI aversion.
-
The methodology used to study AI aversion impacts the findings. Studies with greater ecological validity, such as field studies, those using incentive-compatible designs, perceptually rich stimuli, clear explanations of AI, and behavioral (rather than self-report) measures, document significantly smaller aversion towards AI. This suggests that some documented resistance in purely hypothetical lab settings might be an overestimation of real-world aversion.
Related Articles
The MCP Context Window Problem: Why AI Agent Architecture Matters More Than Model Size
MCP servers are consuming up to 72% of AI agent context windows before a single user message is processed. Here is why smart agent architecture — not bigger models — is the real solution.
Amazon's AI Coding Crisis Reveals What Every Organization Needs: Controlled Agent Infrastructure
Amazon's recent production outages from AI coding agents reveal a fundamental truth: organizations need AI infrastructure they own and control. Here's what the industry can learn.
Why 1 Million Tokens of Context Changes Everything — If You Own the Infrastructure
Anthropic just made 1 million tokens of context generally available. Here's why long context only matters if the infrastructure running it belongs to you.
What Amazon's AI Coding Agent Outage Teaches Us About Deploying Agents in Production
Amazon's AI coding agent Kiro caused a 13-hour AWS outage by deleting a production environment. The incident reveals why organizations need owned, sandboxed AI infrastructure with proper governance — not just smarter models.
See the ibl.ai AI Operating System in Action
Discover how leading universities and organizations are transforming education with the ibl.ai AI Operating System. Explore real-world implementations from Harvard, MIT, Stanford, and users from 400+ institutions worldwide.
View Case StudiesGet Started with ibl.ai
Choose the plan that fits your needs and start transforming your educational experience today.