Abstract
We study human behavior in ultimatum game when interacting with either human or algorithmic opponents. We examine how the type of the AI algorithm (mimicking human behavior, optimising gains, or providing no explanation) and the presence of a human beneficiary affect sending and accepting behaviors. Our experimental data reveal that subjects generally do not differentiate between human and algorithmic opponents, between different algorithms, and between an explained and unexplained algorithm. However, they are more willing to forgo higher payoffs when the algorithm's earnings benefit a human.
Original language | English |
---|---|
Article number | 111979 |
Pages (from-to) | 1-4 |
Number of pages | 4 |
Journal | Economics Letters |
Volume | 244 |
DOIs | |
Publication status | Published - 21 Sept 2024 |
Bibliographical note
Publisher Copyright:© 2024 The Author(s)
Keywords
- Equity
- Explainability
- Fairness
- Human-ai interaction
- Social preferences
- Ultimatum bargaining