Abstract
Although developing successful autonomous vehicles for practical use is of interest, the establishment of moral autonomous vehicles aligned with human values is also crucial. Prior studies have mainly examined moral decision making in the trolley dilemma of autonomous vehicles, i.e., the driverless dilemma. Relatively little is known, however, about the passenger acceptance (specifically like, use, trust, and communication) of autonomous vehicles in the driverless dilemma. Results of a correlational study (Study 1) and an experimental study (Study 2) found that participants as the passengers were more likely to like, use, trust, and communicate with autonomous vehicles programmed to protect self than protect others and be random in a one-passenger-one-pedestrian scenario representing the one-to-one dilemma. However, in most conditions participants showed no preference for either pro-self, pro-social, or random algorithms in the one-passenger-several-pedestrian scenarios implying the utilitarian dilemma (Studies 1 and 2). The variation of passenger did not affect acceptance of autonomous vehicles in the driverless dilemma (Study 2).
Original language | English |
---|---|
Pages (from-to) | 2185-2194 |
Number of pages | 10 |
Journal | International Journal of Human-Computer Interaction |
Volume | 40 |
Issue number | 9 |
Early online date | 2 Jan 2023 |
DOIs | |
Publication status | Published - 2 May 2024 |
Scopus Subject Areas
- Human Factors and Ergonomics
- Human-Computer Interaction
- Computer Science Applications