Human-machine systems

Model
Digital Document
Publisher
Florida Atlantic University
Description
This thesis explores the cross-cultural demands from self-driving cars in regards to their trust, safety, and driving styles. Through the use of international survey data we establish several AI trust and behavior metrics that can be used for understanding cross-cultural expectations from self-driving cars that can potentially address problems of trust between passengers and self-driving cars, social acceptability of self-driving cars, and development of customized autonomous driving technologies. Further this thesis provides a serverless data-collection framework for future research in driving behaviors.
Model
Digital Document
Publisher
Florida Atlantic University
Description
It is becoming increasingly important for an autonomous system to be able to explain its actions to humans in order to improve trust and enhance human-machine collaboration. However, providing the most appropriate kind of explanations – in terms of length, format, and presentation mode of explanations at the proper time – is critical to enhancing their effectiveness. Explanation entails costs, such as the time it takes to explain and for humans to comprehend and respond. Therefore, the actual improvement in human-system tasks from explanations (if any) is not always obvious, particularly given various forms of uncertainty in knowledge about humans.
In this research, we propose an approach to address this issue. The key idea is to provide a structured framework that allows a system to model and reason about human personality traits as critical elements to guide proper explanation in human and system collaboration. In particular, we focus on the two concerns of modality and amount of explanation in order to optimize the explanation experience and improve overall system-human utility. Our models are based on probabilistic modeling and analysis (PRISM-games) to determine at run time what the most effective explanation under uncertainty is. To demonstrate our approach, we introduce a self-adaptative system called Grid – a virtual game – and the Stock Prediction Engine (SPE), which allows an automated system and a human to collaborate on the game and stock investments. Our evaluation of these exemplars, through simulation, demonstrates that a human subject’s performance and overall human-system utility is improved when considering the psychology of human personality traits in providing explanations.