This paper considers the problem of providing advice to an autonomous agent when neither the behavioural policy nor the goals of that agent are known to the advisor. We present an approach based on building a model of “common sense” behaviour in the domain, from an aggregation of different users performing various tasks, modelled as MDPs, in the same domain. From this model, we estimate the normalcy of the trajectory given by a new agent in the domain, and provide behavioural advice based on an approximation of the trade-off in utility between potential benefits to the exploring agent and the costs incurred in giving this advice. This model is evaluated on a maze world domain by providing advice to different types of agents, and we show that this leads to a considerable and unanimous improvement in the completion rate of their tasks.
Reference:
Rosman, B.S. and Ramamoorthy, S. 2014. Robust Autonomy and Decisions Group. In: 2014 IEEE International Conference on Robotics and Automation (ICRA 2014), Hong Kong, China, 31 May - 7 June 2014
Rosman, B. S., & Ramamoorthy, S. (2014). Giving advice to agents with hidden goals. Robust Autonomy and Decisions Group. http://hdl.handle.net/10204/7620
Rosman, Benjamin S, and S Ramamoorthy. "Giving advice to agents with hidden goals." (2014): http://hdl.handle.net/10204/7620
Rosman BS, Ramamoorthy S, Giving advice to agents with hidden goals; Robust Autonomy and Decisions Group; 2014. http://hdl.handle.net/10204/7620 .