<?xml version="1.0" encoding="UTF-8"?><rdf:RDF xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:dcq="http://purl.org/dc/terms/"><records count="1" morepages="false" start="1" end="1"><record rownumber="1"><dc:product_type>Conference Paper</dc:product_type><dc:title>Pragmatic-Pedagogic Value Alignment</dc:title><dc:creator>FIsac, J.; Gates, M.; Hamrick, J.; Liu, C.; Hadfield-Mennell, D.; Palaniappan, M.; Malik, D.; Sastry, S.; Griffiths, T.; Dragan, A.</dc:creator><dc:corporate_author/><dc:editor/><dc:description>As intelligent systems gain autonomy and capability, it becomes vital to
ensure that their objectives match those of their human users; this is known as the
value-alignment problem. In robotics, value alignment is key to the design of collaborative
robots that can integrate into human workflows, successfully inferring and
adapting to their users’ objectives as they go.We argue that a meaningful solution to
value alignment must combine multi-agent decision theory with rich mathematical
models of human cognition, enabling robots to tap into people’s natural collaborative
capabilities. We present a solution to the cooperative inverse reinforcement
learning (CIRL) dynamic game based on well-established cognitive models of decision
making and theory of mind. The solution captures a key reciprocity relation: the
human will not plan her actions in isolation, but rather reason pedagogically about
how the robot might learn from them; the robot, in turn, can anticipate this and interpret
the human’s actions pragmatically. To our knowledge, this work constitutes the
first formal analysis of value alignment grounded in empirically validated cognitive
models.</dc:description><dc:publisher/><dc:date>2018-01-01</dc:date><dc:nsf_par_id>10063837</dc:nsf_par_id><dc:journal_name>International Symposium on Robotics Research (ISRR)</dc:journal_name><dc:journal_volume/><dc:journal_issue/><dc:page_range_or_elocation/><dc:issn/><dc:isbn/><dc:doi>https://doi.org/</dc:doi><dcq:identifierAwardId>1734633</dcq:identifierAwardId><dc:subject/><dc:version_number/><dc:location/><dc:rights/><dc:institution/><dc:sponsoring_org>National Science Foundation</dc:sponsoring_org></record></records></rdf:RDF>