Winfield, A. F.
You really need to know what your bot(s) are thinking about you.
In: Wilks, Y., ed.
Close Engagements with Artificial Companions: Key social, psychological, ethical and design issues.
John Benjamins, pp. 201-208.
- Accepted Version
Publisher's URL: http://www.benjamins.com/cgi-bin/t_bookview.cgi?bo...
The projected ubiquity of personal companion robots raises a range of interesting but also challenging questions. There can be little doubt that an effective artificial companion, whether embodied or not, will need to be both sensitive to the emotional state of its human partner and be able to respond sensitively. It will, in other words, need artificial theory of mind - such an artificial companion would (need to) behave as if it has feelings and as if it understands how its human partner is feeling. This essay explores the implementation and implications of artificial theory of mind, and raises concerns over the asymmetry between and artificial companion's theory of mind for its human partner and the human's theory of mind for his or her artificial companion. The essay argues that social learning (imitation) is an additional requirement of artificial companion robots, then goes on to develop the idea that an artificial companion robot will not be one robot but several. A surprising consequence of these ideas is that a family of artificial companion robots could acquire an artificial culture of its own, and the essay concludes by speculating on what this might mean for human(s) interacting with their artificial companion robots.
|Item Type:||Book Section|
|Uncontrolled Keywords:||bots, natural language processing|
Professor A. Winfield
|Deposited On:||19 Jul 2011 11:49|
|Last Modified:||13 Aug 2013 07:55|
Request a change to this item
Total Document Downloads in Past 12 Months