![]() Pepper, a plastic robot about the size of a toddler and with big eyes, is proficient in small talk, invites people to dance and jokes around. Paro, a digitally enhanced plush toy in the shape of a baby seal, reacts to being stroked and hugged and, as studies show, improves the well-being of patients. The robots Paro and Pepper are already doing such jobs in some special care homes. They can be used in psychotherapy or in elderly care where emotions have to be detected and positive emotions should be elicited. Second, EAI systems open up new operational fields. People tend to work better with a machine that recognizes and simulates emotions that are part of a sympathetic relationship, such as a smile or a word to express joy or gratitude-whether it is a chatbot, a museum guide or a co-worker. First, they help improve human–machine interaction: of course, it helps when the AI systems do not evoke fear, discomfort or similar feelings. Footnote 1ĮAI systems are, on the one hand, welcome for at least two reasons. ![]() I call such systems “emotionalized AI systems” (EAI). That is, they are designed and trained to elicit emotions in humans, to recognize human emotions and, sometimes, to simulate emotions. At the same time, AI-products are increasingly endowed with emotional characteristics. ![]() Examples include workers in factories and extend to personal voice assistants such as Siri or Alexa and to robots that do caretaker work. Interactions between humans and machines that include artificial intelligence are increasingly common in nearly all areas of life. To do so, I discuss three arguments: the argument from self-deception, the argument from lack of mutuality, and the argument from moral negligence. In this paper, I want to tackle these worries by focusing on the last aspect: in what sense could it be problematic or even wrong to establish an emotional relationship with EAI-systems? I want to show that the justifications for the widespread intuition concerning the problems are not as strong as they seem at first sight. There is a rather strong intuition that there is something wrong about getting attached to a machine, about having certain emotions towards it, and about getting involved in a kind of affective relationship with it. The introduction of such systems in our lives is met with some criticism. That is, they are designed and trained to elicit emotions in humans, to recognize human emotions and, sometimes, to simulate emotions (EAI). Meanwhile, AI-products are increasingly endowed with emotional characteristics.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |