Abstract
As robotic agents become increasingly pervasive in our lives, recent works have shown promising results on the use of emotions on social drones. However, further research is needed to determine the appropriateness of these emotions in various contexts. We here investigate the appropriateness of seven drone emotional states. In a series of six workshops (\(N{=}30\)) consisting of both individual and group activities, we studied emotion appropriateness for six drone metaphors across the two most popular and radically different domains for Human–Drone Interaction namely: Emergency and Entertainment. Within diverse situations of interactions, participants were able to identify the appropriateness of each emotion. Our results describe how each emotion was found both appropriate and inappropriate depending on context. We provide insights into unique opportunities generated by the perceived emotion appropriateness, depending on different roles of drone emotions in interactions with people. We conclude with design considerations for future social robotic agents, including the importance of using a broad range of emotions, the use of a neutral expression, the temporality of emotions, and novel applications to interaction design. This work contributes to the understanding of the inner workings of emotion appropriateness in drones, providing researchers with a starting point for future work on social flying robots.
Similar content being viewed by others
Data Availability
The data that support the findings of this study are available upon reasonable request from the corresponding author [VH]. The data are not publicly available due to ethical restrictions.
Code Availability
Not applicable.
Notes
While the term “metaphor” often refers to an idiom or phrase, we here refer to a single descriptor as presented in [20].
Note that the Unusual drone was not considered as it represents a collection of unique applications, which are less formed in the prior body of work.
Note: In the results section, Eme and Ent respectively refer to the Emergency and Entertainment domains.
References
Herdel V, Kuzminykh A, Hildebrandt A, et al (2021) Drone in love: Emotional perception of facial expressions on flying robots. In: Proceedings of the 2021 CHI conference on human factors in computing systems. ACM, New York, NY, USA, CHI ’21, pp 1–20. https://doi.org/10.1145/3411764.3445495
Cauchard JR, Zhai KY, Spadafora M et al (2016) Emotion encoding in human-drone interaction. In: The 11th ACM/IEEE international conference on human robot interaction. IEEE, HRI, vol 16, pp 263–270. https://doi.org/10.1109/HRI.2016.7451761
Sharma M, Hildebrandt D, Newman G, et al (2013) Communicating affect via flight path: exploring use of the Laban effort system for designing affective locomotion paths. In: Proceedings of the 8th ACM/IEEE international conference on human–robot interaction. IEEE, HRI 13, pp 293–300. https://doi.org/10.1109/HRI.2013.6483602
Oxford Learner’s Dictionaries (2022) appropriate. https://www.oxfordlearnersdictionaries.com/definition/english/appropriate_1. Last Accessed 02 Feb 2023
Stock-Homburg R (2021) Survey of emotions in human–robot interactions: perspectives from robotic psychology on 20 years of research. Int J Soc Robot. https://doi.org/10.1007/s12369-021-00778-6
Fosch Villaronga E (2019) I Love You, said the robot: boundaries of the use of emotions in human–robot interactions, Springer, Cham, pp 93–110. https://doi.org/10.1007/978-3-319-96722-6_6
Ekman P (1999) Basic emotions. In: Handbook of cognition and emotion, vol 98, Nos 45–60, p 16. https://doi.org/10.1002/0470013494.ch3
Funk M (2018) Human–drone interaction: let’s get ready for flying user interfaces! Interactions 25(3):78–81. https://doi.org/10.1145/3194317
Suarez-Fernandez RA, Sanchez-Lopez JL, Sampedro C, et al (2016) Natural user interfaces for human-drone multi-modal interaction. In: International conference on unmanned aircraft systems (ICUAS). IEEE, pp 1013–1022. https://doi.org/10.1109/ICUAS.2016.7502665
Cauchard JR, E JL, Zhai KY, et al (2015) Drone & me: An exploration into natural human-drone interaction. In: Proceedings of the 2015 ACM international joint conference on pervasive and ubiquitous computing. ACM, New York, NY, USA, UbiComp ’15, pp 361–365. https://doi.org/10.1145/2750858.2805823
Obaid M, Kistler F, Kasparavičiūtundefined G, et al (2016) How would you gesture navigate a drone? a user-centered approach to control a drone. In: Proceedings of the 20th international academic mindtrek conference. ACM, New York, NY, USA, AcademicMindtrek ’16, pp 113–121. https://doi.org/10.1145/2994310.2994348
Alon O, Rabinovich S, Fyodorov C, et al (2021) Drones in firefighting: A user-centered design perspective. In: Proceedings of the 23rd international conference on mobile human–computer interaction. Association for computing machinery, New York, NY, USA, MobileHCI ’21. https://doi.org/10.1145/3447526.3472030
Cai C, Yang S, Yan P, et al (2019) Real-time human-posture recognition for human–drone interaction using monocular vision. In: Intelligent robotics and applications. Springer, Cham, pp 203–216. https://doi.org/10.1007/978-3-030-27541-9_18
Cauchard JR, Tamkin A, Wang CY, et al (2019) Drone.io: a gestural and visual interface for human–drone interaction. In: 14th ACM/IEEE international conference on human–robot interaction (HRI). IEEE, pp 153–162. https://doi.org/10.1109/HRI.2019.8673011
Avila M, Funk M, Henze N (2015) Dronenavigator: using drones for navigating visually impaired persons. In: Proceedings of the 17th international ACM SIGACCESS conference on computers and accessibility. ACM, New York, NY, USA, ASSETS ’15, pp 327–328. https://doi.org/10.1145/2700648.2811362
Ginosar E, Cauchard JR (2023) At first light: Expressive lights in support of drone-initiated communication. In: Proceedings of the 2023 CHI conference on human factors in computing systems. Association for computing machinery, New York, NY, USA, CHI ’23. https://doi.org/10.1145/3544548.3581062
Salvini P, Ciaravella G, Yu W, et al (2010) How safe are service robots in urban environments? Bullying a robot. In: 19th international symposium in robot and human interactive communication. IEEE, pp 1–7. https://doi.org/10.1109/ROMAN.2010.5654677
Baytas MA, Çay D, Zhang Y, et al (2019) The design of social drones: A review of studies on autonomous flyers in inhabited environments. In: Proceedings of the 2019 CHI conference on human factors in computing systems. Association for computing machinery, New York, NY, USA, CHI ’19, pp 1–13. https://doi.org/10.1145/3290605.3300480
Fartook O, Oron-Gilad T, Cauchard JR (2023) Designing and prototyping drones for emotional support. In: Companion of the 2023 ACM/IEEE international conference on human–robot interaction. Association for computing machinery, New York, NY, USA, HRI ’23, pp 234–237. https://doi.org/10.1145/3568294.3580079
Herdel V, Yamin LJ, Cauchard JR (2022) Above and beyond: A scoping review of domains and applications for human-drone interaction. In: Proceedings of the 2022 CHI conference on human factors in computing systems. ACM, New York, NY, USA, CHI ’22, pp 1–22. https://doi.org/10.1145/3491102.3501881
Herdel V, Yamin LJ, Ginosar E et al (2021) Public drone: attitude towards drone capabilities in various contexts. ACM, New York, NY, USA. https://doi.org/10.1145/3447526.3472053
Scherer KR (2005) What are emotions? And how can they be measured? Soc Sci Inf 44(4):695–729. https://doi.org/10.1177/0539018405058216
Ekman P, Friesen WV (1971) Constants across cultures in the face and emotion. J Personal Soc Psychol 17(2):124–129. https://doi.org/10.1037/h0030377
Baumeister R, Vohs K, DeWall CN et al (2007) How emotion shapes behavior: feedback, anticipation, and reflection, rather than direct causation. Personal Soc Psychol Rev 11:167–203. https://doi.org/10.1177/1088868307301033
Breazeal C (2003) Emotion and sociable humanoid robots. Int J Hum Comput Stud 59(1–2):119–155. https://doi.org/10.1016/S1071-5819(03)00018-1
Frith CD, Frith U (2006) How we predict what other people are going to do. Brain Res 1079(1):36–46. https://doi.org/10.1016/j.brainres.2005.12.126
Reyes ME, Meza IV, Pineda LA (2019) Robotics facial expression of anger in collaborative human–robot interaction. Int J Adv Robot Syst. https://doi.org/10.1177/1729881418817972
Eyssel F, Hegel F, Horstmann G, et al (2010) Anthropomorphic inferences from emotional nonverbal cues: a case study. In: 19th international symposium in robot and human interactive communication. IEEE, pp 646–651. https://doi.org/10.1109/ROMAN.2010.5598687
Wiese E, Metta G, Wykowska A (2017) Robots as intentional agents: using neuroscientific methods to make robots appear more social. Front Psychol 8:1663. https://doi.org/10.3389/fpsyg.2017.01663
Kim EH, Kwak SS, Kwak YK (2009) Can robotic emotional expressions induce a human to empathize with a robot? In: The 18th IEEE international symposium on robot and human interactive communication. IEEE, pp 358–362. https://doi.org/10.1109/ROMAN.2009.5326282
Zhou S, Tian L (2020) Would you help a sad robot? Influence of robots’ emotional expressions on human-multi-robot collaboration. In: 2020 29th IEEE international conference on robot and human interactive communication, pp 1243–1250. https://doi.org/10.1109/RO-MAN47096.2020.9223524
Złotowski J, Proudfoot D, Yogeeswaran K et al (2015) Anthropomorphism: opportunities and challenges in human–robot interaction. Int J Soc Robot 7(3):347–360. https://doi.org/10.1007/s12369-014-0267-6
Leite I, Pereira A, Martinho C, et al (2008) Are emotional robots more fun to play with? In: RO-MAN 2008—the 17th IEEE international symposium on robot and human interactive communication. IEEE, pp 77–82. https://doi.org/10.1109/ROMAN.2008.4600646
Hoffman G, Ju W (2014) Designing robots with movement in mind. J Hum Robot Interact 3(1):91–122. https://doi.org/10.5898/JHRI.3.1.Hoffman
van Breemen A (2004) Animation engine for believable interactive user-interface robots. In: 2004 IEEE/RSJ international conference on intelligent robots and systems (IROS), vol 3, pp 2873–2878. https://doi.org/10.1109/IROS.2004.1389845
Li Z, Cummings C, Sreenath K (2020) Animated Cassie: a dynamic relatable robotic character. In: 2020 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp 3739–3746. https://doi.org/10.1109/IROS45743.2020.9340894
Chan L, Zhang BJ, Fitter NT (2021) Designing and validating expressive cozmo behaviors for accurately conveying emotions. In: 2021 30th IEEE international conference on robot and human interactive communication (RO-MAN), pp 1037–1044. https://doi.org/10.1109/RO-MAN50785.2021.9515425
Wojciechowska A, Frey J, Mandelblum E et al (2019) Designing drones: factors and characteristics influencing the perception of flying robots. Proc ACM Interact Mobile Wearable Ubiquitous Technol 3(3):1–19. https://doi.org/10.1145/3351269
Ruijten PAM, Cuijpers RH (2018) If drones could see: investigating evaluations of a drone with eyes. In: International conference on social robotics. Springer, pp 65–74. https://doi.org/10.1007/978-3-030-05204-1_7
Kirby R, Forlizzi J, Simmons R (2010) Affective social robots. Robot Auton Syst 58(3):322–332. https://doi.org/10.1016/j.robot.2009.09.015
Hegel F, Spexard T, Wrede B, et al (2006) Playing a different imitation game: interaction with an empathic android robot. In: 2006 6th IEEE-RAS international conference on humanoid robots, pp 56–61. https://doi.org/10.1109/ICHR.2006.321363
Tielman M, Neerincx M, Meyer JJ, et al (2014) Adaptive emotional expression in robot-child interaction. In: Proceedings of the 2014 ACM/IEEE international conference on human–robot interaction. ACM, New York, NY, USA, HRI ’14, pp 407–414. https://doi.org/10.1145/2559636.2559663
James J, Watson CI, MacDonald B (2018) Artificial empathy in social robots: an analysis of emotions in speech. In: 2018 27th IEEE international symposium on robot and human interactive communication (RO-MAN). IEEE, pp 632–637. https://doi.org/10.1109/ROMAN.2018.8525652
Brave S, Nass C, Hutchinson K (2005) Computers that care: investigating the effects of orientation of emotion exhibited by an embodied computer agent. Int J Hum Comput Stud 62(2):161–178. https://doi.org/10.1016/j.ijhcs.2004.11.002
de Graaf MMA, Ben Allouch S, van Dijk JAGM (2015) What makes robots social?: A user’s perspective on characteristics for social human–robot interaction. In: Social Robotics. Springer, Cham, pp 184–193. https://doi.org/10.1007/978-3-319-25554-5_19
Pelikan HRM, Broth M, Keevallik L (2020) "are you sad, Cozmo?": How humans make sense of a home robot’s emotion displays. In: Proceedings of the 2020 ACM/IEEE international conference on human–robot interaction. ACM, New York, NY, USA, HRI ’20, pp 461–470. https://doi.org/10.1145/3319502.3374814
Eastwood JD, Smilek D, Merikle PM (2001) Differential attentional guidance by unattended faces expressing positive and negative emotion. Percept Psychophys 63(6):1004–1013. https://doi.org/10.3758/BF03194519
Anderson M, Anderson S, Armen C (2005) Towards machine ethics: implementing two action-based ethical theories. In: AAAI fall symposium—technical report, pp 1–7
Ojha S, Williams MA, Johnston B (2018) The essence of ethical reasoning in robot-emotion processing. Int J Soc Robot 10(2):211–223. https://doi.org/10.1007/s12369-017-0459-y
Petrak B, Stapels JG, Weitz K, et al (2021) To move or not to move? social acceptability of robot proxemics behavior depending on user emotion. In: 2021 30th IEEE international conference on robot and human interactive communication (RO-MAN), pp 975–982. https://doi.org/10.1109/RO-MAN50785.2021.9515502
van Maris A, Zook N, Caleb-Solly P et al (2020) Designing ethical social robots–a longitudinal field study with older adults. Front Robot AI 7:1. https://doi.org/10.3389/frobt.2020.00001
Sharkey A, Sharkey N (2012) Granny and the robots: ethical issues in robot care for the elderly. Ethics Inf Technol 14(1):27–40. https://doi.org/10.1007/s10676-010-9234-6
Karjalainen KD, Romell AES, Ratsamee P, et al (2017) Social drone companion for the home environment: a user-centric exploration. In: Proceedings of the 5th international conference on human agent interaction. ACM, New York, NY, USA, HAI ’17, pp 89–96. https://doi.org/10.1145/3125739.3125774
Lidynia C, Philipsen R, Ziefle M (2017) Droning on about drones–acceptance of and perceived barriers to drones in civil usage contexts. In: Advances in human factors in robots and unmanned systems. Springer, Cham, pp 317–329. https://doi.org/10.1007/978-3-319-41959-6_26
Salvini P, Laschi C, Dario P (2010) Design for acceptability: improving robots’ coexistence in human society. Int J Soc Robot 2(4):451–460. https://doi.org/10.1007/s12369-010-0079-2
Plutchik R (1980) A general psychoevolutionary theory of emotion. In: Theories of Emotion. Elsevier, pp 3–33. https://doi.org/10.1016/B978-0-12-558701-3.50007-7
Russell JA (1980) A circumplex model of affect. J Personal Soc Psychol 39(6):1161–1178. https://doi.org/10.1037/h0077714
Bartneck C, Reichenbach J, van Breemen A (2004) In your face, robot! The influence of a character’s embodiment on how users perceive its emotional expressions. In: Proceedings of design and emotion 2004 conference, pp 32–51. https://doi.org/10.6084/m9.figshare.5160769
Cañamero L, Fredslund J (2001) I show you how i like you—can you read it in my face? [robotics]. IEEE Trans Syst Man Cybern Part A Syst Hum 31(5):454–459. https://doi.org/10.1109/3468.952719
Gasper K, Spencer LA, Hu D (2019) Does neutral affect exist? How challenging three beliefs about neutral affect can advance affective research. Front Psychol 10:2476. https://doi.org/10.3389/fpsyg.2019.02476
Goldschmidt G, Sever AL (2011) Inspiring design ideas with texts. Des Stud 32(2):139–155. https://doi.org/10.1016/j.destud.2010.09.006
Hofstede G, Hofstede GJ, Minkov M (2005) Cultures and organizations: software of the mind, vol 2. Mcgraw-Hill, New York
Nomura T, Suzuki T, Kanda T et al (2006) Measurement of negative attitudes toward robots. Interact Stud 7(3):437–454. https://doi.org/10.1075/is.7.3.14nom
Wojciechowska A, Hamidi F, Lucero A, et al (2020) Chasing lions: Co-designing human-drone interaction in sub-Saharan Africa. In: Proceedings of the 2020 ACM designing interactive systems conference. ACM, New York, NY, USA, DIS ’20, pp 141–152. https://doi.org/10.1145/3357236.3395481
Harrison C, Faste H (2014) Implications of location and touch for on-body projected interfaces. In: Proceedings of the 2014 conference on designing interactive systems. ACM, New York, NY, USA, DIS ’14, pp 543–552. https://doi.org/10.1145/2598510.2598587
Braun V, Clarke V (2006) Using thematic analysis in psychology. Qual Res Psychol 3(2):77–101. https://doi.org/10.1191/1478088706qp063oa
Terry G, Hayfield N (2020) Reflexive thematic analysis. Edward Elgar Publishing, Cheltenham. https://doi.org/10.4337/9781788977159.00049
Cauchard JR, Gover W, Chen W et al (2021) Drones in wonderland—disentangling collocated interaction using radical form. IEEE Robot Autom Lett. https://doi.org/10.1109/LRA.2021.3103653
Shapira S, Cauchard JR (2022) Integrating drones in response to public health emergencies: a combined framework to explore technology acceptance. Front Public Health 10:1. https://doi.org/10.3389/fpubh.2022.1019626
Rosenfeld A (2019) Are drivers ready for traffic enforcement drones? Accid Anal Prevent 122:199–206. https://doi.org/10.1016/j.aap.2018.10.006
Aydin B (2019) Public acceptance of drones: knowledge, attitudes, and practice. Technol Soc 59(101):180. https://doi.org/10.1016/j.techsoc.2019.101180
Tsiourti C, Weiss A, Wac K, et al (2017) Designing emotionally expressive robots: A comparative study on the perception of communication modalities. In: Proceedings of the 5th international conference on human agent interaction. ACM, New York, NY, USA, HAI ’17, pp 213–222. https://doi.org/10.1145/3125739.3125744
Chen JYC, Lakhmani SG, Stowers K et al (2018) Situation awareness-based agent transparency and human-autonomy teaming effectiveness. Theor Issues Ergon Sci 19(3):259–282. https://doi.org/10.1080/1463922X.2017.1315750
Mercado JE, Rupp MA, Chen JYC et al (2016) Intelligent agent transparency in human-agent teaming for multi-UxV management. Hum Factors 58(3):401–415. https://doi.org/10.1177/0018720815621206
Acknowledgements
We would like to thank Prof. Yisrael Parmet for his support and Zachary McKendrick for his consultancy and help in editing the scenarios. This research was supported in part by the Helmsley Charitable Trust through the Agricultural, Biological and Cognitive Robotics Initiative and by the Marcus Endowment Fund both at Ben-Gurion University of the Negev.
Funding
Partial financial support was received from the Helmsley Charitable Trust through the Agricultural, Biological and Cognitive Robotics Initiative and by the Marcus Endowment Fund both at Ben-Gurion University of the Negev.
Author information
Authors and Affiliations
Contributions
All authors contributed to the study conception and design. Material preparation and data collection were performed by Viviane Herdel, and analysis by Viviane Herdel and Jessica R. Cauchard. The first draft of the manuscript was written by Viviane Herdel and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no relevant financial or non-financial interests to disclose.
Ethical Approval
The questionnaires and methodology for this study was approved by the Human Research Ethics committee of Ben-Gurion University of the Negev.
Consent to Participate
Informed consent was obtained from all individual participants included in the study.
Consent for Publication
Patients signed informed consent regarding publishing their data.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Herdel, V., Cauchard, J.R. Emotion Appropriateness in Human–Drone Interaction. Int J of Soc Robotics 16, 579–597 (2024). https://doi.org/10.1007/s12369-023-01094-x
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12369-023-01094-x