Abstract
Robots are becoming increasingly popular as a teaching aid in language learning. For language learning, which relies on inter-personal interactions and references to the physical world, an agent’s embodiment and ability to adapt to the student are both important factors. In this study, adaptive behavior and embodiment were combined in robot-assisted language learning. An online brain-computer interface (BCI) was used to monitor student’s brain activity and prompt adaptive responses from the robot whenever a lapse of attention was detected. The response involved additional repetition of the latest word and an iconic gesture to illustrate its meaning. To isolate the effect of embodied interaction in such a system, participants completed learning tasks in two conditions: one where the adaptive robot was physically present, and another where videos of the robot appeared on a screen. Despite no changes in robot’s behavior, participants reported higher engagement and more positive impressions of the robot, and also showed increased learning outcomes during the embodied interaction. This study confirms the importance of embodied interaction during adaptive learning and highlights the effectiveness of BCI systems in the design of future pedagogical robots.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Belpaeme, T., Kennedy, J., Ramachandran, A., Scassellati, B., Tanaka, F.: Social robots for education: a review. Sci. Rob. 3(8) (2018)
Vogt, P., et al.: Second language tutoring using social robots: a large-scale study. In: 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI). pp. 497–505. IEEE (2019)
Ahmad, M.I., Mubin, O., Shahid, S., Orlando, J.: Robot’s adaptive emotional feed-back sustains children’s social engagement and promotes their vocabulary learning: a long-term child–robot interaction study. Adapt. Behav. 27(4), 243–266 (2019)
Martin, F., Chen, Y., Moore, R.L., Westine, C.D.: Systematic review of adaptive learning research designs, context, strategies, and technologies from 2009 to 2018. Educ. Technol. Res. Dev. 68, 1903–1929 (2020)
Wang, S., Christensen, C., Cui, W., Tong, R., Yarnall, L., Shear, L., Feng, M.: When adaptive learning is effective learning: comparison of an adaptive learning system to teacher-led instruction. Interact. Learn. Environ. (2020)
Schodde, T., Hoffmann, L., Stange, S., Kopp, S.: Adapt, explain, engage—a study on how social robots can scaffold second-language learning of children. ACM Trans. Hum. Robot Inter. 9, 1–27 (2020)
Wit, J.D., et al.: The effect of Arobot’s gestures and adaptive tutoring on children’s acquisition of second language vocabularies. In: The 2018 ACM/IEEE International Conference (2018)
Donnermann, M., Schaper, P., Lugrin, B.: Towards adaptive robotic tutors in universities: a field study. In: Ali, R., Lugrin, B., Charles, F. (eds.) PERSUASIVE 2021. LNCS, vol. 12684, pp. 33–46. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-79460-6_3
Zander, T.O., Krol, L.R.: Team phypa: brain-computer interfacing for everyday human-computer interaction. Period Polytech. Electr. Eng. Comput. Sci. 61, 209 (2017)
Alimardani, M., Hiraki, K.: Passive brain-computer interfaces for enhanced human-robot interaction. Front. Rob. AI July (2020)
Alimardani, M., van den Braak, S., Jouen, A.-L., Matsunaka, R., Hiraki, K.: Assessment of engagement and learning during child-robot interaction using EEG signals. In: Li, H., Ge, S.S., Wu, Y., Wykowska, A., He, H. (eds.) ICSR 2021. LNCS (LNAI), vol. 13086, pp. 671–682. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-90525-5_59
Szafir, D., Mutlu, B.: Pay attention! Designing adaptive agents that monitor and improve user engagement. In: Conference on Human Factors in Computing Systems Proceedings, pp. 11–20 (2012)
Khedher, A.B., Jraidi, I., Frasson, C.: Tracking students’ mental engagement using EEG signals during an interaction with a virtual learning environment. J. Intell. Learn. Syst. Appl. 11, 1–14 (2019)
Rohani, D.A., Puthusserypady, S.: BCI inside a virtual reality classroom: a potential training tool for attention. EPJ Nonlinear Biomed. Phys. 3 (2015)
Looije, R., van der Zalm, A., Neerincx, M.A., Beun, R.J.: Help, I need somebody the effect of embodiment on playful learning. In: 2012 IEEE RO-MAN: The 21st IEEE International Symposium on Robot and Human Interactive Communication, pp. 718–724 (2012)
Kennedy, J., Baxter, P., Belpaeme, T.: Comparing robot embodiments in a guided discovery learning interaction with children. Int. J. Soc. Rob. 7 (2015)
Köse, H., Uluer, P., Akalın, N., Yorgancı, R., Özkul, A., Ince, G.: The effect of embodiment in sign language tutoring with assistive humanoid robots. Int. J. Soc. Robot. 7, 537–548 (2015)
Kennedy, J., Baxter, P., Belpaeme, T.: The robot who tried too hard: Social behaviour of a robot tutor can negatively affect child learning. In: ACM/IEEE Inter-national Conference on Human-Robot Interaction 2015-March, pp. 67–74, March 2015
Prinsen, J., Pruss, E., Vrins, A., Ceccato, C.: A passive brain-computer interface for monitoring engagement during robot-assisted language learning. In: Proceedings of 2022 IEEE International Conference on Systems, Man and Cybernetics (2022)
g.tec Medical Engineering GmbH Austria: User Manual for Unicorn Brain Interface Hybrid Black, 1.18.00 ed (2019). https://www.unicorn-bi.com/
Pope, A.T., Bogart, E.H., Bartolome, D.S.: Biocybernetic system evaluates indices of operator engagement in automated task. Biol. Psychol. 40(1–2), 187–195 (1995)
Bartneck, C., Kulić, D., Croft, E., Zoghbi, S.: Measurement instruments for the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety robots. Int. J. Soc. Robot. 1(1), 71–81 (2009)
O’Brien, H.L., Cairns, P., Hall, M.: A practical approach to measuring user engagement with the refined user engagement scale (UES) and new UES short form. Int. J. Hum. Comput. Stud. 112, 28–39 (2018)
Kim, Y., Tscholl, M.: Young children’s embodied interactions with a social robot. Educ. Technol. Res. Dev. 69, 2059–2081 (2021)
Konijn, E.A., Jansen, B., Bustos, V.M., Hobbelink, V.L.N.F., Daniel, V.P.: Social robots for (second) language learning in (migrant) primary school children. Int. J. Soc. Robot. 14, 827–843 (2022)
Kanero, J., Tunalı, E.T., Oranç, C., Göksun, T., Küntay, A.C.: When even a robot tutor zooms: a study of embodiment, attitudes, and impressions. Front. Rob. AI 8, 169 (2021)
Pütten, A., Straßmann, C., Krämer, N.: Language learning with artificial entities: effects of an artificial tutor’s embodiment and behavior on users’ alignment and evaluation. In: International Conference on Social Robotics, pp. 96–107 (2020)
Sauro, J., Lewis, J.R.: Quantifying the User Experience: Practical Statistics for User Research, Morgan Kaufmann (2016)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Vrins, A., Pruss, E., Prinsen, J., Ceccato, C., Alimardani, M. (2022). Are You Paying Attention? The Effect of Embodied Interaction with an Adaptive Robot Tutor on User Engagement and Learning Performance. In: Cavallo, F., et al. Social Robotics. ICSR 2022. Lecture Notes in Computer Science(), vol 13818. Springer, Cham. https://doi.org/10.1007/978-3-031-24670-8_13
Download citation
DOI: https://doi.org/10.1007/978-3-031-24670-8_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-24669-2
Online ISBN: 978-3-031-24670-8
eBook Packages: Computer ScienceComputer Science (R0)