Piavca: a framework for heterogeneous interactions with virtual characters | Virtual Reality Skip to main content
Log in

Piavca: a framework for heterogeneous interactions with virtual characters

  • Original Article
  • Published:
Virtual Reality Aims and scope Submit manuscript

    We’re sorry, something doesn't seem to be working properly.

    Please try refreshing the page. If that doesn't work, please contact support so we can address the problem.

Abstract

This paper presents a virtual character animation system for real- time multimodal interaction in an immersive virtual reality setting. Human to human interaction is highly multimodal, involving features such as verbal language, tone of voice, facial expression, gestures and gaze. This multimodality means that, in order to simulate social interaction, our characters must be able to handle many different types of interaction and many different types of animation, simultaneously. Our system is based on a model of animation that represents different types of animations as instantiations of an abstract function representation. This makes it easy to combine different types of animation. It also encourages the creation of behavior out of basic building blocks, making it easy to create and configure new behaviors for novel situations. The model has been implemented in Piavca, an open source character animation system.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Notes

  1. For quaternion animations multiplication is replaced by scaling the rotation angle by s.

  2. For quaternions, quaternion multiplication is used instead of addition.

References

  • Alexa M, Müller W (2000) Representing animations by principal components. Comput Graph Forum 19(3):411–418

    Article  Google Scholar 

  • Argyle M, Cookv M (1976) Gaze and mutual gaze. Cambridge University Press, Cambridge

    Google Scholar 

  • Arikan O, Forsyth DA (2002) Interactive motion generation from examples. ACM Trans Graph 21(3):483-490

    Article  Google Scholar 

  • Badler N, Philips C, Webber B (eds) (1993) Simulating humans: computer graphics, animation and control. Oxford University Press, Oxford

    MATH  Google Scholar 

  • Bui TD, Heylen D, Nijholt A (2004) Combination of facial movements on a 3D talking head. In: Proceedings of computer graphics international 2004, pp 284–290

  • Cassell J, Bickmore T, Campbell L, Chang K,Vilhjálmsson H, Yan H (1999) Embodiment in conversational interfaces: Rea. In: ACM SIGCHI , ACM Press, pp 520–527

  • Cassell J, Bickmore T, Campbell L, Chang K Vilhjálmsson H, Yan H (1999) Requirements for an architecture for embodied conversational characters. In: Proceedings of computer animation and simulation ’99

  • Cassell J, Stone M (1999) Living hand to mouth: psychological theories about speech and gesture in interactive dialogue systems. In: AAI 1999 fall symposium on psychological models of communication in collaborative systems, pp 34–42

  • Egges A, Molet T, Magnenat-Thalmann N (2004) Personalised real-time idle motion synthesis. In: 12th Pacific conference on computer graphics and applications, pp 121–130

  • Elliott C, Schechter G, Yeung R, Abi-Ezzi S (July 1994) Tbag: a high level framework for interactive, animated 3d graphics applications. In: Proceedings of SIGGRAPH 94, computer graphics proceedings, annual conference series, pp 421–434

  • Figueroa P, Green M, Hoover HJ (2002) Intml: a description language for vr applications. In: Proceedings of Web3D ’02

  • Gillies M, Ballin D (2003) A model of interpersonal attitude and posture generation. In: Rist T, Aylett R, Ballin D, Rickel J (eds) Fourth workshop on intelligent virtual agents. Kloster Irsee, Germany

  • Grassia FS (1998) Practical parameterization of rotations using the exponential map. J Graph Tools 3(3):29–48

    Google Scholar 

  • Gratch J, Marsella S (2001) Tears and fears: modeling emotions and emotional behaviors in synthetic agents. In: AGENTS ’01, Proceedings of the fifth international conference on autonomous agents, pp 278–285, ACM Press, New York

  • Guye-Vuilléme A, Capin TK, Pandzic IS, Magnenat-Thalmann N, Thalmann D (1999) Non-verbal communication interface for collaborative virtual environments. Virtual Real J 4:49–59

    Article  Google Scholar 

  • Kendon A (1970) Movement coordination in social interaction. Acta Psychol 32:1–25

    Article  Google Scholar 

  • Kopp S, Jung B, Lessmann N, Wachsmuth I (2003) Max–a multimodal assistant in virtual reality construction. KI-Knstliche Intelligenz 3(4):11–17

    Google Scholar 

  • Kopp S, Tepper P, Cassell J (2005) Towards integrated microplanning of language and iconic gesture for multimodal output. In: International Conference on Multimodal Interfaces (ICMI’04), ACM Press, pp 97–104

  • Kopp S, Wachsmuth I (2004) Synthesizing multimodal utterances for conversational agents. J Comput Animat Virtual Worlds 15(1):39–52

    Article  Google Scholar 

  • Kovar L, Gleicher M, Pighin F (2002) Motion graphs. ACM Trans Graph 21(3):473–482 July 2002

    Article  Google Scholar 

  • Kshirsagar S, Magnenat-Thalmann N (2002) A multilayer personality model. In: 2nd international symposium on smart graphics, pp 107–115

  • Lee J, Chai J, Reitsma PSA, Hodgins JK, Pollard NS (2002) Interactive control of avatars animated with human motion data. ACM Trans Graph 21(3):491–500

    Google Scholar 

  • Lee SP, Badler JB, Badler NI (2002) Eyes alive. ACM Trans Graph 21(3):637–644

    Google Scholar 

  • Maatman RM, Gratch J, Marsella S (2005) Natural behavior of a listening agent. In: Panayiotopoulos T, Gratch J, Aylett R, Ballin D, Olivier P, Rist T (eds) Intelligent virtual agents. 5th international working conference, Kos

  • Pan X, Gillies M, Slater M (2008) Male bodily responses during an interaction with a virtual woman. In: Prendinger H, Lester JC, Ishizuka M (eds) Intelligent virtual agents. 8th international conference, IVA 2008, Tokyo, Japan, September 1–3, 2008. Proceedings, Vol 5208 of lecture notes in computer science, Springer, pp 89–96

  • Pan X, Slater M (2007) A preliminary study of shy males interacting with a virtual female. In: Presence 2007. The 10th annual international workshop on presence, pp 101–108

  • Poggi I, Pelachaud C, De Rosis F, Carofiglio V, De Carolis B (2005) Greta. A believable embodied conversational agent. In: Stock O, Zancanaro M (eds) Multimodal intelligent information presentation. Text, speech and language technology, vol 27. Kluwer

  • Rose C, Cohen MF, Bodenheimer B (1998) Verbs and adverbs: multidimensional motion interpolation. IEEE Comput Graph Appl 18(5):32–40

    Google Scholar 

  • Slater M, Usoh M (1994) Body centred interaction in immersive virtual environments. In: Thalmann NM, Thalmann D (eds) Artificial life and virtual reality. Wiley, pp 125–148

  • Stone M, DeCarlo D, Oh I, Rodriguez C, Stere A, Lees A, Bregler C (2004) Speaking with hands: creating animated conversational characters from recordings of human performance. In: SIGGRAPH ’04, ACM SIGGRAPH 2004 papers. ACM Press, New York, pp 506–513

  • Thórisson K (1998) Real-time decision making in multimodal face-to-face communication. In: Second ACM international conference on autonomous agents, pp 16–23

  • Vilhjálmsson HH (2005) Augmenting online conversation through automated discourse tagging. In: 6th annual minitrack on persistent conversation at the 38th Hawaii international conference on system sciences

  • Vinayagamoorthy V, Garau M, Steed A, Slater M (2004) An eye gaze model for dyadic interaction in an immersive virtual environment: practice and experience. Comput Graph Forum 23(1):1–12

    Article  Google Scholar 

  • Vinayagamoorthy V, Gillies M, Steed A, Tanguy E, Pan X, Loscos C, Slater M (2006) Building expression into virtual characters. In: Eurographics conference state of the art reports

  • Witkin A, Popović Z (1995) Motion warping. In: ACM SIGGRAPH, pp 105–108

Download references

Acknowledgments

We would like to thank the funders of this work: BT plc, the European Union FET project PRESENCIA (contract number 27731) and the Empathic Avatars project funded by the UK Engineering and Physical Sciences Research Council. We also would like to thank the members of the University College London Department of Computer Science Virtual Environments and Graphics Group.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Marco Gillies.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Gillies, M., Pan, X. & Slater, M. Piavca: a framework for heterogeneous interactions with virtual characters. Virtual Reality 14, 221–228 (2010). https://doi.org/10.1007/s10055-010-0167-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10055-010-0167-5

Keywords

Navigation