ABSTRACT
Good public speaking skills convey strong and effective communication, which is critical in many professions and used in everyday life. The ability to speak publicly requires a lot of training and practice. Recent technological developments enable new approaches for public speaking training that allow users to practice in a safe and engaging environment. We explore feedback strategies for public speaking training that are based on an interactive virtual audience paradigm. We investigate three study conditions: (1) a non-interactive virtual audience (control condition), (2) direct visual feedback, and (3) nonverbal feedback from an interactive virtual audience. We perform a threefold evaluation based on self-assessment questionnaires, expert assessments, and two objectively annotated measures of eye-contact and avoidance of pause fillers. Our experiments show that the interactive virtual audience brings together the best of both worlds: increased engagement and challenge as well as improved public speaking skills as judged by experts.
- Anderson, K., and et al. The TARDIS framework: Intelligent virtual agents for social coaching in job interviews. In Proceedings of International Conference on Advances in Computer Entertainment (2013), 476--491. Google ScholarDigital Library
- Baltrusaitis, T., Robinson, P., and Morency, L.-P. Constrained local neural fields for robust facial landmark detection in the wild. In IEEE International Conference on Computer Vision Workshops (ICCVW), IEEE (2013), 354--361. Google ScholarDigital Library
- Batrinca, L., Stratou, G., Shapiro, A., Morency, L.-P., and Scherer, S. Cicero - towards a multimodal virtual audience platform for public speaking training. In Proceedings of Intelligent Virtual Agents (IVA) 2013, Springer (2013), 116--128.Google ScholarCross Ref
- Chen, L., Feng, G., Joe, J., Leong, C. W., Kitchen, C., and Lee, C. M. Towards automated assessment of public speaking skills using multimodal cues. In Proceedings of the 16th International Conference on Multimodal Interaction, 200--203. Google ScholarDigital Library
- Chollet, M., Stratou, G., Shapiro, A., Morency, L.-P., and Scherer, S. An interactive virtual audience platform for public speaking training. In Proceedings of International Conference on Autonomous Agents and Multiagent Systems (AAMAS) (2014), 1657--1658. Google ScholarDigital Library
- DeVault, D., and et al. Simsensei kiosk: A virtual human interviewer for healthcare decision support. In Proceedings of Autonomous Agents and Multiagent Systems (AAMAS) (2014), 1061--1068. Google ScholarDigital Library
- DiMatteo, M. R., Hays, R. D., and Prince, L. M. Relationship of physicians' nonverbal communication skill to patient satisfaction, appointment noncompliance, and physician workload. Health Psychology 5, 6 (1986), 581.Google ScholarCross Ref
- Durlak, J. A. How to select, calculate, and interpret effect sizes. Journal of Pediatric Psychology 34, 9 (2009), 917--928.Google ScholarCross Ref
- Feng, A., Huang, Y., Xu, Y., and Shapiro, A. Fast, automatic character animation pipelines. Computer Animation and Virtual Worlds (2013). Google ScholarDigital Library
- Harris, S. R., Kemmerling, R. L., and North, M. M. Brief virtual reality therapy for public speaking anxiety. Cyberpsychology and Behavior 5 (2002), 543--550.Google ScholarCross Ref
- Hart, J., Gratch, J., and Marsella, S. How Virtual Reality Training Can Win Friends and Influence People. Human Factors in Defence. Ashgate, 2013, ch. 21, 235--249.Google Scholar
- Hedges, L. V. Distribution theory for glass's estimator of effect size and related estimators. Journal of Educational Statistics 6, 2 (1981), 107--128.Google ScholarCross Ref
- Hernandez, M., Choi, J., and Medioni, G. Laser scan quality 3-d face modeling using a low-cost depth camera. In Proceedings of the 20th European Signal Processing Conference (2012), 1995--1999.Google Scholar
- Hook, J. N., Smith, C. A., and Valentiner, D. P. A short-form of the personal report of confidence as a speaker. Personality and Individual Differences 44, 6 (2008), 1306--1313.Google ScholarCross Ref
- Hoque, M., Courgeon, M., Martin, J.-C., Bilge, M., and Picard, R. Mach: My automated conversation coach. In Proceedings of International Joint Conference on Pervasive and Ubiquitous Computing (2013). Google ScholarDigital Library
- Jennett, C., Cox, A. L., Cairns, P., Dhoparee, S., Epps, A., Tijs, T., and Walton, A. Measuring and defining the experience of immersion in games. International Journal of Human-Computer Studies 66 (2008), 641--661. Google ScholarDigital Library
- Johnson, W. L., Rickel, J. W., and Lester, J. C. Animated pedagogical agents: Face-to-face interaction in interactive learning environments. International Journal of Artificial Intelligence in Education 11, 1 (2000), 47--78.Google Scholar
- Kwon, J. H., Powell, J., and Chalmers, A. How level of realism influences anxiety in virtual reality environments for a job interview. International Journal of Human-Computer Studies 71, 10 (2013), 978--987. Google ScholarDigital Library
- Lane, H. C., Hays, M. J., Core, M. G., and Auerbach, D. Learning intercultural communication skills with virtual humans: Feedback and fidelity. Journal of Educational Psychology Special Issue on Advanced Learning Technologies 105, 4 (2013), 1026--1035.Google Scholar
- Lane, H. C., and Wray, R. E. Individualized Cultural and Social Skills Learning with Virtual Humans. Adaptive Technologies for Training and Education. Cambridge University Press, 2012, ch. 10.Google Scholar
- Li, H., Weise, T., and Pauly, M. Example-based facial rigging. ACM Transactions on Graphics (Proceedings SIGGRAPH 2010) 29, 3 (July 2010). Google ScholarDigital Library
- Lucas, G., Gratch, J., King, A., and Morency, L.-P. It's only a computer: Virtual humans increase willingness to disclose. Computers in Human Behavior 37 (2014), 94--100.Google ScholarDigital Library
- MacIntyre, P. D., Thivierge, K. A., and MacDonald, J. R. The effects of audience interest, responsiveness, and evaluation on public speaking anxiety and related variables. Communication Research Reports 14, 2 (1997), 157--168.Google ScholarCross Ref
- North, M. M., North, S. M., and Coble, J. R. Virtual reality therapy: An effective treatment for the fear of public speaking. International Journal of Virtual Reality 3 (1998), 2--6.Google ScholarCross Ref
- Park, S., Shoemark, P., and Morency, L.-P. Toward crowdsourcing micro-level behavior annotations: the challenges of interface, training, and generalization. In Proceedings of the 18th International Conference on Intelligent User Interfaces (IUI '14), ACM (2014), 37--46. Google ScholarDigital Library
- Paul, G. L. Insight vs. Desensitization in Psychotherapy: An Experiment in Anxiety Reduction. Stanford University Press, 1966.Google Scholar
- Pertaub, D. P., Slater, M., and Barker, C. An experiment on public speaking anxiety in response to three different types of virtual audience. Presence: Teleoperators and virtual environments 11 (2002), 68--78. Google ScholarDigital Library
- Peterson, R. T. An examination of the relative effectiveness of training in nonverbal communication: Personal selling implications. Journal of Marketing Education 27, 2 (2005), 143--150.Google ScholarCross Ref
- Rosenberg, A., and Hirschberg, J. Acoustic/prosodic and lexical correlates of charismatic speech. In Proceedings of Interspeech 2005, ISCA (2005), 513--516.Google ScholarCross Ref
- Rowe, J., Shores, L., Mott, B., and Lester, J. C. Integrating learning and engagement in narrative-centered learning environments. In Proceedings of the Tenth International Conference on Intelligent Tutoring Systems (2010). Google ScholarDigital Library
- Scherer, S., Layher, G., Kane, J., Neumann, H., and Campbell, N. An audiovisual political speech analysis incorporating eye-tracking and perception data. In Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC'12), ELRA (2012), 1114--1120.Google Scholar
- Scherer, S., Marsella, S., Stratou, G., Xu, Y., Morbini, F., Egan, A., Rizzo, A., and Morency, L.-P. Perception markup language: Towards a standardized representation of perceived nonverbal behaviors. In Proceedings of Intelligent Virtual Agents (IVA'12), LNAI 7502, Springer (2012), 455--463. Google ScholarDigital Library
- Schreiber, L. M., Gregory, D. P., and Shibley, L. R. The development and test of the public speaking competence rubric. Communication Education 61, 3 (2012), 205--233.Google ScholarCross Ref
- Shapiro, A. Building a character animation system. In Motion in Games, J. Allbeck and P. Faloutsos, Eds., vol. 7060 of Lecture Notes in Computer Science. Springer Berlin / Heidelberg, 2011, 98--109. Google ScholarDigital Library
- Shapiro, A., Feng, A., Wang, R., Li, H., Bolas, M., Medioni, G., and Suma, E. Rapid avatar capture and simulation using commodity depth sensors. Computer Animation and Virtual Worlds 25, 3--4 (2014), 201--211. Google ScholarDigital Library
- Sloetjes, H., and Wittenburg, P. Annotation by category: Elan and iso dcr. In Proceedings of the Sixth International Conference on Language Resources and Evaluation (LREC'08), European Language Resources Association (ELRA) (2008).Google Scholar
- Spence, S. H. Social skills training with children and young people: Theory, evidence, and practice. Child and Adolescent Mental Health 8, 2 (2003), 84--96.Google ScholarCross Ref
- Strangert, E., and Gustafson, J. What makes a good speaker? subject ratings, acoustic measurements and perceptual evaluations. In Proceedings of Interspeech 2008, ISCA (2008), 1688--1691.Google ScholarCross Ref
- Swartout, W., Artstein, R., Forbell, E., Foutz, S., Lane, H. C., Lange, B., Morie, J., Rizzo, A., and Traum, D. Virtual humans for learning. AI Magazine 34, 4 (2013), 13--30.Google ScholarCross Ref
- Tanaka, H., Sakti, S., Neubig, G., Toda, T., Negoro, H., Iwasaka, H., and Nakamura, S. Automated social skills trainer. In ACM International Conference on Intelligent User Interfaces (IUI) (2015). Google ScholarDigital Library
- Tanveer, M., Lin, E., and Hoque, M. E. Rhema: A real-time in-situ intelligent interface to help people with public speaking,. In ACM International Conference on Intelligent User Interfaces (IUI) (2015). Google ScholarDigital Library
- Wang, N., and Gratch, J. Don't Just Stare at Me! In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI) (Chicago, IL, 2010), 1241--1250. Google ScholarDigital Library
- Wang, R., Choi, J., and Medioni, G. Accurate full body scanning from a single fixed 3d camera. In 3D Imaging, Modeling, Processing, Visualization and Transmission (3DIMPVT), 2012 Second International Conference on, IEEE (2012), 432--439. Google ScholarDigital Library
Index Terms
- Exploring feedback strategies to improve public speaking: an interactive virtual audience framework
Recommendations
Effects of Good Speaking Techniques on Audience Engagement
ICMI '15: Proceedings of the 2015 ACM on International Conference on Multimodal InteractionUnderstanding audience engagement levels for presentations has the potential to enable richer and more focused interaction with audio-visual recordings. We describe an investigation into automated analysis of multimodal recordings of scientific talks ...
Multimodal Public Speaking Performance Assessment
ICMI '15: Proceedings of the 2015 ACM on International Conference on Multimodal InteractionThe ability to speak proficiently in public is essential for many professions and in everyday life. Public speaking skills are difficult to master and require extensive training. Recent developments in technology enable new approaches for public ...
Presentation Trainer, your Public Speaking Multimodal Coach
ICMI '15: Proceedings of the 2015 ACM on International Conference on Multimodal InteractionThe Presentation Trainer is a multimodal tool designed to support the practice of public speaking skills, by giving the user real-time feedback about different aspects of her nonverbal communication. It tracks the user's voice and body to interpret her ...
Comments