[go: up one dir, main page]

Skip to main content

Recognition of Students’ Multiple Mental States in Conversation Based on Multimodal Cues

  • Conference paper
  • First Online:
Computer Supported Education (CSEDU 2020)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1473))

Included in the following conference series:

  • 906 Accesses

Abstract

Learning activities, especially face-to-face conversational coaching may invite students experience a set of learning-centered mental states including concentration, confusion, frustration, and boredom, those mental sates have been widely used as vital proxies for inferring their learning processes and are closely linked with learning outcomes. Recognizing students’ learning-centered mental states, particularly effectively detecting negative mental states such as confusion and frustration in teacher-student conversation could help teacher effectively monitor students’ learning situations in order to direct personalized and adaptive coaching resources to maximum students’ learning outcome. Most of research focused on analyzing students’ mental states using univariate modality when they completing pre-designed tasks in a computer related environment. It is still an open question on how to effectively measure students’ multiple mental states when they interacting with human teacher in coach-led conversations from various aspects. To achieve this goal, in this work, we developed an advanced multi-sensor-based system to record multi-modal conversational data of student-teacher conversations generated in a real university lab. We then attempt to derive a series of interpretable features from multiple perspectives including facial and physiological (heart rate) to characterize students’ multiple mental states. A set of supervised classifiers were built based on those features with different modality fusion methods to recognize multiple mental states of students. Our results have provided the experimental evidence to validate the outstanding predictive ability of our proposed features and the possibility of using multimodal data to recognize students’ multiple mental states in ‘in-the-wild’ student-teacher conversation.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Apple Inc.: ARKit—Apple developer documentation. https://developer.apple.com/documentation/arkit. Accessed 04 Dec 2017

  2. Busso, C., et al.: IEMOCAP: interactive emotional dyadic motion capture database. Lang. Resour. Eval. 42(4), 335 (2008)

    Article  Google Scholar 

  3. Burt, K.B., Obradović, J.: The construct of psychophysiological reactivity: statistical and psychometric issues. Dev. Rev. 33(1), 29–57 (2013)

    Article  Google Scholar 

  4. Craig, S., Graesser, A., Sullins, J., Gholson, B.: Affect and learning: an exploratory look into the role of affect in learning with AutoTutor. J. Educ. Media 29(3), 241–250 (2004)

    Article  Google Scholar 

  5. Calvo, R.A., D’Mello, S.: Affect detection: an interdisciplinary review of models, methods, and their applications. IEEE Trans. Affect. Comput. 1(1), 18–37 (2010)

    Article  Google Scholar 

  6. Cowley, B., Ravaja, N., Heikura, T.: Cardiovascular physiology predicts learning effects in a serious game activity. Comput. Educ. 60(1), 299–309 (2013)

    Article  Google Scholar 

  7. Devillers, L., Vidrascu, L.: Real-life emotion recognition in speech. In: Müller, C. (ed.) Speaker Classification II. LNCS (LNAI), vol. 4441, pp. 34–42. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-74122-0_4

    Chapter  Google Scholar 

  8. D’Mello, S.K., Craig, S.D., Sullins, J., Graesser, A.C.: Predicting affective states expressed through an emote-aloud procedure from AutoTutor’s mixed-initiative dialogue. Int. J. Artif. Intell. Educ. 16(1), 3–28 (2006)

    Google Scholar 

  9. D’Mello, S.K., Craig, S.D., Witherspoon, A., Mcdaniel, B., Graesser, A.: Automatic detection of learner’s affect from conversational cues. User Model. User-Adap. Inter. 18(1–2), 45–80 (2008)

    Article  Google Scholar 

  10. D’Mello, S., Graesser, A.: Dynamics of affective states during complex learning. Learn. Instr. 22(2), 145–157 (2012)

    Article  Google Scholar 

  11. D’Mello, S., Mills, C.: Emotions while writing about emotional and non-emotional topics. Motiv. Emot. 38(1), 140–156 (2013). https://doi.org/10.1007/s11031-013-9358-1

    Article  Google Scholar 

  12. Forbes-Riley, K., Litman, D.: When does disengagement correlate with learning in spoken dialog computer tutoring? In: Biswas, G., Bull, S., Kay, J., Mitrovic, A. (eds.) AIED 2011. LNCS (LNAI), vol. 6738, pp. 81–89. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21869-9_13

    Chapter  Google Scholar 

  13. Feidakis, M., Daradoumis, T., Caballé, S.: Building emotion-aware features in computer supported collaborative learning (CSCL) systems. In: Alpine Rendez-Vous (ARV) Workshop on Tools and Technologies for Emotion Awareness in Computer-Mediated Collaboration and Learning (ARV 2013) (2013)

    Google Scholar 

  14. Gomes, J., Yassine, M., Worsley, M., Blikstein, P.: Analysing engineering expertise of high school students using eye tracking and multimodal learning analytics. In: Educational Data Mining (2013)

    Google Scholar 

  15. Grafsgaard, J., Wiggins, J., Boyer, K., Wiebe, E., Lester, J.: Embodied affect in tutorial dialogue: student gesture and posture. In: Lane, H.C., Yacef, K., Mostow, J., Pavlik, P. (eds.) AIED 2013. LNCS (LNAI), vol. 7926, pp. 1–10. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-39112-5_1

    Chapter  Google Scholar 

  16. Hussain, M. S., AlZoubi, O., Calvo, R., D’Mello, S.: Affect detection from multichannel physiology during learning sessions with AutoTutor. In: Biswas, G., Bull, S., Kay, J., Mitrovic, A. (eds.) AIED 2011. LNCS (LNAI), vol. 6738, pp. 131–138. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21869-9_19

    Chapter  Google Scholar 

  17. Kononenko, I.: Estimating attributes: analysis and extensions of RELIEF. In: Bergadano, F., De Raedt, L. (eds.) ECML 1994. LNCS, vol. 784, pp. 171–182. Springer, Heidelberg (1994). https://doi.org/10.1007/3-540-57868-4_57

    Chapter  Google Scholar 

  18. Kapoor, A., Picard, R.W.: Multimodal affect recognition in learning environments. In: Proceedings of the 13th Annual ACM International Conference on Multimedia, pp. 677–682 (2005)

    Google Scholar 

  19. de Koning, B.B., Tabbers, H.K., Rikers, R.M., Paas, F.: Attention guidance in learning from a complex animation: seeing is understanding? Learn. Instr. 20(2), 111–122 (2010)

    Article  Google Scholar 

  20. Landis, J.R., Koch, G.G.: The measurement of observer agreement for categorical data. Biometrics 33(1), 159 (1977)

    Article  Google Scholar 

  21. Li, Y., Su, H., Shen, X., Li, W., Cao, Z., Niu, S.: DailyDialog: a manually labelled multi-turn dialogue dataset (2017). arXiv preprint arXiv:1710-03957

  22. Luft, C.D., Nolte, G., Bhattacharya, J.: High-learners present larger mid-frontal theta power and connectivity in response to incorrect performance feedback. J. Neurosci. 33(5), 2029–2038 (2013)

    Article  Google Scholar 

  23. O’Brien, H.L., Toms, E.G.: The development and evaluation of a survey to measure user engagement. J. Am. Soc. Inform. Sci. Technol. 61(1), 50–69 (2010)

    Article  Google Scholar 

  24. Pardos, Z.A., Baker, R.S., San Pedro, M.O., Gowda, S.M., Gowda, S.M.: Affective states and state tests: investigating how affect and engagement during the school year predict end-of-year learning outcomes. J. Learn. Anal. 1(1), 107–128 (2014)

    Article  Google Scholar 

  25. Parsons, J., Taylor, L.: Student engagement: what do we know and what should we do? University of Alberta (2012)

    Google Scholar 

  26. Peng, S., Chen, L., Gao, C., Tong, R.J.: Predicting students’ attention level with inter-pretable facial and head dynamic features in an online tutoring system (student abstract). In: AAAI, pp. 13895–13896 (2020)

    Google Scholar 

  27. Peng, S., Ohira, S., Nagao, K.: Automatic evaluation of students’ discussion skill based on their heart rate. In: McLaren, B.M., Reilly, R., Zvacek, S., Uhomoibhi, J. (eds.) CSEDU 2018. CCIS, vol. 1022, pp. 572–585. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-21151-6_27

    Chapter  Google Scholar 

  28. Peng, S., Ohira, S., Nagao, K.: Prediction of students’ answer relevance in discussion based on their heart-rate data. Int. J. Innov. Res. Educ. Sci. (IJIRES) 6(3), 414–424 (2019)

    Google Scholar 

  29. Peng, S., Ohira, S., Nagao, K.: Reading students’ multiple mental states in conversation from facial and heart rate cues. In: CSEDU (1), pp. 68–76 (2020)

    Google Scholar 

  30. Rodrigo, M.M.T., et al.: The effects of an interactive software agent on student affective dynamics while using; an intelligent tutoring system. IEEE Trans. Affect. Comput. 3(2), 224–236 (2012)

    Article  Google Scholar 

  31. Stevens, R., Galloway, T., Berka, C.: EEG-related changes in cognitive workload, engagement and distraction as students acquire problem solving skills. In: Conati, C., McCoy, K., Paliouras, G. (eds.) UM 2007. LNCS (LNAI), vol. 4511, pp. 187–196. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-73078-1_22

    Chapter  Google Scholar 

  32. Urbanowicz, R.J., Olson, R.S., Schmitt, P., Meeker, M., Moore, J.H.: Benchmarking relief-based feature selection methods for bioinformatics data mining. J. Biomed. Inform. 85, 168–188 (2018)

    Article  Google Scholar 

  33. Whitehill, J., et al.: Towards an optimal affect-sensitive instructional system of cognitive skills. In: CVPR 2011 Workshops, pp. 20–25. IEEE (2011)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Katashi Nagao .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Peng, S., Ohira, S., Nagao, K. (2021). Recognition of Students’ Multiple Mental States in Conversation Based on Multimodal Cues. In: Lane, H.C., Zvacek, S., Uhomoibhi, J. (eds) Computer Supported Education. CSEDU 2020. Communications in Computer and Information Science, vol 1473. Springer, Cham. https://doi.org/10.1007/978-3-030-86439-2_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-86439-2_24

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-86438-5

  • Online ISBN: 978-3-030-86439-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics