M3B corpus: multi-modal meeting behavior corpus for group meeting assessment

This paper is the first trial to create a corpus on human-to-human multi-modal communication among multiple persons in group discussions. Our corpus includes not only video conversations but also the head movement and eye gaze. In addition, it includes detailed labels about the behaviors appeared in the discussion. Since we focused on the micro-behavior, we classified the general behavior into more detailed behaviors based on those meaning. For example, we have four types of smile: response, agree, interesting, sympathy. Because it takes much effort to create such corpus having multiple sensor data and detailed labels, it seems that no one has created it. In this work, we first attempted to create a corpus called "M3B Corpus (Multi-Modal Meeting Behavior Corpus)," which includes 320 minutes discussion among 21 Japanese students in total by developing the recording system that can handle multiple sensors and 360-degree camera simultaneously and synchronously. In this paper, we introduce our developed recording system and report the detail of M3B Corpus.

[1]  Philipp Scholl,et al.  A multi-media exchange format for time-series dataset curation , 2016, UbiComp Adjunct.

[2]  Masakiyo Fujimoto,et al.  Low-Latency Real-Time Meeting Recognition and Understanding Using Distant Microphones and Omni-Directional Camera , 2012, IEEE Transactions on Audio, Speech, and Language Processing.

[3]  Peter Robinson,et al.  OpenFace: An open source facial behavior analysis toolkit , 2016, 2016 IEEE Winter Conference on Applications of Computer Vision (WACV).

[4]  Jean-Marc Odobez,et al.  The vernissage corpus: A conversational Human-Robot-Interaction dataset , 2013, 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[5]  Peter Bull Posture and Gesture , 2016 .

[6]  H. Gunes,et al.  Multimodal Human-Human-Robot Interactions (MHHRI) Dataset for Studying Personality and Engagement , 2019, IEEE Transactions on Affective Computing.

[7]  Sebastian Drude,et al.  The Language Archive , 2013 .

[8]  Toshio Irino,et al.  Manual and Accelerometer Analysis of Head Nodding Patterns in Goal-oriented Dialogues , 2011, HCI.

[9]  Kazuya Murao,et al.  A method for structuring meeting logs using wearable sensors , 2019, Internet Things.