Cross-modal approach for conversational well-being monitoring with multi-sensory earables
More Info
expand_more
expand_more
Abstract
We propose a cross-modal approach for conversational well-being monitoring with a multi-sensory earable. It consists of motion, audio, and BLE models on earables. Using the IMU sensor, the microphone, and BLE scanning, the models detect speaking activities, stress and emotion, and participants in the conversation, respectively. We discuss the feasibility in qualifying conversations with our purpose-built cross-modal model in an energy-efficient and privacy-preserving way. With the cross-modal model, we develop a mobile application that qualifies on-going conversations and provides personalised feedback on social well-being.
Files
3267305.3267695.pdf
(pdf | 1.39 Mb)
Unknown license
Download not available