Projects per year
Abstract
Continuous detection of social interactions from wearable sensor data streams has a range of potential applications in domains including health and social care, security, and assistive technology. We contribute an annotated, multimodal dataset capturing such interactions using video, audio, GPS and inertial sensing. We present methods for automatic detection and temporal segmentation of focused interactions using support vector machines and recurrent neural networks with features extracted from both audio and video streams. Focused interaction occurs when co-present individuals, having mutual focus of attention, interact by first establishing face-to-face engagement and direct conversation. We describe an evaluation protocol including framewise, extended framewise and event-based measures and provide empirical evidence that fusion of visual face track scores with audio voice activity scores provides an effective combination. The methods, contributed dataset and protocol together provide a benchmark for future research on this problem. The dataset is available at https://doi.org/10.15132/10000134
Original language | English |
---|---|
Pages (from-to) | 37493-37505 |
Number of pages | 13 |
Journal | IEEE Access |
Volume | 6 |
Early online date | 25 Jun 2018 |
DOIs | |
Publication status | Published - 25 Jun 2018 |
Keywords
- Social interaction
- egocentric sensing
- multimodal analysis
- temporal segmentation
ASJC Scopus subject areas
- General Computer Science
- General Materials Science
- General Engineering
Fingerprint
Dive into the research topics of 'Multimodal Egocentric Analysis of Focused Interactions'. Together they form a unique fingerprint.Projects
- 1 Finished
-
ACE-LP - Augmenting Communication Using Environmental Data to Drive Language Prediction (Joint with University of Cambridge)
Black, R. (Investigator), McKenna, S. (Investigator), Waller, A. (Investigator) & Zhang, J. (Investigator)
Engineering and Physical Sciences Research Council
1/02/16 → 31/01/20
Project: Research
Datasets
-
Multimodal Focused Interaction Dataset
Bano, S. (Creator) & McKenna, S. (Creator), University of Dundee, Jun 2018
DOI: 10.15132/10000134, http://cvip.computing.dundee.ac.uk/datasets/focusedinteraction/
Dataset
File
Press/Media
-
Context Aware Text Prediction for Enhancing Augmentative Communication
Black, R., Waller, A., Nandadasa, H., Suveges, T., McKenna, S., Kristensson, P. O., Zhang, J., McKillop, C. & Mckillop, C.
19/11/19
2 items of Media coverage
Press/Media: Research