https://github.com/cwi-dis/rcea360vr-chi2021
CHI 2021 paper (RCEA-360VR) source code for our (a) Viewport-dependent annotation fusion method (b) Viewport-based fine-grained V-A video overlay generator
Science Score: 13.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
○codemeta.json file
-
○.zenodo.json file
-
✓DOI references
Found 4 DOI reference(s) in README -
○Academic publication links
-
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (8.3%) to scientific vocabulary
Keywords
affective-computing
emotion
ground-truth
hci
method
vr
Last synced: 5 months ago
·
JSON representation
Repository
CHI 2021 paper (RCEA-360VR) source code for our (a) Viewport-dependent annotation fusion method (b) Viewport-based fine-grained V-A video overlay generator
Basic Info
Statistics
- Stars: 2
- Watchers: 6
- Forks: 2
- Open Issues: 0
- Releases: 0
Topics
affective-computing
emotion
ground-truth
hci
method
vr
Created about 5 years ago
· Last pushed almost 5 years ago
https://github.com/cwi-dis/RCEA360VR-CHI2021/blob/master/
## CHI 2021 paper: RCEA-360VR
Source code for our CHI 2021 paper:
>*RCEA-360VR: Real-time, Continuous Emotion Annotation in 360 VR Videos for Collecting Precise Viewport-dependent Ground Truth Labels*
Repository contains:
1. [Viewport-dependent annotation fusion method](source)
2. [Viewport-based fine-grained V-A video overlay generator + example videos](video_results)
Example results for our viewport-dependent label fusion method:

*[Explanation of source code will be finalized soon]*
## Preprint + Video Preview
Preprint (.pdf) here: [chi2021_rcea360VR_preprint](preprint/chi2021_rcea360VR_preprint.pdf)
30s video preview (on YouTube):
[](https://www.youtube.com/watch?v=dSeCyH6OuIc "CHI 2021 RCEA-360VR")
## Citing this paper or code
**Please cite our paper in any published work that uses any of these resources.**
BiBTeX:
```
@inproceedings{Xue2021,
title = {RCEA-360VR: Real-time, Continuous Emotion Annotation in 360 VR Videos for Collecting Precise Viewport-dependent Ground Truth Labels},
author = {Tong Xue and Abdallah El Ali and Tianyi Zhang and Gangyi Ding and Pablo Cesar},
booktitle = {Proceedings of the International Conference on Human Factors in Computing Systems 2021},
series = {CHI '21},
year = {2021},
location = {Yokohama, Japan},
pages = {1-15},
keywords = {Emotion, annotation, 360, virtual reality, ground truth, labels, viewport-dependent, real-time, continuous},
url = {https://doi.org/10.1145/3411764.3445487},
abstract = {Precise emotion ground truth labels for 360 virtual reality (VR) video watching are essential for fine-grained predictions under varying viewing behavior. However, current annotation techniques either rely on post-stimulus discrete self-reports, or real-time, continuous emotion annotations (RCEA) but only for desktop/mobile settings. We present RCEA for 360 VR videos (RCEA-360VR), where we evaluate in a controlled study (N=32) the usability of two peripheral visualization techniques: HaloLight and DotSize. We furthermore develop a method that considers head movements when fusing labels. Using physiological, behavioral, and subjective measures, we show that (1) both techniques do not increase users' workload, sickness, nor break presence (2) our continuous valence and arousal annotations are consistent with discrete within-VR and original stimuli ratings (3) users exhibit high similarity in viewing behavior, where fused ratings perfectly align with intended labels. Our work contributes usable and effective techniques for collecting fine-grained viewport-dependent emotion labels in 360 VR.},
}
```
ACM Ref Citation:
*Tong Xue, Abdallah El Ali, Tianyi Zhang, Gangyi Ding, and Pablo Cesar (2021). RCEA-360VR: Real-time, Continuous Emotion Annotation in 360 VR Videos for Collecting Precise Viewport-dependent Ground Truth Labels. In CHI Conference on Human Factors in Computing Systems (CHI 21), May 813, 2021, Yokohama, Japan. ACM, New York, NY, USA, 15 pages. https://doi.org/10.1145/3411764.3445487*
## Licenses
Code in this repo is released under [Mozilla Public
License 2.0](https://github.com/ayman/hubs-research-2021/blob/main/LICENSE).
Owner
- Name: cwi-dis
- Login: cwi-dis
- Kind: organization
- Location: Amsterdam, the Netherlands
- Website: http://www.dis.cwi.nl
- Repositories: 21
- Profile: https://github.com/cwi-dis
CWI Distributed and Interactive Systems Group