TY - GEN
T1 - Automatic Labeling of Training Data by Vowel Recognition for Mouth Shape Recognition with Optical Sensors Embedded in Head-Mounted Display
AU - Nakamura, Fumihiko
AU - Suzuki, Katsuhiro
AU - Masai, Katsutoshi
AU - Itoh, Yuta
AU - Sugiura, Yuta
AU - Sugimoto, Maki
N1 - Funding Information:
This work was supported by JSPS KEKENHI Grant Number 16H05870.
Publisher Copyright:
© 2019 The Author(s)
PY - 2019
Y1 - 2019
N2 - Facial expressions enrich communication via avatars. However, in common immersive virtual reality (VR) systems, facial occlusions by head-mounted displays (HMD) lead to difficulties in capturing users’ faces. In particular, the mouth plays an important role in facial expressions because it is essential for rich interaction. In this paper, we propose a technique that classifies mouth shapes into six classes using optical sensors embedded in HMD and gives labels automatically to the training dataset by vowel recognition. We experiment with five subjects to compare the recognition rates of machine learning under manual and automated labeling conditions. Results show that our method achieves average classification accuracy of 99.9% and 96.3% under manual and automated labeling conditions, respectively. These findings indicate that automated labeling is competitive relative to manual labeling, although the former’s classification accuracy is slightly higher than that of the latter. Furthermore, we develop an application that reflects the mouth shape on avatars. This application blends six mouth shapes and then applies the blended mouth shapes to avatars.
AB - Facial expressions enrich communication via avatars. However, in common immersive virtual reality (VR) systems, facial occlusions by head-mounted displays (HMD) lead to difficulties in capturing users’ faces. In particular, the mouth plays an important role in facial expressions because it is essential for rich interaction. In this paper, we propose a technique that classifies mouth shapes into six classes using optical sensors embedded in HMD and gives labels automatically to the training dataset by vowel recognition. We experiment with five subjects to compare the recognition rates of machine learning under manual and automated labeling conditions. Results show that our method achieves average classification accuracy of 99.9% and 96.3% under manual and automated labeling conditions, respectively. These findings indicate that automated labeling is competitive relative to manual labeling, although the former’s classification accuracy is slightly higher than that of the latter. Furthermore, we develop an application that reflects the mouth shape on avatars. This application blends six mouth shapes and then applies the blended mouth shapes to avatars.
UR - http://www.scopus.com/inward/record.url?scp=85122864831&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85122864831&partnerID=8YFLogxK
U2 - 10.2312/egve.20191274
DO - 10.2312/egve.20191274
M3 - Conference contribution
AN - SCOPUS:85122864831
T3 - ICAT-EGVE 2019 - 29th International Conference on Artificial Reality and Telexistence and 24th Eurographics Symposium on Virtual Environments
SP - 9
EP - 16
BT - ICAT-EGVE 2019 - 29th International Conference on Artificial Reality and Telexistence and 24th Eurographics Symposium on Virtual Environments
A2 - Kakehi, Yasuaki
A2 - Hiyama, Atsushi
A2 - Fellner, Dieter W.
A2 - Hansmann, Werner
A2 - Purgathofer, Werner
A2 - Sillion, Francois
PB - The Eurographics Association
T2 - 29th International Conference on Artificial Reality and Telexistence and 24th Eurographics Symposium on Virtual Environments, ICAT-EGVE 2019
Y2 - 11 September 2019 through 13 September 2019
ER -