First-Person View Hand Segmentation of Multi-Modal Hand Activity Video Dataset

BMVC 2020

First-Person View Hand Segmentation of Multi-Modal Hand Activity Video Dataset

Dec 19, 2020
|
33 views
Details
Convergence Design Lab Purdue University Authors: Sangpil Kim, Hyung-gun Chi, Xiao Hu, Anirudh Vegesana, Karthik Ramani In proceedings of the 31st British Machine Vision Conference (BMVC) Abstract: First-person-view videos of hands interacting with tools are widely used in the computer vision industry. However, creating a dataset with pixel-wise segmentation of hands is challenging since most videos are captured with fingertips occluded by the hand dorsum and grasped tools. Current methods often rely on manually segmenting hands to create annotations, which is inefficient and costly. To relieve this challenge, we create a method that utilizes thermal information of hands for efficient pixel-wise hand segmentation to create a multi-modal activity video dataset. Our method is not affected by fingertip and joint occlusions and does not require hand pose ground truth. We show our method to be 24 times faster than the traditional polygon labeling method while maintaining high quality. With the segmentation method, we propose a multi-modal hand activity video dataset with 790 sequences and 401,765 frames of “hands using tools” videos captured by thermal and RGB-D cameras with hand segmentation data. We analyze multiple models for hand segmentation performance and benchmark four segmentation networks. We show that our multi-modal dataset with fusing LWIR and RGB-D frames achieves 5% better hand IoU performance than using RGB frames. More details can be found on our web site: https://engineering.purdue.edu/cdesign/wp/first-person-view-hand-segmentation-of-multi-modal-hand-activity-video-dataset/

Comments
loading...