first person benchmark dataset - hassony2/inria-research-wiki GitHub Wiki
First-Person Hand Action Benchmark 2017
[first-person-benchmark-dataset] First-Person Hand Action Benchmark with RGB-D Videos and 3D Hand Pose Annotations [PDF] [project page] [dataset not available here] [notes]
Guillermo Garcia-Hernando, Shanxin Yuan, Seungryul Baek, Tae-Kyun Kim
read xx/05/2017
Synthesis
Benchmark hand pose estimation and action recognition from RGB-D data in ego-centric setting
No prior 3D model
New Dataset : Daily Hand-Object Actions Dataset
100.000 RGBD annotated with 3D hand poses
3D pose acquired using magnetic 6D sensors + inverse kinematics over 21 joint hand model
1175 action samples : 45 categories of hand actions (write, scratch, wash, squeeze,...) manually annotated, 25 objects in 3 scenarios
for 4 objects in 10 actions : 6-D object pose ground truth and mesh
camera on the shoulder
RGB : 1920x1080 (contaminated with magnets and tapes) Depth : 640x480 30 fps
Pipeline
Preprocessing
compensate for anthropomorphic differences by normalizing hand poses (same distance between pairs of joints)
Network training
- Shallow LSTM
- Deep LSTM
Results
purely depth-based results: poor using hand poses produces the best performances with best-performer Gram Matrix and Lie group
Hand object interaction has to be present in training set