English 中文 日本語 Русский
NOKOV Showcases Banner

Capturing Motion,
Crafting Stories

Explore Our Case Studies: Transforming Motion into Masterpieces Across Industries

A deep learning-enabled visual-inertial fusion method for human pose estimation in occluded human-robot collaborative assembly scenarios

Client
School of Mechanical Engineering, Zhejiang University
Capture volume
Application
Human-Robot Collaboration, Motion Capture, Pose Estimation
Objects
Human body
Equipment used

A research team from the School of Mechanical Engineering, Zhejiang University, has published a paper titled "A deep learning-enabled visual-inertial fusion method for human pose estimation in occluded human-robot collaborative assembly scenarios" in Robotics and Computer-Integrated Manufacturing. 

图1.png

This research focuses on human-centric smart manufacturing scenarios, proposing a visual-inertial fusion-based method for human pose estimation to mitigate the impact of occlusions in industrial environments. NOKOV motion capture system provided ground-truth human pose data for the study, for the validation of the proposed method.


图8.png

Compared to state-of-the-art methods, the proposed Human Pose Estimation (HPE) method demonstrates higher accuracy on both public and custom datasets and exhibits robustness in occluded scenarios. The contributions of this work are summarized as follows:

1.  A novel visual-inertial fusion-based HPE method is developed for occluded Human-Robot Collaboration (HRC) scenarios. It utilizes a single RGB camera and sparse IMUs as inputs, balancing algorithmic performance and operator comfort.

2.  A part-specific cross-modal transformer-based fusion block is designed to integrate spatial features from different modalities, facilitating effective data fusion.

3.  a. Comprehensive experiments conducted on two public datasets and a synthetic-occlusion dataset demonstrate the competitive performance and occlusion robustness of the proposed method. b. An HPE dataset collected from an HRC assembly process, containing video, IMU data, and 3D ground truth annotations, is developed to validate the application performance.

 

This work validates the performance and robustness of the proposed visual-inertial fusion-based HPE method across various scenarios. During the construction of the custom dataset, the high-precision ground truth data provided by NOKOV optical motion capture system was utilized to assess the accuracy of the proposed method (Figs. 9, 10, 11). 

NOKOV motion capture system provides high-precision human pose ground truth, for the validation of the proposed method.

 

Authors Profile:

Baicun Wang (Corresponding Author): Associate Dean, Qiushi Distinguished Professor, Doctoral Supervisor, School of Mechanical Engineering, Zhejiang University.

Ci Song: Ph.D. Candidate, School of Mechanical Engineering, Zhejiang University.

Xingyu Li: Assistant Professor, School of Engineering Technology, Purdue University.

Huiying Zhou: Ph.D. Candidate, School of Mechanical Engineering, Zhejiang University.

Huayong Yang: Academician of the Chinese Academy of Engineering (CAE); Director of the Faculty of Engineering, Professor, Doctoral Supervisor, School of Mechanical Engineering, Zhejiang University.

Lihui Wang: Fellow of the Canadian Academy of Engineering (CAE); Professor and Head of Sustainable Production, Department of Production Engineering, KTH Royal Institute of Technology

Prev
IEEE RA-L | GeoPF: Infusing Geometry into Potential Fields for Reactive Planning in Non-trivial Environments

NOKOV Motion Capture Basketball Game Demo

UMI Game
2022-03-29

Kung Fu Motion Capture Performance

Shu-Gu Entertainment
2023-02-06

Applications of motion capture systems in wire-driven continuum robot research

Sichuan University
2022-06-17

IEEE TRO Proprioceptive State Estimation for Amphibious Tactile Sensing

Southern University of Science and Technology
2024-12-10

By using this site, you agree to our terms, which outline our use of cookies. CLOSE ×

Contact us
We are committed to responding promptly and will connect with you through our local distributors for further assistance.
Engineering Virtual Reality Life Sciences Entertainment
I would like to receive a quote
Beijing NOKOV Science & Technology Co., Ltd (Headquarter)
LocationRoom820, China Minmetals Tower, Chaoyang Dist., Beijing
Emailinfo@nokov.cn
Phone+ 86-10-64922321
Capture Volume*
Objective*
Full Bodies Drones/Robots Others
Quantity
Camera Type
Pluto1.3C Mars1.3H Mars2H Mars4H Underwater Others/I do not know
Camera Count
4 6 8 12 16 20 24 Others/I don't know