KLI

Human motion estimation using wearable RGB-D camera and inertial sensors

Metadata Downloads
Abstract
Recently, human motion estimation is an important task and attracting significant attention
in sports and medical applications, as it presents both theoretical and practical interest from
bio-mechanical, computer vision and robotics perspectives.
We focus on developing algorithms to: 1) estimate gait parameters, and 2) analysis human
motion during walking activity, with reduced number of sensor count.
In this dissertation, gait parameters are estimated as well as human lower-limb motion is
reconstructed using a single waist-mounted Intel Realsense Depth Camera D455, which has
an integrated 6-DOF Inertial Measurement Unit (IMU).

An inertial navigation algorithm is primarily proposed with only IMU data to obtain spatio-
temporal gait parameters, such as: attitude, position, velocity of a human body; and stance
phase duration, stride length, and walking distance. However, low-cost inertial sensors come
with noise and bias that lead to unavoidable accumulative errors.
Therefore, straight-line walking with a constant speed constraints is imposed in the filtering
algorithm to improve attitude estimation.
Visual odometry (VO) algorithm provides relative pose from image sequence, which plays as a
measurement updating role for the filter. Detected stance foot from color and depth image
data are used as landmarks to update foot position. Finally, a smoothing algorithm is proposed
as a linear optimization problem to minimize estimation errors. Stance foot position is derived
and other gait parameters can be calculated from step information.

A deep learning-based segmentation model with custom dataset is proposed to improve foot
detection not only in stance phases but also in swing phases. 3-D dual foot trajectories are
then calculated from proposed filter results and relative position of dual foot with respect to
the camera. However, foot position in between Toe-Off and Mid-Stance phases are missing
due to obscurity. To tackle this problem, a Graph Convolutional Network (GCN) based model
is proposed to predict pose from previous poses. Complete foot trajectories are finally
reconstructed with only a single waist-mounted RGB-D camera.

Through experiments, the proposed system shows promising results and could be applied for
human motion estimation in real application with a considerable accuracy.
Author(s)
당 득 꽁
Issued Date
2023
Awarded Date
2023-08
Type
Dissertation
Keyword
inertial sensorsRGB-D camerakalman filtergait analysis
URI
https://oak.ulsan.ac.kr/handle/2021.oak/12898
http://ulsan.dcollection.net/common/orgView/200000692809
Alternative Author(s)
DANG DUC CONG
Affiliation
울산대학교
Department
일반대학원 전기전자컴퓨터공학과
Advisor
Young Soo Suh
Degree
Doctor
Publisher
울산대학교 일반대학원 전기전자컴퓨터공학과
Language
eng
Rights
울산대학교 논문은 저작권에 의해 보호 받습니다.
Appears in Collections:
Computer Engineering & Information Technology > 2. Theses (Ph.D)
공개 및 라이선스
  • 공개 구분공개
파일 목록

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.