PhD Defence Notice - Vladimir JoukovExport this event to calendar

Monday, June 7, 2021 — 9:30 AM EDT

Candidate: Vladimir Joukov

Title: Human motion estimation and controller learning

Date: June 7, 2021

Time: 9:30 AM

Place: REMOTE ATTENDANCE

Supervisor(s): Kulic, Dana (Adjunct) - Melek, William (Mechanical & Mechatronics Engineering)

 

Abstract:

Humans are capable of complex manipulation and locomotion tasks. They are able to achieve energy-efficient gait, reject disturbances, handle changing loads, and adapt to environmental constraints. Using inspiration from the human body, robotics researchers aim to develop systems with similar capabilities. Research suggests that humans minimize a task specific cost function when performing movements. In order to learn this cost function from demonstrations and incorporate it into a controller, it is first imperative to accurately estimate the expert motion the captured motions can then be analyzed to extract the objective function the expert was minimizing.

 

We propose a framework for human motion estimation from wearable sensors. Human body joints are modeled by matrix Lie groups, using special orthogonal groups SO(2) and SO(3) for joint pose and special Euclidean group SE(3) for base link pose representation. To estimate the human joint pose, velocity and acceleration, we provide the equations for employing the Extended Kalman Filter on Lie Groups, thus explicitly accounting for the non-Euclidean geometry of the state space. Incorporating interaction constraints with respect to the environment or within the participant allows us to track global body position without an absolute reference and ensure viable pose estimate. The algorithms are extensively validated in both simulation and real-world experiments.

 

Next, to learn underlying expert control strategies from the expert demonstrations we present a novel fast approximate multi-variate Gaussian Process regression. The method estimates the underlying cost function, without making assumptions on its structure. The computational efficiency of the approach allows for real time forward horizon prediction. Using a linear model predictive control framework we then reproduce the demonstrated movements on a robot. The learned cost function captures the variability in expert motion as well as the correlations between states, leading to a controller that both produces motions and reacts to disturbances in a human-like manner. The model predictive control formulation allows the controller to satisfy task and joint space constraints avoiding obstacles and self collisions, as well as torque constraints, ensuring operational feasibility. The approach is validated on the Franka Emika robot using real human motion exemplars.

Location 
REMOTE PARTICIPATION


,

S M T W T F S
29
30
31
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
1
2
  1. 2021 (130)
    1. December (10)
    2. November (18)
    3. October (10)
    4. September (4)
    5. August (12)
    6. July (4)
    7. June (5)
    8. May (10)
    9. April (25)
    10. March (14)
    11. February (10)
    12. January (9)
  2. 2022 (1)
    1. January (1)
  3. 2020 (246)
    1. December (20)
    2. November (19)
    3. October (15)
    4. September (15)
    5. August (16)
    6. July (32)
    7. June (29)
    8. May (32)
    9. April (27)
    10. March (13)
    11. February (20)
    12. January (16)
  4. 2019 (282)
  5. 2018 (150)
  6. 2017 (212)
  7. 2016 (242)
  8. 2015 (242)
  9. 2014 (268)
  10. 2013 (190)
  11. 2012 (31)