Made byBobr AI

MoCap-to-IMU Transfer Learning for Industrial Exoskeletons

Explore a teacher-student transfer learning pipeline for industrial exoskeleton intent prediction using knowledge distillation from MoCap to IMU sensors.

#exoskeleton#transfer-learning#knowledge-distillation#imu-sensors#mocap#robotics#industrial-automation#ai
Watch
Pitch

MSC THESIS PROPOSAL

Robust MoCap-to-IMU Transfer for Industrial Exoskeleton Intent Modeling Under Domain Shift

Alba Huti

Exoskeleton Human

EMPOWER Project — Industrial Human-Robot Collaboration

Made byBobr AI

Problem Statement

Industrial exoskeletons need real-time intent prediction — but current approaches rely on lab-grade sensors that don't scale to real-world deployment.

Transferability

Transferability Gap

Current research relies on MoCap & EMG; no standardized pipeline exists for transferring knowledge to IMU-only wearable deployment models.

Robustness

Robustness Gap

Performance degrades under domain shift (across users, tasks, sensor configs) but rarely addressed with explicit mitigation strategies.

Acceptance

Acceptance Gap

Many studies under-measure usability; industrial adoption depends on comfort, perceived usefulness, and predictable behavior.

EMPOWER Project — Industrial Human-Robot Collaboration

Made byBobr AI

Research Objective

To develop a robust teacher–student transfer learning pipeline that distills MoCap-based kinematic knowledge into a compact, IMU-only model for real-time exoskeleton intent prediction in industrial settings.

Teacher Model

High-capacity temporal network trained on full-body MoCap kinematics (MS-G3D skeleton graph).

Rich Supervision — Lab Only

Distillation

Student Model

Compact TCN/LSTM using 2–4 wearable IMUs only.

Edge Deployment — Industrial Ready
Knowledge Distillation
Cross-Modality Transfer
Edge Feasibility

EMPOWER Project — Industrial Human-Robot Collaboration

Made byBobr AI

Research Questions

RQ1
Can MoCap data, used as privileged supervision during training, improve an IMU-only model's performance compared to a model trained solely on wearable IMU data?
RQ2
Can a model using only minimal wearable IMU sensors (2–4) accurately predict user intent or generate outputs relevant for exoskeleton support and assistance?
RQ3
Can laboratory-quality kinematic information from MoCap be effectively distilled into a practical IMU-based system while maintaining robustness under real-world domain shifts (unseen users, scenarios, sensor placement variation, noise)?
Made byBobr AI

Methodology

Data

1. Data

  • LARa Dataset
    MoCap + IMU
    Logistics activities
  • OpenPack Dataset
    IMU only
    Packaging operations
Teacher AI

2. Teacher Training

  • MS-G3D Network
    Graph neural network
    Trained on MoCap 3D joints
  • Outputs:
    Intent class logits &
    biomechanical proxy
Distillation

3. Distillation

  • Task Loss (CE / MSE)
  • Distill Loss (KL, temp scale)
  • Feature Alignment (MSE)
Ltotal = αLtask + βLdistill + γLfeat
Student Deployment

4. Deployment

  • Compact Model
    TCN / LSTM architectures
  • 2–4 Wearable IMUs
    6-axis (Acc + Gyro)
  • Edge-Feasible
    Real-time inference support

Controlled Benchmarking

Evaluation track on predefined activities using LARa dataset

Deployment-Oriented

Real-world application focus evaluating on OpenPack dataset

EMPOWER Project — Industrial Human-Robot Collaboration

Made byBobr AI

Proposed Analysis Tools & Evaluation

Models & Tools

Teacher MS-G3D (Spatio-Temporal Graph CNN) on MoCap skeleton
Student Lightweight TCN / LSTM-based architecture
Baseline 1D CNN + LSTM (IMU-only, no distillation)

Optimizer

Adam (lr ≈ 0.001), Early Stopping, Mini-batch training

Sensor Configs Tested

5 configurations (torso only → torso + arms + legs)

Robustness Stress Tests

Neural Net Cross-Subject
Generalization
Arrows Cross-Task /
Scenario Shift
Database Cross-Dataset
(LARa → OpenPack)
Sensor Sensor Placement
(Noise, Rotations)

Key Metrics

Macro-F1
Accuracy
RMSE / MAE
Latency (ms/window)
Model Size
Robustness Curves

EMPOWER Project — Industrial Human-Robot Collaboration

Made byBobr AI

Calendar of Activities — Gantt Chart

Activity M1 M2 M3 M4 M5 M6 M7 M8 M9
1. Literature Review & Dataset Acquisition
2. Data Preprocessing & Baseline Implementation
(IMU-only CNN+LSTM)
3. Teacher Model Training
(MS-G3D on MoCap)
4. Knowledge Distillation Framework
(Student training)
5. Robustness Evaluation
(cross-user, cross-task, sensor perturbations)
6. Deployment Feasibility & Edge Testing
7. Writing & Thesis Submission
Made byBobr AI

Expected Contributions

1

Reproducible MoCap-to-IMU Transfer Pipeline

A concrete teacher–student distillation framework for temporal motion data, directly operationalizing key literature recommendations for industrial exoskeleton intent modeling.

2

Minimal Sensor Configuration Guidance

A clear performance vs. sensor-count trade-off curve identifying the "minimal sufficient" IMU setup for robust industrial deployment.

3

Robustness Under Domain Shift

Empirical evidence that MoCap-based privileged supervision improves not only peak accuracy but also generalization stability across users, tasks, and sensor perturbations.

4

Standardized Benchmarking Protocol

Clearly defined cross-user/cross-scenario splits, synthetic perturbations, and metrics (Macro-F1, RMSE, Latency, Model Size) to address evaluation fragmentation in the field.

5

Deployment-Oriented Insights

Quantitative links between sensor count, robustness, latency, and model size — transforming research into practical design guidance for embedded exoskeleton systems.

Made byBobr AI
Bobr AI

DESIGNER-MADE
PRESENTATION,
GENERATED FROM
YOUR PROMPT

Create your own professional slide deck with real images, data charts, and unique design in under a minute.

Generate For Free

MoCap-to-IMU Transfer Learning for Industrial Exoskeletons

Explore a teacher-student transfer learning pipeline for industrial exoskeleton intent prediction using knowledge distillation from MoCap to IMU sensors.

Robust MoCap-to-IMU Transfer for Industrial Exoskeleton Intent Modeling Under Domain Shift

MSC THESIS PROPOSAL

Alba Huti

EMPOWER Project — Industrial Human-Robot Collaboration

Problem Statement

Industrial exoskeletons need real-time intent prediction — but current approaches rely on lab-grade sensors that don't scale to real-world deployment.

Transferability Gap

Current research relies on MoCap & EMG; no standardized pipeline exists for transferring knowledge to IMU-only wearable deployment models.

Robustness Gap

Performance degrades under domain shift (across users, tasks, sensor configs) but rarely addressed with explicit mitigation strategies.

Acceptance Gap

Many studies under-measure usability; industrial adoption depends on comfort, perceived usefulness, and predictable behavior.

EMPOWER Project — Industrial Human-Robot Collaboration

Research Objective

To develop a robust teacher–student transfer learning pipeline that distills MoCap-based kinematic knowledge into a compact, IMU-only model for real-time exoskeleton intent prediction in industrial settings.

Teacher Model

High-capacity temporal network trained on full-body MoCap kinematics (MS-G3D skeleton graph).

Rich Supervision — Lab Only

Student Model

Compact TCN/LSTM using 2–4 wearable IMUs only.

Edge Deployment — Industrial Ready

Knowledge Distillation

Cross-Modality Transfer

Edge Feasibility

EMPOWER Project — Industrial Human-Robot Collaboration

Research Questions

RQ1

Can MoCap data, used as privileged supervision during training, improve an IMU-only model's performance compared to a model trained solely on wearable IMU data?

RQ2

Can a model using only minimal wearable IMU sensors (2–4) accurately predict user intent or generate outputs relevant for exoskeleton support and assistance?

RQ3

Can laboratory-quality kinematic information from MoCap be effectively distilled into a practical IMU-based system while maintaining robustness under real-world domain shifts (unseen users, scenarios, sensor placement variation, noise)?

Methodology

EMPOWER Project — Industrial Human-Robot Collaboration

Proposed Analysis Tools & Evaluation

EMPOWER Project — Industrial Human-Robot Collaboration

Calendar of Activities — Gantt Chart

Activity

M1

M2

M3

M4

M5

M6

M7

M8

M9

1. Literature Review & Dataset Acquisition

2. Data Preprocessing & Baseline Implementation

(IMU-only CNN+LSTM)

3. Teacher Model Training

(MS-G3D on MoCap)

4. Knowledge Distillation Framework

(Student training)

5. Robustness Evaluation

(cross-user, cross-task, sensor perturbations)

6. Deployment Feasibility & Edge Testing

7. Writing & Thesis Submission

Expected Contributions

Reproducible MoCap-to-IMU Transfer Pipeline

A concrete teacher–student distillation framework for temporal motion data, directly operationalizing key literature recommendations for industrial exoskeleton intent modeling.

Minimal Sensor Configuration Guidance

A clear performance vs. sensor-count trade-off curve identifying the "minimal sufficient" IMU setup for robust industrial deployment.

Robustness Under Domain Shift

Empirical evidence that MoCap-based privileged supervision improves not only peak accuracy but also generalization stability across users, tasks, and sensor perturbations.

Standardized Benchmarking Protocol

Clearly defined cross-user/cross-scenario splits, synthetic perturbations, and metrics (Macro-F1, RMSE, Latency, Model Size) to address evaluation fragmentation in the field.

Deployment-Oriented Insights

Quantitative links between sensor count, robustness, latency, and model size — transforming research into practical design guidance for embedded exoskeleton systems.

  • exoskeleton
  • transfer-learning
  • knowledge-distillation
  • imu-sensors
  • mocap
  • robotics
  • industrial-automation
  • ai