Hybrid Deep Learning-Reinforcement Learning for Adaptive Human-Robot Task Allocation in Industry 5.0
Abstract
Human-Robot Collaboration (HRC) is pivotal for flexible, worker-centric manufacturing in Industry 5.0, yet dynamic task allocation remains difficult because operator states-fatigue and skill-fluctuate abruptly. I address this gap with a hybrid framework that couples real-time perception and double-estimating reinforcement learning. A Convolutional Neural Network (CNN) classifies nine fatigue-skill combinations from synthetic physiological cues (heart-rate, blink rate, posture, wrist acceleration); its outputs feed a Double Deep Q-Network (DDQN) whose state vector also includes task-queue and robot-status features. The DDQN optimises a multi-objective reward balancing throughput, workload and safety and executes at 10 Hz within a closed-loop pipeline implemented in MATLAB R2025a and RoboDK v5.9. Benchmarking on a 1000-episode HRC dataset (2500 allocations
Más información
| Título según WOS: | ID WOS:001559938200001 Not found in local WOS DB | 
| Título de la Revista: | SYSTEMS | 
| Volumen: | 13 | 
| Número: | 8 | 
| Editorial: | MDPI AG | 
| Fecha de publicación: | 2025 | 
| DOI: | 
 10.3390/systems13080631  | 
| Notas: | ISI |