Learning to Break Rocks With Deep Reinforcement Learning

Abstract

This work proposes a scheme for learning how to break rocks with an impact hammer. The problem is formulated as a Partially Observable Markov's Decision Process, and then solved through deep reinforcement learning. We propose a simple formulation, requiring only a basic sensorization of the hammer's manipulator, and involving just two discrete actions. We use Dueling Double Deep-Q Networks to parameterize the policy, and wield it with an auxiliary output. The proposed auxiliary task is also trained in simulation, and allows deciding when to stop the operation by detecting the absence of a rock from the observed joints' movement. The resulting policy is tested in a real world experimental environment, using a Bobcat E10 mini-excavator, and various rock types. The results show that a good performance can be obtained in a safe, and robust manner.

Más información

Título según WOS: Learning to Break Rocks With Deep Reinforcement Learning
Título de la Revista: IEEE ROBOTICS AND AUTOMATION LETTERS
Volumen: 8
Número: 2
Editorial: IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
Fecha de publicación: 2023
Página de inicio: 1077
Página final: 1084
DOI:

10.1109/LRA.2023.3236562

Notas: ISI