Control of heap leach piles using deep reinforcement learning

Canales, Claudio; Diaz-Quezada, Simon; Leiva, Francisco; Estay, Humberto; Ruiz-del-Solar, Javier

Abstract

In this study, we propose a novel methodology for the automatic control of heap leaching by means of policies obtained using Reinforcement Learning (RL). This methodology models the leaching dynamics as a Markov Decision Process (MDP) whose reward function captures the economic profit of the heap leaching operation. As a case study, the leaching process of copper oxide heaps is simulated and controlled under various conditions. Results show that controlling this process using the proposed approach outperforms a heuristic control strategy that emulates real mining operations by increasing recovery rates by 2.25 times, reducing water consumption by 32.4% and acid consumption by 19.9%, and enhancing economic returns by 17.5%. This approach highlights the robustness of a Deep Reinforcement Learning (DRL) policy in heap leaching operations under significant out -of -distribution (OOD) conditions, demonstrating its adaptability and effectiveness under diverse and unpredictable conditions. Furthermore, this research highlights the potential for this methodology to be applied to other leachable ores, as it could reduce the overall environmental impact of this operation by using fewer resources.

Más información

Título según WOS: Control of heap leach piles using deep reinforcement learning
Título según SCOPUS: ID SCOPUS_ID:85191335413 Not found in local SCOPUS DB
Título de la Revista: MINERALS ENGINEERING
Volumen: 212
Editorial: PERGAMON-ELSEVIER SCIENCE LTD
Fecha de publicación: 2024
DOI:

10.1016/J.MINENG.2024.108707

Notas: ISI, SCOPUS