Deep Reinforcement Learning for Intraday Multireservoir Hydropower Management

Castro Freibott, Rodrigo ORCID: https://orcid.org/0009-0001-3171-2583, García Sánchez, Álvaro ORCID: https://orcid.org/0000-0001-5774-1430, Espiga Fernández, Francisco ORCID: https://orcid.org/0009-0004-3072-6934 and González-Santander de la Cruz, Guillermo ORCID: https://orcid.org/0009-0007-9117-3554 (2025). Deep Reinforcement Learning for Intraday Multireservoir Hydropower Management. "Mathematics", v. 13 (n. 1); p. 151. ISSN 22277390. https://doi.org/10.3390/math13010151.

Descripción

Título: Deep Reinforcement Learning for Intraday Multireservoir Hydropower Management
Autor/es:
Tipo de Documento: Artículo
Título de Revista/Publicación: Mathematics
Fecha: 3 Enero 2025
ISSN: 22277390
Volumen: 13
Número: 1
Materias:
ODS:
Palabras Clave Informales: Daily optimization; Hydropower generation; Mixed integer linear programming; Multireservoir; Reinforcement Learning
Escuela: E.T.S.I. Industriales (UPM)
Departamento: Ingeniería de Organización, Administración de Empresas y Estadística
Licencias Creative Commons: Reconocimiento

Texto completo

[thumbnail of 10310748.pdf] PDF (Portable Document Format) - Se necesita un visor de ficheros PDF, como GSview, Xpdf o Adobe Acrobat Reader
Descargar (621kB)

Resumen

This study investigates the application of Reinforcement Learning (RL) to optimize intraday operations of hydropower reservoirs. Unlike previous approaches that focus on long-term planning with coarse temporal resolutions and discretized state-action spaces, we propose an RL framework tailored to the Hydropower Reservoirs Intraday Economic Optimization problem. This framework manages continuous state-action spaces while accounting for fine-grained temporal dynamics, including dam-to-turbine delays, gate movement constraints, and power group operations. Our methodology evaluates three distinct action space formulations (continuous, discrete, and adjustments) implemented using modern RL algorithms (A2C, PPO, and SAC). We compare them against both a greedy baseline and Mixed-Integer Linear Programming (MILP) solutions. Experiments on real-world data from a two-reservoir system and a simulated six-reservoir system demonstrate that while MILP achieves superior performance in the smaller system, its performance degrades significantly when scaled to six reservoirs. In contrast, RL agents, particularly those using discrete action spaces and trained with PPO, maintain consistent performance across both configurations, achieving considerable improvements with less than one second of execution time. These results suggest that RL offers a scalable alternative to traditional optimization methods for hydropower operations, particularly in scenarios requiring real-time decision making or involving larger systems.

Proyectos asociados

Tipo
Código
Acrónimo
Responsable
Título
Gobierno de España
PID2022-137748OB-C31
Sin especificar
Sin especificar
Sin especificar
Gobierno de España
TSI-100408-2021
IA4TES
Sin especificar
Advanced Intelligent Technologies for Sustainable Energy Transition

Más información

ID de Registro: 88711
Identificador DC: https://oa.upm.es/88711/
Identificador OAI: oai:oa.upm.es:88711
URL Portal Científico: https://portalcientifico.upm.es/es/ipublic/item/10310748
Identificador DOI: 10.3390/math13010151
URL Oficial: https://www.mdpi.com/2227-7390/13/1/151
Depositado por: iMarina Portal Científico
Depositado el: 09 Abr 2025 08:32
Ultima Modificación: 09 Abr 2025 08:32