Texto completo
|
PDF (Portable Document Format)
- Se necesita un visor de ficheros PDF, como GSview, Xpdf o Adobe Acrobat Reader
Descargar (621kB) |
ORCID: https://orcid.org/0009-0001-3171-2583, García Sánchez, Álvaro
ORCID: https://orcid.org/0000-0001-5774-1430, Espiga Fernández, Francisco
ORCID: https://orcid.org/0009-0004-3072-6934 and González-Santander de la Cruz, Guillermo
ORCID: https://orcid.org/0009-0007-9117-3554
(2025).
Deep Reinforcement Learning for Intraday Multireservoir Hydropower Management.
"Mathematics", v. 13
(n. 1);
p. 151.
ISSN 22277390.
https://doi.org/10.3390/math13010151.
| Título: | Deep Reinforcement Learning for Intraday Multireservoir Hydropower Management |
|---|---|
| Autor/es: |
|
| Tipo de Documento: | Artículo |
| Título de Revista/Publicación: | Mathematics |
| Fecha: | 3 Enero 2025 |
| ISSN: | 22277390 |
| Volumen: | 13 |
| Número: | 1 |
| Materias: | |
| ODS: | |
| Palabras Clave Informales: | Daily optimization; Hydropower generation; Mixed integer linear programming; Multireservoir; Reinforcement Learning |
| Escuela: | E.T.S.I. Industriales (UPM) |
| Departamento: | Ingeniería de Organización, Administración de Empresas y Estadística |
| Licencias Creative Commons: | Reconocimiento |
|
PDF (Portable Document Format)
- Se necesita un visor de ficheros PDF, como GSview, Xpdf o Adobe Acrobat Reader
Descargar (621kB) |
This study investigates the application of Reinforcement Learning (RL) to optimize intraday operations of hydropower reservoirs. Unlike previous approaches that focus on long-term planning with coarse temporal resolutions and discretized state-action spaces, we propose an RL framework tailored to the Hydropower Reservoirs Intraday Economic Optimization problem. This framework manages continuous state-action spaces while accounting for fine-grained temporal dynamics, including dam-to-turbine delays, gate movement constraints, and power group operations. Our methodology evaluates three distinct action space formulations (continuous, discrete, and adjustments) implemented using modern RL algorithms (A2C, PPO, and SAC). We compare them against both a greedy baseline and Mixed-Integer Linear Programming (MILP) solutions. Experiments on real-world data from a two-reservoir system and a simulated six-reservoir system demonstrate that while MILP achieves superior performance in the smaller system, its performance degrades significantly when scaled to six reservoirs. In contrast, RL agents, particularly those using discrete action spaces and trained with PPO, maintain consistent performance across both configurations, achieving considerable improvements with less than one second of execution time. These results suggest that RL offers a scalable alternative to traditional optimization methods for hydropower operations, particularly in scenarios requiring real-time decision making or involving larger systems.
| ID de Registro: | 88711 |
|---|---|
| Identificador DC: | https://oa.upm.es/88711/ |
| Identificador OAI: | oai:oa.upm.es:88711 |
| URL Portal Científico: | https://portalcientifico.upm.es/es/ipublic/item/10310748 |
| Identificador DOI: | 10.3390/math13010151 |
| URL Oficial: | https://www.mdpi.com/2227-7390/13/1/151 |
| Depositado por: | iMarina Portal Científico |
| Depositado el: | 09 Abr 2025 08:32 |
| Ultima Modificación: | 09 Abr 2025 08:32 |
Publicar en el Archivo Digital desde el Portal Científico