The deployment of microgrids could be fostered by control systems that do not require very complex modelling, calibration, prediction and/or optimisation processes. This paper explores the application of Reinforcement Learning (RL) techniques for the operation of a microgrid. The implemented Deep Q-Network (DQN) can learn an optimal policy for the operation of the elements of an isolated microgrid, based on the interaction agent-environment when particular operation actions are taken in the microgrid components. In order to facilitate the scaling-up of this solution, the algorithm relies exclusively on historical data from past events, and therefore it does not require forecasts of the demand or the renewable generation. The objective is to minimise the cost of operating the microgrid, including the penalty of non-served power. This paper analyses the effect of considering different definitions for the state of the system by expanding the set of variables that define it. The obtained results are very satisfactory as it can be concluded by their comparison with the perfect-information optimal operation computed with a traditional optimisation model, and with a Naive model.
Palabras clave: machine learning; microgrids; optimisation methods; power systems; reinforcement learning
Energies. Volumen: 13 Número: 11 Páginas: 2830-1-2830-19
Índice de impacto JCR y cuartil WoS: 2.707 - Q3 (2018)
Referencia DOI: 10.3390/en13112830
Publicado en papel: Junio 2020. Publicado on-line: Junio 2020.
D. Domínguez-Barbero, J. García-González, M.A. Sanz-Bobi, E.F. Sánchez-Úbeda. Optimising a microgrid system by deep reinforcement learning techniques. Energies. vol. 13, no. 11, pp. 2830-1-2830-19, Junio 2020. [Online: Junio 2020]