<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-17T16:41:46Z</responseDate><request verb="GetRecord" identifier="oai:riubu.ubu.es:10259/7421" metadataPrefix="etdms">https://riubu.ubu.es/oai/request</request><GetRecord><record><header><identifier>oai:riubu.ubu.es:10259/7421</identifier><datestamp>2024-05-14T11:44:00Z</datestamp><setSpec>com_10259_6155</setSpec><setSpec>com_10259_4266</setSpec><setSpec>com_10259.4_106</setSpec><setSpec>com_10259_2604</setSpec><setSpec>col_10259_6156</setSpec></header><metadata><thesis xmlns="http://www.ndltd.org/standards/metadata/etdms/1.0/" xmlns:doc="http://www.lyncode.com/xoai" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.ndltd.org/standards/metadata/etdms/1.0/ http://www.ndltd.org/standards/metadata/etdms/1.0/etdms.xsd">
<title>Wind turbine pitch reinforcement learning control improved by PID regulator and learning observer</title>
<creator>Sierra Garcia, Jesús Enrique</creator>
<creator>Santos, Matilde</creator>
<creator>Pandit, Ravi</creator>
<subject>Intelligent control</subject>
<subject>Reinforcement learning</subject>
<subject>Learning observer</subject>
<subject>Pitch control</subject>
<subject>Wind turbines</subject>
<description>Wind turbine (WT) pitch control is a challenging issue due to the non-linearities of the wind device and its&#xd;
complex dynamics, the coupling of the variables and the uncertainty of the environment. Reinforcement learn-&#xd;
ing (RL) based control arises as a promising technique to address these problems. However, its applicability&#xd;
is still limited due to the slowness of the learning process. To help alleviate this drawback, in this work we&#xd;
present a hybrid RL-based control that combines a RL-based controller with a proportional–integral–derivative&#xd;
(PID) regulator, and a learning observer. The PID is beneficial during the first training episodes as the RL based&#xd;
control does not have any experience to learn from. The learning observer oversees the learning process by&#xd;
adjusting the exploration rate and the exploration window in order to reduce the oscillations during the training&#xd;
and improve convergence. Simulation experiments on a small real WT show how the learning significantly&#xd;
improves with this control architecture, speeding up the learning convergence up to 37%, and increasing the&#xd;
efficiency of the intelligent control strategy. The best hybrid controller reduces the error of the output power&#xd;
by around 41% regarding a PID regulator. Moreover, the proposed intelligent hybrid control configuration has&#xd;
proved more efficient than a fuzzy controller and a neuro-control strategy.</description>
<date>2023-02-08</date>
<date>2023-02-08</date>
<date>2022-05</date>
<type>info:eu-repo/semantics/article</type>
<identifier>0952-1976</identifier>
<identifier>http://hdl.handle.net/10259/7421</identifier>
<identifier>10.1016/j.engappai.2022.104769</identifier>
<language>eng</language>
<relation>Engineering Applications of Artificial Intelligence. 2022, V. 111, 104769</relation>
<relation>https://doi.org/10.1016/j.engappai.2022.104769</relation>
<relation>info:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2017-2020/RTI2018-094902-B-C21/ES/ANALISIS Y CONTROL DE UN DISPOSITIVO FLOTANTE HIBRIDO DE ENERGIA EOLICA Y MARINA/</relation>
<rights>http://creativecommons.org/licenses/by-nc-nd/4.0/</rights>
<rights>info:eu-repo/semantics/openAccess</rights>
<rights>Attribution-NonCommercial-NoDerivatives 4.0 Internacional</rights>
<publisher>Elsevier</publisher>
</thesis></metadata></record></GetRecord></OAI-PMH>