Hernández del Olmo, FélixGaudioso Vázquez, ElenaDormido Canto, RaquelDuro Carralero, Natividad2024-05-202024-05-2020180950-7051, eISSN 1872-7409https://doi.org/10.1016/j.knosys.2017.12.019https://hdl.handle.net/20.500.14468/12455Reinforcement learning problems involve learning by doing. Therefore, a reinforcement learning agent will have to fail sometimes (while doing) in order to learn. Nevertheless, even with this starting error, introduced at least during the non-optimal learning stage, reinforcement learning can be affordable in some domains like the control of a wastewater treatment plant. However, in wastewater treatment plants, trying to solve the day-to-day problems, plant operators will usually not risk to leave their plant in the hands of an inexperienced and untrained reinforcement learning agent. In fact, it is somewhat obvious that plant operators will require firstly to check that the agent has been trained and that it works as it should at their particular plant. In this paper, we present a solution to this problem by giving a previous instruction to the reinforcement learning agent before we let it act on the plant. In fact, this previous instruction is the key point of the paper. In addition, this instruction is given effortlessly by the plant operator. As we will see, this solution does not just solve the starting up problem of leaving the plant in the hands of an untrained agent, but it also improves the future performance of the agent.enAtribución 4.0 Internacionalinfo:eu-repo/semantics/openAccessTackling the start-up of a reinforcement learning agent for the control of wastewater treatment plantsartículo