2011
DOI: 10.1016/j.asoc.2009.10.018
|View full text |Cite
|
Sign up to set email alerts
|

Model-free control based on reinforcement learning for a wastewater treatment problem

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
15
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 52 publications
(17 citation statements)
references
References 25 publications
(48 reference statements)
0
15
0
Order By: Relevance
“…A real-world implementation of the RL control method is made possible by applying data-mining algorithms to construct models from data and generate synthetic data for pre-training the RL algorithm, without the need to physically interact with the system. This also represents an original contribution compared to other control problems like [25,24]. In order to address this problem and in a different context for building heating systems, an interesting and different approach is proposed by Costanzo et al [35], and inspired by the work in [36], which consists in using a neural network to create virtual data that are used together with experimental data to approximate the Q-function (state-action value function) of fitted Q-learning and reduce the amount of system interaction required to learn a "good" policy.…”
Section: Related Work and Contributionsmentioning
confidence: 94%
See 1 more Smart Citation
“…A real-world implementation of the RL control method is made possible by applying data-mining algorithms to construct models from data and generate synthetic data for pre-training the RL algorithm, without the need to physically interact with the system. This also represents an original contribution compared to other control problems like [25,24]. In order to address this problem and in a different context for building heating systems, an interesting and different approach is proposed by Costanzo et al [35], and inspired by the work in [36], which consists in using a neural network to create virtual data that are used together with experimental data to approximate the Q-function (state-action value function) of fitted Q-learning and reduce the amount of system interaction required to learn a "good" policy.…”
Section: Related Work and Contributionsmentioning
confidence: 94%
“…Syafiie et al proposed a model-free control approach for advanced oxidation processes (or Fenton process), since according to the authors it is extremely difficult do develop a precise mathematical model and the system is subject to several uncertainties and time-evolving characteristics [24]. As an alternative to proportional-integral-derivative (PID) controllers, Hernández-del-Olmo et al explored RL for oxygen control in the N-ammonia removal process, whose main objective was to minimize WTTP operational costs (including energy costs) [25].…”
Section: Related Work and Contributionsmentioning
confidence: 99%
“…A biosensor based on gas phase monitoring was designed to measure the hydrogen peroxide concentrations in Modrzejewska, Guwy, Dinsdale, and Hawkes (2007). It is worth to mention also a specific continuous monitoring of hydrogen peroxide formed in situ was proposed by Oh, Kim, Kang, Oh, and Kang (2005), or a model-free learning control to deal with the chemical characteristic variations with time is given by Syafiie, Tadeo, Martinez, and Alvarez (2011).…”
Section: Introductionmentioning
confidence: 99%
“…For a class of unknown general nonlinear systems such as WWTP, adaptive dynamic programming (ADP)-also known as reinforcement learning, has attracted more and more attention of experts recently [8][9][10][11][12][13][14][15][16][17][18][19][20][21][22][23][24]. The core of ADP is the approximation of Bellman's equation or the Hamilton-Jacobi-Bellman (HJB) equation by learning [10,12,20].…”
Section: Introductionmentioning
confidence: 99%