Abstract: In this work a visual and reactive contour following
behaviour is learned by reinforcement. With artificial vision the
environment is perceived in 3D, and it is possible to avoid obstacles
that are invisible to other sensors that are more common in mobile
robotics. Reinforcement learning reduces the need for intervention in
behaviour design, and simplifies its adjustment to the environment,
the robot and the task. In order to facilitate its generalisation to other
behaviours and to reduce the role of the designer, we propose a
regular image-based codification of states. Even though this is much
more difficult, our implementation converges and is robust. Results
are presented with a Pioneer 2 AT on a Gazebo 3D simulator.
Abstract: A self tuning PID control strategy using reinforcement
learning is proposed in this paper to deal with the control of wind
energy conversion systems (WECS). Actor-Critic learning is used to
tune PID parameters in an adaptive way by taking advantage of the
model-free and on-line learning properties of reinforcement learning
effectively. In order to reduce the demand of storage space and to
improve the learning efficiency, a single RBF neural network is used
to approximate the policy function of Actor and the value function of
Critic simultaneously. The inputs of RBF network are the system
error, as well as the first and the second-order differences of error.
The Actor can realize the mapping from the system state to PID
parameters, while the Critic evaluates the outputs of the Actor and
produces TD error. Based on TD error performance index and
gradient descent method, the updating rules of RBF kernel function
and network weights were given. Simulation results show that the
proposed controller is efficient for WECS and it is perfectly
adaptable and strongly robust, which is better than that of a
conventional PID controller.