Difference between using a rlContinuousGaussianActor and a rlContinuousDeterministicActor with a Gaussian Explorationmodel
Mostrar comentarios más antiguos
Hi,
Someone please explain the difference between using a rlContinuousGaussianActor and using a rlContinuousDeterministicActor with a Gaussian Explorationmodel (namely the GaussianActionNoise) in reinforcement learning with e.g. a rlTD3agent.
With possibly using the rlContinuousGaussianActor using the Gaussian Explorationmodel would not be impossible: What is the point/ use case of this combination?
Thank you.
1 comentario
Jonas Woeste
el 25 de Jun. de 2022
Respuestas (1)
Hi Jonas Woeste,
I understand that you wish to know the difference between "rlContinuousGaussianActor" and "rlContinuousDeterministicActor" with a Gaussian exploration model.
The choice between these two approaches depends on the specific problem and the trade-off between exploration and exploitation.
“rlContinuousGaussainActor” is suitable when exploration is crucial and stochastic actions are desired. “rlContinuousDeterministicActor” is suitable for providing a stable policy with controlled exploration. It strikes a balance between exploration and exploitation. To introduce exploration, a Gaussian exploration model like “GaussianActiveNoise” is used. The exploration model adds Gaussian noise to the actor's output action, making it slightly perturbed and allowing exploration.
In summary, the choice of the combination depends on the nature of the problem and the desired behaviour of the agent you are seeking.
Hope this helps!
Categorías
Más información sobre Reinforcement Learning en Centro de ayuda y File Exchange.
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!