PPO agent applied to ACC model

I'm applying the PPO algorithm to the ACC model in this DDPG example:Train DDPG Agent for Adaptive Cruise Control
How ever the agent went wrong at 50th episode with error:
An error occurred while simulating "rlACCMdl" with the agent "agent".
[varargout{1},varargout{2}] = simWithPolicy(this.Env,this.Agent,simOpts);
[varargout{1:nargout}] = runImpl(this);
[varargout{1:nargout}] = run(task);
[this.Outputs{1:getNumOutputs(this)}] = internal_run(this);
runDirect(this);
runScalarTask(task);
run(seriestaskspec);
run(trainer);
train(this);
TrainingStatistics = run(trainMgr);
Caused by:
Invalid input argument type or size such as observation, reward, isdone or loggedSignals.
Standard deviation must be nonnegative. Ensure your representation always outputs nonnegative values for outputs that correspond to the standard deviation.
Is there anybody could solve this problem? Thanks for your help!

1 comentario

qun wang
qun wang el 26 de Oct. de 2021
您好,请问这个问题你解决了嘛?我最近也在用PPO遇到类似的问题,可否讨论一下?期待回复

Iniciar sesión para comentar.

Respuestas (1)

Emmanouil Tzorakoleftherakis
Emmanouil Tzorakoleftherakis el 3 de Sept. de 2020

0 votos

Hello,
Can you make sure that you set up your actor following a structure similar to this one? It seems that your variance path is not set up properly and gives negative values.

Preguntada:

el 22 de Ag. de 2020

Comentada:

el 26 de Oct. de 2021

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by