Respondida
Question about using the nonlinear model predictive control toolbox
nlmpcmove function uses fmincon from Optimization Toolbox. Pleaase take a look at this doc page for more details

más de 3 años hace | 0

Respondida
Adaptive MPC for LPV systems
This is possible with Model Predictive Control Toolbox yes, please take a look at this example.

más de 3 años hace | 0

Respondida
Is it possible to pass simulink NMPC parameters as struct?
Please take a look at my answer here

más de 3 años hace | 0

| aceptada

Respondida
Get nonlinear model from Simulink to Matlab
Hi Daniel, Unfortunately Model Predictive Control Toolbox does not currently support using a Simulink model directly as a predi...

más de 3 años hace | 0

| aceptada

Respondida
Binary Manipulated Variable in NLMPC
Unfortunately Model Predictive Control Toolbox does not currently support mixed-integer problems for NLMPC.

más de 3 años hace | 0

Respondida
Import Timeseries on Simulink as raw structure and not as single data samples at certain simulation time
It seems to me that you want to implement the MPC preview capability in Simulink. This example actually has a custom previewer b...

más de 3 años hace | 0

Respondida
NLMPC state function parameters
Hi, Given that it is not possible to package all variables into a struct and feed them into the MPC controller that way (it's n...

más de 3 años hace | 0

| aceptada

Respondida
Can I Generate C Code for TD3 or DDPG agent?
Yes. Here is an example

más de 3 años hace | 0

Respondida
Applying reinforcement learning with two continuous actions. During training one varies but the other is virtually static.
You should be able to use a single agent for this task. Since you are using DDPG, the first thing I would check is whether the n...

más de 3 años hace | 0

Respondida
How the generated parameters of agent are related PI/PID controller in reinforcement learning algorithm
This example uses an actor network that works as a PI controller. This is possible because we are using a customer layer called ...

más de 3 años hace | 0

| aceptada

Respondida
Constantly high negative reward in RL agent
You need to see what the actions generated by the RL Agent block are and how they affect the quadrotor dynamics. That's what it ...

más de 3 años hace | 0

| aceptada

Respondida
Load PPO agent and check layer
The neural network architectures are not immediately available from the agent object. You can extract and inspect layers using g...

más de 3 años hace | 0

Respondida
cumulative reward in RL AGENT block.
It's not very clear what you are asking, but the cumulative reward changes during each episode. It also depends on the sample ti...

más de 3 años hace | 0

Respondida
Sending episode number to simulink during training
The training stats are not available until after training is done, so you cannot use trainingStats.EpisodeIndex during training....

más de 3 años hace | 0

| aceptada

Respondida
How to modify actions in experiences during a reinforcement learning training
If you are working in Simulink, you can use the "Last Action" port in the RL Agent block to indicate what was the action that wa...

casi 4 años hace | 1

Respondida
How to get the actor network of a trained policy gradient agent?
Hello, To get the neural network model you can use net = getModel(getActor(agent)) To get learnable parameters you can use g...

casi 4 años hace | 0

| aceptada

Respondida
How to see actions when using the train() function in RL tool box.
Hello, To log action data throughout an episode, you would need to do so from inside the step function of your environment. You...

casi 4 años hace | 1

| aceptada

Respondida
Training Quadrotor using PPO agent
Hello, There are multiple things not set up properly, including: 1) The isdone flag seems to be 1 all the time leading to epis...

alrededor de 4 años hace | 0

Respondida
How to train RL-DQN agent with varying environment?
What you are describing is actually pretty standard process to create robust policies. To change the driving profiles, you can u...

casi 5 años hace | 2

| aceptada

Respondida
Editing the Q-table before Training in Basic Grid World?
Hello, Please take a look at this link that mentions how you can initialize the table.

casi 5 años hace | 0

| aceptada

Respondida
Could I learn from past data INCLUDING actions? Could I make vector with actions to be used in a certain order?
Hello, If the historical observations do not depend on the actions taken, (think of stock values, or historical power demand), ...

casi 5 años hace | 1

| aceptada

Respondida
update reinforcement policy.m weights
Hello, When you want to perform inference on an RL policy, there is no need to consider rewards. The trained policy already kno...

casi 5 años hace | 0

| aceptada

Respondida
I believe the RL environment template creator has an error in the reset function but I'm not sure
Hello, You are correct the order is wrong. That being said, the order of states depends on your dynamics and how you set up the...

casi 5 años hace | 0

| aceptada

Respondida
What exactly is Episode Q0? What information is it giving?
Q0 is calculated by performing inference on the critic at the beginning of each episode. Effectively, it is a metric that tells ...

casi 5 años hace | 2

| aceptada

Respondida
Resume training of a DQN agent. How to avoid Epsilon from being reset to max value?
Hello, This is currently not possible, but it is a great enhancement idea. I have informed the developers about your request an...

casi 5 años hace | 0

| aceptada

Respondida
Reinforcement learning with Simulink and Simscape
Even outside the thermal domain, you most likely need to start with a simulation model. RL does not need to build that model nec...

casi 5 años hace | 0

Respondida
RL training result very different from the result of 'sim'
Please see this post that explains why simulation results may differ during training and after training. If the simulation resu...

casi 5 años hace | 0

| aceptada

Respondida
RL in dynamic environment
The following example seems relevant, please take a look: https://www.mathworks.com/help/robotics/ug/avoid-obstacles-using-rein...

casi 5 años hace | 0

| aceptada

Respondida
MPC Controller giving nice performance during designing but fails on testing
Hello, It sounds to me that the issue is with the linearized model. When you are exporting the controller from MPC Designer, yo...

casi 5 años hace | 0

Respondida
What is in a reinforcement learning saved agent .mat file
Why don't you load the file and check? When you saved the agen tin the .mat file, did you save anything else with it? Are you m...

casi 5 años hace | 0

Cargar más