Replies: 3 comments
-
Unfortunately we don't have good documentation on these parameters internally, and I think we need help from @petergjoel to completely understand how your desired results. |
Beta Was this translation helpful? Give feedback.
-
Most of the parameters have a direct equivalent in Fig. 2 of On Time with Minimal Expected Cost!. For your particular case, you need to limit to one iteration, as per your example.
This assume that your model ALWAYS reaches the trace-termination condition. |
Beta Was this translation helpful? Give feedback.
-
Setting |
Beta Was this translation helpful? Give feedback.
-
By default when using reinforcement learning to generate a strategy using
minE
ormaxE
queries, UPPAAL automatically chooses a number of traces to train on. On a very simple example, it seems to default to 9000 runs.However, for experiments it can be beneficial to fix the number of training episodes while changing other variables. How can this be done? I have previously been trying something like these learning parameters to train for 10 traces only:
But this (seemingly) results in twice that amount of episodes being trained for.
Is the number in the result box correct, and should I put in half as many training episodes as I want in the boxes? Or is the reporting misleading somehow, and it actually trained for the 10 traces I wanted?
Beta Was this translation helpful? Give feedback.
All reactions