You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi there, the readme says that distributed parallel sampling can be implemented. But it doesn't look like this feature is presented in examples, for example the td3_script.py.
In issure #24 , you said "You can start the Server Manager once and then call env.make() multiple times , with the algorithm that we are using right now we have multiple workers running in parallel and each worker is calling env.make() and the Server Manager spawns a new instance of the env. ". Does you mean that there are only certain specific algorithms which contains multiple workers like D4PG could pararrel sample? It's confused that in openai gym, parallel envs could be accomplished by VecEnv based on multithreading of python, so that any RL algorithm could make it.
What can i have to do to start a parallel envs with the help of robo-gym? Is there any examples or documents to reference?
I would really appreciate it if someone can help me out. Thanks in advance.
The text was updated successfully, but these errors were encountered:
D4PG is one option that we have been using internally.
With stable-baselines3 it is also possible to simply wrap many robo-gym envs in a SubprocVecEnv, like in the following snippet - please just take it as a cheap example, not a recommendation.
Notes:
You could of course use different IPs for the environments, corresponding to server managers on different machines. With the single IP you would get multiple robot servers in parallel created by the same server manager.
Considering that parallel steps are synchronized, it is not impossible (though maybe insignificant) that bottlenecks could deteriorate the results by reducing the rate of steps while the simulations keep running at their own pace. Handling the individual environments in separate workers can be more precise and efficient. It also allows you to react to termination or truncation individually, or to collect step results for asynchronous bulk processing.
Hi there, the readme says that distributed parallel sampling can be implemented. But it doesn't look like this feature is presented in examples, for example the td3_script.py.
In issure #24 , you said "You can start the Server Manager once and then call env.make() multiple times , with the algorithm that we are using right now we have multiple workers running in parallel and each worker is calling env.make() and the Server Manager spawns a new instance of the env. ". Does you mean that there are only certain specific algorithms which contains multiple workers like D4PG could pararrel sample? It's confused that in openai gym, parallel envs could be accomplished by VecEnv based on multithreading of python, so that any RL algorithm could make it.
What can i have to do to start a parallel envs with the help of robo-gym? Is there any examples or documents to reference?
I would really appreciate it if someone can help me out. Thanks in advance.
The text was updated successfully, but these errors were encountered: