You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
On second thought, parallel training may be achieved without #11 by wrapping our current environment in a 'pool' wrapper.
This pool would have a manager or a cron job based on time or episodes that will periodically soft-sync models to a 'master' model to rapidly accumulate experience, assuming all jobs are normalized identically. This master model would then be distributed to env-threads in this pool for further (distributed) training.
Extend trainer/runner/environment from #11 for parallel training.
The text was updated successfully, but these errors were encountered: