Build your own agent
On this page we quickly describe the steps we take before training our agents
Checklist
- Decide on Observation-Space and create an
ObservationManager, if the already implemented ones do not suffice - Create Reward Function and implement a
RewardManager, if nececssary - Check if config is as you'd like, most important configs to check are
configs/train.yamlandconfigs/rl_env/single_agent_env.yaml. Most importantly, set the managers in thesingle_agent_env.yaml:
defaults:
- obs_manager: next_point_obs.yaml
- reward_manager: next_point_rewards.yaml
- termination_manager: timeout.yaml
Note
Be sure to check the config-files in detail in order not to miss any important setting; this may really result in unnecessary troubleshooting.
Check your reward function
We highly suggest you check your reward function, using the testing-framework to avoid unncessary errors, or scaling issues, as errors in the reward functions are the most likely reason for lack of learning.
What to expect
When using a single environment and images in your observation-space, on a reasonably capable GPU, it tatkes around 5-10 hours to see results.