You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I want to use PPO and A2C with a custom LSTM and transformer network. PPO only have native support for "MlpPolicy" "CnnPolicy" and "MultiInputPolicy". Can I still use "MlpPolicy" but specify a custom LSTM or transformer network, or can it purely be used with networks that only have linear layers? Does it mess up the training if I specify a LSTM network?
Checklist
I have checked that there is no similar issue in the repo
I can recommend to use a custom features extractor that does everything you want. However, you might want to adjust the MLP heads for critic and actor accordingly - but still using the MlpPolicy.
❓ Question
I want to use PPO and A2C with a custom LSTM and transformer network. PPO only have native support for "MlpPolicy" "CnnPolicy" and "MultiInputPolicy". Can I still use "MlpPolicy" but specify a custom LSTM or transformer network, or can it purely be used with networks that only have linear layers? Does it mess up the training if I specify a LSTM network?
Checklist
The text was updated successfully, but these errors were encountered: