-
Notifications
You must be signed in to change notification settings - Fork 74
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DRL algorithm with api #2528
Comments
cibot: Thank you for posting issue #2528. The person in charge will reply soon. |
|
Another question. |
Hello! Thank you for your question and concern. Here're my answer on your questions: |
You can do checkout and continue training process, but that's just not based on gradient saving. |
ok, thanks for your reply , |
For model copy, if there is no copy-constructor for model class and the default behavior does not do what you want, you may try "original.save()" and "cloned.load()". For Polyak update, it appears that the DQN application (or simple "reinforcement learning" app) has its own "custom" op. But I'm not too sure about this. I guess @jijoongmoon may answer this when he returns from trip. |
Hello, I checked the reinforcement learning app , |
Hello , Dear Contributors
I notice that the application DQN don't use the api .h file.
And there only exists defined loss function, so if I want to develop a DQN methods, I would like to ask you to confirm the following.
Or you have better advice.
The text was updated successfully, but these errors were encountered: