-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Code related #1
Comments
NOW! |
Terrific! Thanks a lot. Could you specify the version of mmdetection btw? |
And the pytorch version, it seems like you have pytorch>=2.0.0 |
System environment:
|
yes, I try to catch up with the rapid world. |
Hi, while I went through your code, I encountered some issues, hope you could help me out :)
|
The refinement aware gradient is equivalent to the truncated bptt to some extend, cutting off the higher order terms of the rnn iterations. Due to that then each supervision is independent, we can use gradient accumulation between each supervision to avoid the extra memory consumption, but the Autograd in pytorch will push the gradient calculated in single supervision to every parameters, resulting serval back pass to backbone though, so I use this hook to stash gradient to mlvl features, the last backward of the supervision will restore the stashed gradient, and bring stashed gradient to backbone weights |
For the question 2, yes, the RAG formulation is derived from 2-step unrolled fix-point formulation in paper, the implementation in codebase is that 2-step unrolled fix-point. The equation mainly helps to analyze the reason why two-step better than simple estimation method used in deq-flow. You can find the pesudo code in appendix. |
Thanks for the excellent work. I wonder the release time of the source code.
Looking forward to your reply. : )
The text was updated successfully, but these errors were encountered: