-
Notifications
You must be signed in to change notification settings - Fork 67
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The reason why constant loss #2
Comments
Could you post your a rendered view image? |
Would you try inverting the views so that they have black background?
|
Thank you very much! |
@WeiTang114 I used the rendered views(given by your link) as input and still get a constant loss of approximately 3.69 even after 25 epochs! Could you tell me what might be the issue? Also, is there a way to visualize the loss/accuracy vs epoch? Thanks! |
|
@WeiTang114 |
@Priyam1994 Then I'm not very sure what problem it may be. |
In my perspective, the learning rate for deep layer (fc layers) should be multiplied by 10. It works for me. I recommend it for you @Priyam1994. |
@weiweisunWHU Could you tell me how you changed the learning rate only for the fc layers? |
@Priyam1994 |
@weiweisunWHU |
@weiweisunWHU @Priyam1994 @Xmen0123 Also, I found 0.0001 is more reliable for training. I've updated the code (commit b476e17f11bd540f4f962ae157f20c17067996b2). |
@youkaichao If simply inverting your images works, I'll consider adding an option such as "--white_background=True" 😅 |
@WeiTang114 But I'm puzzled. What's the difference between white and black? If I feed the MVCNN with white-background views, it's expected to identify the white-background views, isn't it? |
@youkaichao |
Now the problem is solved, I got an accuracy of 85%. Not state-of-art, but reasonably good. Thank you! I'll go fine-tuning now ^_^ |
@WeiTang114 Your suggestion worked and I could achieve a test accuracy of 88%. But it always results in the constant 3.69 loss if I use the caffe alexnet model, any thoughts? @youkaichao Did you train from scratch or did you use the alexnet model? If you did use the pretrained model, did you make any changes? |
@Priyam1994 I'm using the pretrained alexnet model, and it works. I have made no changes to the pretrained alexnet model. |
@youkaichao Thank you for the clarification. Did you maybe change any other parameters than what has been originally suggested? |
@Priyam1994 Nope. I run MVCNN with default settings. And I don't know how to replace the alexnet model …… |
@WeiTang114 Could you tell me if there is a specific reason the rendered images(specified in the link given by you) are of dimension 600 * 600? Can I also feed in images of other dimensions, say 300 * 300 or 400*400? Thank you |
@Priyam1994 That's just random. After input into the network the images are resized to 256 and cropped to 227, so I just wanted a large enough size so that the images aren't distorted after resized. Of course any other size is fine! |
@WeiTang114 Thank you for your and informative reply! |
@weiweisunWHU |
@youkaichao |
@youkaichao @WeiTang114 hello, after i use your code "cv2.bitwise_not(im)" after line 27, my acc is still about 2%, for example, sometimes it's "acc=1.953125", sometimes it's "acc=2.734375" and other value, and i haven't changed anything in the code, the dataset i train is ModelNet40 with write background like your airplane image. Do you know what should i do? thank you so much. |
@weiweisunWHU your method really works!i tried it successfully to more than 85% accuracy. |
@491506870 |
Hello Wei,
I have prepared the data and trained the models without changing anything. However, I found the loss converging to 3.69. Then I changed initial learning rate but obtaining same converged loss (3.69). Do you know is there any problem? By the way, could you please provide the trained weight? Thanks a lot.
The text was updated successfully, but these errors were encountered: