-
Notifications
You must be signed in to change notification settings - Fork 372
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
learning rate decay is depending on batch size (was: Training never achieves results of original caffe models) #39
Comments
The questions is did this project are able to achieve results of original training at all? |
May be I've found problem:
Note: Actually we achieve small enough learning rate after generation 100. Original code has 25 generations, each twice size as ours(meta.write_number: 121000), btw why? Did they had more images or made more agmentation? |
And all of this affected by batch size, I have batch size = 20, so probably there is my problems in training |
Probably I need help here - should be learning rate changed with batch size? |
Update: I've probably found the problem, see last comments
I tried to train model with original C++ augmentation (rmpe_server) and my own python implementation (py_rmpe_server) it never train correctly.
To prove my point this is demo.py output with weights.best.h5 converted from caffe

Note it is perfect joint match and no additional unconnected points.
The text was updated successfully, but these errors were encountered: