-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Accuracy is ~80 after 350 epochs #4
Comments
Hi @ChesterAiGo |
Hi @vibrantabhi19 Thanks for your prompt reply! I will have a try of different optimizers as well as try vary γ during training(I think that's probably why) In addition, there was something very interesting about the accuracies..i.e. the training accuracy keeps increasing steadily (from epoch 1 to epoch 350) while the validation accuracy became stable (was not increasing but was not decreasing as well..that's weird xD) after epoch 49.. Something looks like: Epoch 1: Val: 0.1, Train: 0.1 Thanks again ! :) |
Oh, that's weird, the network cannot overfit, we are already using a dropout of 0.5. |
I can confirm, that using the original code (with the fix in #5) and removal of multi_gpu code reveals an accuracy over 81%. My best after 350 epochs using the code of this repository was 90.88%. However, it cracked 90% already in epoch 140. See accuracy (as CSV): And loss (as CSV): The learning rate decay produced this (as CSV): See also full console log. and all source code + weights here: https://aetros.com/marcj/keras:all-conv/view/refs/aetros/job/92fcd671c6814c375edd404a65edc66c00ba5aec or in the analytics tool at https://trainer.aetros.com/model/marcj/keras:all-conv/job/92fcd671c6814c375edd404a65edc66c00ba5aec (requires login first) Hyper parameter and other information here: So what I can say: I can not reproduce the stuck at 81%. @ChesterAiGo, you can fork my model at https://aetros.com/marcj/keras:all-conv and try to run it on your hardware, so we have all information to debug it. However, I'd also like to know why this code does not produce the results from the linked paper and what you need concretely to achieve 95.59% for cifar10 using all-conv. |
this is some sexy plots 90 percent accuracy |
hi vibrantabhi19 :
Thank you for sharing your code! That's very helpful for me to understand All-CNN.
In addition, I've trained it last with your model night with 350 epochs, however found its accuracy (i.e. val_acc) became stable (about 0.81) after epoch 49 and remained the same to the end
Any ideas? :) 👍
The model I used:
`
model = Sequential()
sgd = SGD(lr=0.01, decay=1e-6, momentum=0.9, nesterov=True)
model.compile(loss='categorical_crossentropy', optimizer=sgd, metrics=['accuracy'])`
The text was updated successfully, but these errors were encountered: