-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
A question about the split of dataset #77
Comments
Hi, the test set should be reserved before the training, and the training process further splits the training data into a training set and a validation set. After the training is done you can then evaluate with the best model on the test data. |
I incorporated the training/testing into the same pipeline in the latest commit. You can set --eval_scheme=5-fold-cv-standalone-test which will perform a train/valid/test like this:
You can also simply run a 5-fold cv --eval_scheme=5-fold-cv There were some issues with the testing script when loading pretrained weights (i.e., sometimes the weights are not fully loaded or there are missing weights, setting strict=False can reveal the problems.). The purpose of the testing script is to generate the heatmap, you should now read the performance directly from the training script. I will fix the issues in a couple of days. |
Hi!
In the paper, dataset is split into two parts, training slides and testing slides. In the code, testing dataset is evaluated after every epoch of training. It seems that testing dataset is used as validation dataset. Do the classification results come from results of the last evaluation? Which do the classification results come from, results of the last evaluation or the best evaluation?
The text was updated successfully, but these errors were encountered: