-
Notifications
You must be signed in to change notification settings - Fork 40
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training vs validation accuracy #28
Comments
Hi @AnabetsyR, thanks for the questions 😃 There is a degree to which the training data is harder to classify (since the images are all mixed), but also the accuracy metric is different during training. That's also why it says It's the same as regular accuracy during validation / testing, but a bit different for training (since there are two partially correct classes during training). Hope that helps! |
@ethanwharris Thank you so much for your great response. |
Hi there! Thanks for sharing your work! The paper is very impressive!
I ran the cifar_experiment.sh with cifar data, resnet model, and fmix. However, I'm a little confused about two things:
1- Training accuracy is much lower than validation accuracy. Is this an artifact of using the masks and therefore making the learning phase much harder? Am I missing something? I confess that I'm not 100% sure I understand all the moving parts.
2- When it runs, it says running_mixup_acc etc. I had selected fmix so it's confusing to see mixup here. Can you help me understand this?
Thank you in advance!
The text was updated successfully, but these errors were encountered: