-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The precision of ResNet18 is .0? #1
Comments
Huh. Can you please share the entire log? Is this with a single GPU? |
Hey, I was using a single GPU! where can I find the entire log? |
|
Yep that looks like the log, but definitely not all of it. |
This should be the whole log info |
It seems like the accuracy after the initial compression is too low (5%). I'll try to reproduce on my end, thanks for bringing this up! |
Thanks! Waiting for your info~ |
I also have similar situation, the accuracy was maintained at 5% and has not changed! |
Are you seeing this with other models too? Or just with Resnet 18? |
I have managed to reproduce the issue that you are reporting. I apologize. Since we developed this code on machines with distributed training with horovod, we missed a bug in the dataloader. As written, the training imagenet dataloader is not shuffling the training data. You can replace
with loader = DataLoader(
dataset,
batch_size=batch_size,
num_workers=num_workers,
shuffle=(sampler is None),
sampler=sampler,
pin_memory=True
) and that should bring back training numbers that make sense. Also, please note that I don't have write access to this repo anymore, so I will push patches to my personal fork at https://github.com/una-dinosauria/permute-quantize-finetune/. I'll let you know here when that repo is patched. Again, sorry for this mistake, and thank you so much for reporting this issue. Cheers, |
Fixed by una-dinosauria#2. Cheers, |
I've also added a docker image to make it easier to reproduce our results: una-dinosauria#3 |
The command I ran is "python -m src.train_resnet --config ../config/train_resnet18.yaml", I got the accuracy is 0.0 after finetune! Any idea of what's causing it?
The text was updated successfully, but these errors were encountered: