-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
mae_visualize models vs mae_pretrain_full models #12
Comments
As noted in the issue where you find this checkpoint (#8),
If your goal is to reconstruct a good-looking image, use unnormalized pixels. If your goal is to finetune for a downstream recognition task, use normalized pixels. |
That makes total sense. I missed the |
Hello,
thank you for the great work and the great repo. I was playing with different pre-trained models for visualization. When I use a
mae_visualize_vit_base.pth
I get the reconstruction results as in the demo and the paper such as below:However when I use the
mae_pretrain_vit_base_full.pth
checkpoint the results are as below:mask_ratio=0.75
for both results.So here are my questions:
visualize
andfull
checkpoints and why the results look worse withfull
checkpoints?I would appreciate it if you could help me with these questions.
The text was updated successfully, but these errors were encountered: