Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

pretrained weights are probably incorrect #7

Open
BestSonny opened this issue May 22, 2022 · 7 comments
Open

pretrained weights are probably incorrect #7

BestSonny opened this issue May 22, 2022 · 7 comments

Comments

@BestSonny
Copy link

The pretrained weights seem to be wrong.

For example, the vit_base has a dimension of 1024.

Could you upload the correct version? Thanks

@brewormly
Copy link

I have also some issues regarding the pre-trained checkpoints. The checkpoints only include the keys "target_encoder" and "prototypes". If I want to load the checkpoint via the training script, I get errors because the keys "epoch" and "encoder" are missing.

@MidoAssran
Copy link
Contributor

Hi @BestSonny, There are 1024 prototypes used in the loss, but I just checked the ViT-B/16 and ViT-B/4 pre-trained weights, and they both have the correct output dimension of 768. Please let me know if you would like some more clarification or help loading the models!

@MidoAssran
Copy link
Contributor

Hi @brewormly, yes the current checkpoints only include the "target_encoder" since those are the network used at the end of pre-training to obtain the results in the paper, but I would be happy to release the full checkpoints as well in case you find these useful! Will ping you once these are online!

@sayakpaul
Copy link

@MidoAssran possible to release the ImageNet-1k specific checkpoints (fine-tuned and / or linear-eval'd)?

By "linear-eval'd" I mean keeping the target encoder frozen and just training a linear layer on top of it. So, essentially, the target encoder params (which are already released) and the linear layer params.

@sayakpaul
Copy link

sayakpaul commented Aug 25, 2022

Also, the target_encoder key in the released weights -- seems like it contains two things - the actual encoder plus the projection head (module.fc* params). Is the projection head needed for downstream tasks?

@MidoAssran

@CemKaratastan
Copy link

I have also some issues regarding the pre-trained checkpoints. The checkpoints only include the keys "target_encoder" and "prototypes". If I want to load the checkpoint via the training script, I get errors because the keys "epoch" and "encoder" are missing.

I have the same issue !

@ludles
Copy link

ludles commented Jul 17, 2023

Hi @brewormly, yes the current checkpoints only include the "target_encoder" since those are the network used at the end of pre-training to obtain the results in the paper, but I would be happy to release the full checkpoints as well in case you find these useful! Will ping you once these are online!

Sorry for the late reply after one year. I wonder if there is still a plan to release the full checkpoints? I think they will be very helpful in continuing the training for other tasks.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants