Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

a small model such as Mobilenet v2 for pre-training #25

Open
mmmz28 opened this issue Apr 13, 2023 · 3 comments
Open

a small model such as Mobilenet v2 for pre-training #25

mmmz28 opened this issue Apr 13, 2023 · 3 comments

Comments

@mmmz28
Copy link

mmmz28 commented Apr 13, 2023

Thank you for your excellent work. Replacing the transformer with CNN does make deployment more friendly. Furthermore, I'm wondering if using a smaller model such as Mobilenet v2 for pre-training and then fine-tuning downstream would be effective?

@keyu-tian
Copy link
Owner

keyu-tian commented Apr 19, 2023

Thank you and we agree that this could be of general interest and value. We will consider running SparK on mobilenet recently (perhaps v2 and v3), or you can try it out too. (see tutorial at https://github.com/keyu-tian/SparK/tree/main/pretrain#tutorial-for-pretraining-your-own-cnn-model).

@xylcbd
Copy link

xylcbd commented May 9, 2023

@keyu-tian Can I use swinv2-base as the backbone for pre-training?

@keyu-tian
Copy link
Owner

keyu-tian commented May 12, 2023

@xylcbd sorry but SparK is not suitable for this. Our SparK can pretrain any CNN model but swinv2 is a transformer. Maybe you can use MAE or SimMIM to pretrain swin transformer.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants