-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Target prune FLOPs ratio #4
Comments
Hi King4819, Thanks for your interest in our work! The level of pruning depends on the hyper-parameters Regards, |
@Charleshhy thanks for your reply. Is there a proper way to set hypaparameter theta when the target FLOPs is change? how do I know which value to set? |
In practice, I set these two values to let the FLOPs loss slightly higher than the cross-entropy loss at the beginning of training and find it works well. However, I didn't experiment much about the trade-off between these two losses and different settings may lead to better performance :) |
@Charleshhy Thanks for your reply. I want to ask it more explicitly. For example, I want to prune DeiT-S model at three different level: prune FLOPs ratio 30%, prune FLOPs ratio 60% and prune FLOPs ratio 90%. How do I alter hypaparameter theta for these three target FLOPs ? Or this is an unexperimented direction ? |
Different target FLOPs would lead to different FLOPs loss and we need to adjust \theta to hack its value to be slightly higher than the cross-entropy loss in practice. Note that pruning 90% FLOPs is too aggressive and I have not tried it :) |
@Charleshhy Thanks for your reply! |
Hello, I would like to ask you about the "theta": 1.5 in swin-transformer pruning, what does it mean?what do theta and 1.5 mean?In addition, the recognition accuracy is relatively low in the search process?ACC1% is only 10% in the number of search rounds about 40, is this normal?(I set theta to 0.5 in this process, and the target_flops is 2.9) |
@King4819 I made a mistake in explaining the hyper-parameters and have corrected it just now. Sorry for the confusion. |
Hi Bo102, |
Hello, excuse me, I used your configuration to search for the swin-transformer pruning architecture, I just modified the dataset to imagenet-tiny-200, and the accuracy is still very low during the search process. Is this due to the fact that I need to modify the parameters in the configuration file according to my actual data set? What is the reason for this? Also, I use the above searched files to guide pruning, and the generated model, I can't improve the accuracy of the training level, and it has been hovering around 50%. |
During searching, a low accuracy means the model searched a trivial solution and that won't have good performance. My suggestion is to set a lower |
Thank you so much, I wish you all the best, I'll give it a try |
Excellent work !!! I want to ask whether the method is able to decide different prune FLOPs ratio. For example, I want to perform different level of pruning, such as prune FLOPs 30%, 60%, 90%, respectively. Thanks !
The text was updated successfully, but these errors were encountered: