BERT distillation and quantization in a distributed setting using Fairscale library. The repository contains different versions of the training code on the GLUE task using the mRPC dataset with different levels of parallelism using PyTorch and FairScale's constructs:
-
Notifications
You must be signed in to change notification settings - Fork 0
rkv0id/fair-distribert
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
BERT distillation and quantization in a distributed setting using Fairscale library.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published