diff --git a/finetuning/generalists/training/electron_microscopy/train_em_generalist.py b/finetuning/generalists/training/electron_microscopy/train_em_generalist.py index 89e7d03f..adb87568 100644 --- a/finetuning/generalists/training/electron_microscopy/train_em_generalist.py +++ b/finetuning/generalists/training/electron_microscopy/train_em_generalist.py @@ -13,7 +13,7 @@ def finetune_em_generalist(args): """Example code for finetuning SAM on multiple electron microscopy datasets""" # override this (below) if you have some more complex set-up and need to specify the exact gpu - device = "cuda" if torch.cuda.is_available() else "gpu" + device = "cuda" if torch.cuda.is_available() else "cpu" # training settings model_type = args.model_type