Skip to content
This repository has been archived by the owner on Nov 12, 2024. It is now read-only.

Distilling failed #7

Open
Yanwanxi opened this issue Sep 27, 2023 · 1 comment
Open

Distilling failed #7

Yanwanxi opened this issue Sep 27, 2023 · 1 comment

Comments

@Yanwanxi
Copy link

Hello, I try to run and train
python run.py --from_pretrained google/t5-v1_1-base --dataset cqa --model_type task_prefix --label_type llm --llm palm --alpha 0.5 --batch_size 64
The google/t5-v1_1-base model was downloaded from Hugging Face, but the tokenizer was having problems.
7633a7b7-bd47-4528-b6cd-c5d2d01c695f

@rohnson1999
Copy link

you can manually download files on Hugging Face and set offline mode to run training :)
python run.py --from_pretrained '/root/.cache/huggingface/hub/models--google--t5-v1_1-base' --dataset cqa --model_type task_prefix --label_type gt --llm palm --alpha 0.5 --batch_size 64

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants