You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I trained using your pipeline on the whisper-large-v3 model and the word error rate is more than 100% and the value is {'eval/wer': 100.71108101244393, 'eval/normalized_wer': 100.2449409394744}
When I commented out fp16=True it worsened more to :
wer=333.30652670786424 and normalized_wer=190.30308579800385
I used common voice -11 dataset for Hindi language, here is my code and config for the trainer:
from transformers import Seq2SeqTrainingArguments
training_args = Seq2SeqTrainingArguments(
output_dir="Abhinav28/large-v3-hi-commonvoice-11-peft-trained-adapter-withfp16", # change to a repo name of your choice
per_device_train_batch_size=8,
gradient_accumulation_steps=1, # increase by 2x for every 2x decrease in batch size
learning_rate=1e-3,
warmup_steps=50,
num_train_epochs=1,
evaluation_strategy="steps",
fp16=True,
per_device_eval_batch_size=8,
generation_max_length=128,
# metric_for_best_model="wer",
logging_steps=819,
# max_steps=10, # only for testing purposes, remove this from your final run :)
# greater_is_better=False,
remove_unused_columns=False, # required as the PeftModel forward doesn't have the signature of the wrapped model's forward
label_names=["labels"], # same reason as above
)
I trained using your pipeline on the whisper-large-v3 model and the word error rate is more than 100% and the value is
{'eval/wer': 100.71108101244393, 'eval/normalized_wer': 100.2449409394744}
When I commented out fp16=True it worsened more to :
wer=333.30652670786424 and normalized_wer=190.30308579800385
I used common voice -11 dataset for Hindi language, here is my code and config for the trainer:
from transformers import Seq2SeqTrainingArguments
How do I make it have better wer?
The text was updated successfully, but these errors were encountered: