diff --git a/docs/source/usage_guides/accelerate_training.mdx b/docs/source/usage_guides/accelerate_training.mdx index 2eb84da5ad..b3a3934dd6 100644 --- a/docs/source/usage_guides/accelerate_training.mdx +++ b/docs/source/usage_guides/accelerate_training.mdx @@ -57,7 +57,7 @@ To not take them into account in the computation of the throughput at the end of ## Mixed-Precision Training Mixed-precision training enables to compute some operations using lighter data types to accelerate training. -Optimum Habana enables mixed precision training in a similar fasion as 🤗 Transofrmers: +Optimum Habana enables mixed precision training in a similar fashion as 🤗 Transformers: - argument `--bf16` enables usage of PyTorch autocast - argument `--half_precision_backend [hpu_amp, cpu_amp]` is used to specify a device on which mixed precision operations should be performed