gpt_big_code: make flash attention impl quantization friendly #5117
fast_tests.yml
on: pull_request
Run tests for optimum.habana.transformers
5m 28s
Run tests for optimum.habana.diffusers
25m 4s
Annotations
1 error and 4 warnings