Skip to content

gpt_big_code: make flash attention impl quantization friendly (#1282) #5127

gpt_big_code: make flash attention impl quantization friendly (#1282)

gpt_big_code: make flash attention impl quantization friendly (#1282) #5127

Triggered via push September 25, 2024 15:59
Status Success
Total duration 28m 57s
Artifacts

fast_tests.yml

on: push
Run tests for optimum.habana.transformers
3m 35s
Run tests for optimum.habana.transformers
Run tests for optimum.habana.diffusers
25m 0s
Run tests for optimum.habana.diffusers
Fit to window
Zoom out
Zoom in

Annotations

4 warnings
Run tests for optimum.habana.transformers
The following actions uses node12 which is deprecated and will be forced to run on node16: actions/checkout@v2. For more info: https://github.blog/changelog/2023-06-13-github-actions-all-actions-will-run-on-node16-instead-of-node12-by-default/
Run tests for optimum.habana.transformers
The following actions use a deprecated Node.js version and will be forced to run on node20: actions/checkout@v2. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/
Run tests for optimum.habana.diffusers
The following actions uses node12 which is deprecated and will be forced to run on node16: actions/checkout@v2. For more info: https://github.blog/changelog/2023-06-13-github-actions-all-actions-will-run-on-node16-instead-of-node12-by-default/
Run tests for optimum.habana.diffusers
The following actions use a deprecated Node.js version and will be forced to run on node20: actions/checkout@v2. For more info: https://github.blog/changelog/2024-03-07-github-actions-all-actions-will-run-on-node20-instead-of-node16-by-default/