-
Notifications
You must be signed in to change notification settings - Fork 200
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Optimized inference of XGLM model on HPU #1323
Conversation
Signed-off-by: Ye, Xinyu <[email protected]>
Signed-off-by: Ye, Xinyu <[email protected]>
Signed-off-by: Ye, Xinyu <[email protected]>
@XinyuYe-Intel can you provide gaudi2 test on latest 1/17/1.18 docker RUN_SLOW=true GAUDI2_CI test and gaudi1 test result? |
perf on gaudi2 on 1.17.1 with RUN_SLOW=true is as below: For gaudi1, I don't have the machine, so I can't provide the result. |
Signed-off-by: Ye, Xinyu <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@XinyuYe-Intel could you please resolve the conflicts on this PR, looks good otherwise
Resolved conflicts. |
@XinyuYe-Intel can you rebase? |
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
What does this PR do?
Optimized inference of XGLM model on HPU.
Before submitting