-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
0.4.0 lm-evaluation-harness #15
Comments
Yes, there are! :) stay tuned! |
Do you have any particular expectations for improvements with the upgrade to the 0.4.0+ backend? |
@LSinev Hi, |
hello guys! can I ask you, do you work on this topic, maybe you have some estimated dates? |
will give more information next week, or may be even branch for playing/testing work in progress |
new_harness_codebase — "work in progress" branch with submoduled patched (waiting for PR to be merged) lm-evaluation-harness. |
great, thank you! |
Hi @LSinev, I noticed that the tasks from the branch do not include the MERA tasks in Could you please confirm if the MERA tasks will be added to this branch, or if there is another location where they might be available? Thanks! |
This link goes to fork of lm-evaluation-harness. In this fork there is a code needed for RuTiE task, which is PRed in lm-evaluation-harness, but not yet approved and merged. new_harness_codebase is using 0.4.x code, but tasks are not in fully yaml format yet (will be, but not yet, just like, for example, SQUADv2 task in lm-evaluation-harness). MERA tasks are stored in https://github.com/ai-forever/MERA/tree/update/new_harness_codebase/benchmark_tasks as new code allows to use tasks from external directory. |
Hi!
Your benchmarks are functioning well with version 0.3.0 of lm-evaluation-harness. Are there any plans to update and support version 0.4.0?
The text was updated successfully, but these errors were encountered: