You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When using CustomTask to evaluate the dataset, I found that whether I update the revision and main_score of the dataset, the result will not be modified synchronously.
So you are suggestion that we overwrite results if the revision does not match. I would be perfectly fine with that. One could be even more conservative and overwrite if the version does not match (would probably be too much), however, maybe we should overwrite results is the mteb version is below a threshold (e.g. v1.12.0, where the new results were defined)
When using CustomTask to evaluate the dataset, I found that whether I update the revision and main_score of the dataset, the result will not be modified synchronously.
The revision of "result" is still v1.
The result still remains unchanged.
I understand that adding
overwrite_results=True
can be used to forcibly update the results, but this is more like a hacky way.I observe that when the revision of the Model is modified, a re-evaluation process is triggered, which is as expected.
Additionally, when a large number of CustomTasks are defined,
overwrite_results=True
is not a good solution.The text was updated successfully, but these errors were encountered: