We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Using together ai as in the notebook. Not using GPUs.
I'm running the evals benchmark notebook provided. https://github.com/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb
Its throwing an error: NameError: name 'UserMessage' is not defined.
In the below code, eval_rows is the eval dataset which contains 3 fields, ['chat_completion_input', 'input_query', 'expected_answer']
Its clear where/how to define the 'UserMessage' in inference call but not with evals.
response = client.eval.evaluate_rows( task_id="meta-reference::mmmu", input_rows=eval_rows, scoring_functions=["basic::regex_parser_multiple_choice_answer"], task_config={ "type": "benchmark", "eval_candidate": { "type": "model", "model": "meta-llama/Llama-3.2-90B-Vision-Instruct", "sampling_params": { "temperature": 0.0, "max_tokens": 4096, "top_p": 0.9, "repeat_penalty": 1.0, }, "system_message": system_message } } )
File ~/anaconda3/envs/py99/lib/python3.10/site-packages/llama_stack/providers/inline/eval/meta_reference/eval.py:200, in (.0) 196 chat_completion_input_str = str( 197 x[ColumnName.chat_completion_input.value] 198 ) 199 input_messages = eval(chat_completion_input_str) --> 200 input_messages = [UserMessage(**x) for x in input_messages] 201 messages = [] 202 if candidate.system_message:
NameError: name 'UserMessage' is not defined
I'm running the code as it is from notebook and documentation. It is supposed to produce the evals output.
The text was updated successfully, but these errors were encountered:
No branches or pull requests
System Info
Using together ai as in the notebook.
Not using GPUs.
Information
🐛 Describe the bug
I'm running the evals benchmark notebook provided.
https://github.com/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb
Its throwing an error: NameError: name 'UserMessage' is not defined.
In the below code, eval_rows is the eval dataset which contains 3 fields, ['chat_completion_input', 'input_query', 'expected_answer']
Its clear where/how to define the 'UserMessage' in inference call but not with evals.
response = client.eval.evaluate_rows(
task_id="meta-reference::mmmu",
input_rows=eval_rows,
scoring_functions=["basic::regex_parser_multiple_choice_answer"],
task_config={
"type": "benchmark",
"eval_candidate": {
"type": "model",
"model": "meta-llama/Llama-3.2-90B-Vision-Instruct",
"sampling_params": {
"temperature": 0.0,
"max_tokens": 4096,
"top_p": 0.9,
"repeat_penalty": 1.0,
},
"system_message": system_message
}
}
)
Error logs
File ~/anaconda3/envs/py99/lib/python3.10/site-packages/llama_stack/providers/inline/eval/meta_reference/eval.py:200, in (.0)
196 chat_completion_input_str = str(
197 x[ColumnName.chat_completion_input.value]
198 )
199 input_messages = eval(chat_completion_input_str)
--> 200 input_messages = [UserMessage(**x) for x in input_messages]
201 messages = []
202 if candidate.system_message:
NameError: name 'UserMessage' is not defined
Expected behavior
I'm running the code as it is from notebook and documentation. It is supposed to produce the evals output.
The text was updated successfully, but these errors were encountered: