Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

run.parent.log fails randomly and silently and breaks Evaluate step in pipeline #394

Open
ciprianjichici opened this issue Feb 13, 2022 · 0 comments

Comments

@ciprianjichici
Copy link

ciprianjichici commented Feb 13, 2022

With the latest releases of AML SDK, we find line 157 in diabetes_regression/training/train_aml.py failing randomly and silently. This breaks line 122 in diabetes_regression/evaluate/evaluate_model.py. While the fix in evaluate_model.py solves the problem of the missing metric in the parent run, it does not address the root cause of the problem, which is the random failure of run.parent.log.

Could an OutputFileDatasetConfig be used to send a model_metrics.json file containing the newly trained model metrics between the train and evaluate steps of the pipeline?

@microsoft microsoft deleted a comment from arch-exe Feb 22, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant