-
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add running jobs in Planetary Computer #50
Comments
Docs for kbatch: https://kbatch.readthedocs.io/en/latest/examples/index.html |
Could you point to one specific file where this was done for the Satellite date? Perhaps that is to come? |
Sure, here is a config file: And to run it, after updating the HF token, EUMETSAT API key and secret: kbatch job submit -f pc/eumetsat-0deg.yaml |
Nice thanks @jacobbieker |
We would like to run some of our jobs with public data in Planetary Computer, to help with the bandwidth issues with have on Leonardo.
Detailed Description
I've used the
kbatch
service to submit jobs to Planetary Computer quite successfully, command line args can be passed, as well as code locations. This should allow us to do whatever we want in the VM.Each VM seems to have: 64GB RAM, 150ish GB storage
We can pass in credentials, including to GCP, to write directly there if need be.
Sometimes, processing seems to fail for an unknown reason, for example it processes HILAM surface NetCDFs fine, but not the pressure ones. But overall, that should just be some debugging and then it will be good to go.
Context
It offloads processing to the cloud, freeing up bandwidth on Leonardo for batch creation or getting of data we can't make public.
Possible Implementation
I have a lot of examples in my planetary-datasets repo
The text was updated successfully, but these errors were encountered: