Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Massive suspend data clogs up the diff_suspend_data task #294

Open
christianp opened this issue Nov 7, 2023 · 1 comment
Open

Massive suspend data clogs up the diff_suspend_data task #294

christianp opened this issue Nov 7, 2023 · 1 comment

Comments

@christianp
Copy link
Member

We had an attempt whose suspend data was over 20MB long because they wrote python code that printed like 5 million lines, and the output is saved in the pre-submit cache.

When a huey worker tried to diff the suspend data elements, it took so long that the database lock timed out. The attempt was still not diffed, so the next time the diff job came up a worker tried again. This ended up locking up all the workers.

We could try compressing very large suspend data instead of diffing it. Somewhere around 500KB might be a cut-off for compressing instead of diffing. I don't think it's feasible to diff two compressed values.

@christianp
Copy link
Member Author

We should also consider making the programming extension truncate output after a certain point - maybe 1MB.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant