-
Notifications
You must be signed in to change notification settings - Fork 53
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Request: increase memory allocation for microsatbed #831
Comments
Deeptools use a wide variety of memory allocations per individual tool, but I have increased it to 14GB on .org for now. |
Thanks @natefoo - when will the updated RAM allocation become activated? Here's a shared history with the failed job at the top if that helps. |
It is, the job had 14GB to work with.
Forgive my ignorance, but something as simple as creating bigwig files from
a sorted bed files should not be using more than a couple MB of memory.
Could you use something like
https://github.com/deeptools/pyBigWig?tab=readme-ov-file#adding-entries-to-a-bigwig-file
?
I'd also be relatively surprised if deeptools didn't use pyBigWig since
Devon is the author of both.
…On Thu, 12 Sept 2024 at 02:03, Ross Lazarus ***@***.***> wrote:
Thanks @natefoo <https://github.com/natefoo> - when will the updated RAM
allocation become activated?
Reran the same job and got exactly same OOM failure - info page has no
details on how much RAM was allocated so cannot tell if the change has been
activated. Here's a shared history with the failed job
<https://usegalaxy.org/u/fubar/h/dimer-density-and-hap-coverage-aug17-hg002>
at the top if that helps.
—
Reply to this email directly, view it on GitHub
<#831 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABT5LJOUGLGZRLSYNCZEAJLZWDK4NAVCNFSM6AAAAABN7OT3V6VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGNBUHE4TQNBTGA>
.
You are receiving this because you are subscribed to this thread.Message
ID: ***@***.***>
|
@mvdbeek: Thanks, and yes, there's a little more to it than that - all the short tandem repeats in 3-4GB of fasta need to be found first. Birds seem to have a lot of them. Will take a look, but it works without any problem as is on EU, with the same data, with whatever set up they have there - so at least we know it can be made to work. Used to be able to see maximum ram use on the information page, but no longer unfortunately. |
@mvdbeek: Configured not to make a bigwig, the most recent failure in that history also went OOM so making the bigwig is not the source of the OOM problem - that just ran pytrf. |
tl;dr Looks like EU's 8GB is highly elastic and the OOM killer is sleepy. The pytrf STR finder package uses 32GB or more to generate these VGP dimer tracks. Replacing the bigwig generator code as suggested will probably not help. @natefoo @jennaj @mvdbeek: Just ran the failing .org job command line WITHOUT producing a bigwig for the TC dimers for a 2.9GB fasta, and watched top - it rose steadily and hit 31.5GB toward the end of the run.
The resulting bed is 2.5GB which is why the tool offers bigwig output - the beds for dimer density are far too busy - but bigwigs are ~200MB so they load faster. psrecord gives a lower estimate for RAM useage - not sure why the difference from top - this with 3 second samples. Reran the CL to make a bigwig instead of a bed and used 20 second samples so coarser but seemed to peak at slightly less RAM - so bigwig is not the RAM problem. |
If the issue is not bigwig I don't see a problem with allocating more memory. Thanks for checking @fubar2 ! |
I'll give it 58GB, Runtime is pretty short anyway, but if you feel like it, https://github.com/fubar2/microsatbed/blob/main/find_str.py#L51 seems like it could be parallelized effecively, and you could write out bed entries to disk instead of collecting them in-memory in a list. |
@mvdbeek: That worked for that job - thanks! Can make it go OOM with a small adjustment to the parameters. Will retest those other parameters once the new version is available with less list comprehension bloat, but let's see how it goes in VGP workflows once the other tools are working. |
Is this one still good at 58 GB? |
All good today - need to check the figures to see how much they really used... |
http://toolshed.g2.bx.psu.edu/repos/iuc/microsatbed/microsatbed/1.3.2+galaxy1
Ross @fubar2 suggests increasing to be the same as Deeptools since it also makes use of ucsc-bedgraphtobigwig.
The text was updated successfully, but these errors were encountered: