Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Konnector merges very few reads for high-coverage genome #489

Open
zcalamari opened this issue Dec 26, 2024 · 0 comments
Open

Konnector merges very few reads for high-coverage genome #489

zcalamari opened this issue Dec 26, 2024 · 0 comments

Comments

@zcalamari
Copy link

abyss-pe version 2.1.5 (singularity container)
Linux Distribution: Rocky Linux release 8.9 (Green Obsidian)

Hello,

I am trying to troubleshoot using Konnector to generate pseudoreads for a mammalian genome around 2.7 Gb. The genome in question is approximately 100x coverage, but Konnector merges very few reads. Here is an example output:

Bloom filter FPR: 0.672%
Connecting read pairs
Processed 920776180 read pairs
Merged (Unique path + Multiple paths): 18 (1.95e-06%)
No start/goal kmer: 96582293 (10.5%)
No path: 515094849 (55.9%)
Unique path: 18 (1.95e-06%)
Multiple paths: 0 (0%)
Too many paths: 667414 (0.0725%)
Too many branches: 0 (0%)
Too many path/path mismatches: 16476 (0.00179%)
Too many path/read mismatches: 0 (0%)
Contains cycle: 0 (0%)
Max cost exceeded: 308415130 (33.5%)
Skipped: 0 (0%)
Bloom filter FPR: 0.672%

This was generated with this command:
singularity exec /home/calamari/abyss.sif konnector -j 128 -k 95 -b 220G --fastq -o Kon_2 -v Mg_1P.fq.gz Mg_2P.fq.gz

I have tried different k-mers, bloom filter sizes, and minimum coverage settings, with no improvement. I also increased the maximum search cost (--max-cost), but the operation timed out after a week without producing a result. Konnector has worked for me with lower coverage genomes (30x coverage for a species with a similar sized genome), so I am hoping I can get it to work as well again with these higher-coverage genomes. Is there anything I can do to address the number of reads with "no path" or "max cost exceeded"?

Thanks for the help!

Best,
Zac

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant