-
Notifications
You must be signed in to change notification settings - Fork 38
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem with dRep dereplication with large contigs list #247
Comments
Hi @yeoeunyun , Sorry to hear this happened. When crashing at that step, it almost certainly crashed because it ran out of RAM. If you run the same command again, with the Unfortunately, the only real solution here is to either make more RAM available or to decrease the number of genomes. I'll also just say that dRep uses genomes as inputs, not contigs (unless of course each contig is one genome). Best, |
Dear @MrOlm Thank you for your kind response! As you suggested, I'll try rerunning it with -d option. I have one additional question: would lowering the I'll explore ways to reduce the number of contigs, but actually using contigs as input is needed for my research purpose. Thanks a lot. |
Hi @yeoeunyun , Great. Paradoxically, increasing the --primary_chunksize option (to around 10,000 or 20,000?) will help the process complete. Best of luck, |
Hi @MrOlm Thank you so much. I’ll give it a try following your suggestions. Have a nice day :) |
Hi, Matt. Thanks for developing such a useful tool.
I'm trying to dereplicate a dataset using drep with 132,642 contigs.
Based on the manual and your comments in Github, I understand that for datasets with more than 5,000 contigs, it is recommended to use a contigs list text file with paths and the options
--multiround_primary_clustering
and--primary_chunksize 3000
So, here's my command I used:
However, the process was not completed successfully. It terminated with the message just "Killed" and did not generate an error log.
Below is the log content saved in the output file:
How could I solve this problem?
Thank you!
The text was updated successfully, but these errors were encountered: