Large memory usage by gIMble #124
-
Hi :^) Thanks for creating this very neat program! All instructions (manual) given on the GitHub main page are clear to follow. I have been running gIMble on a cluster with Slurm, and I found that it can take extremely large memory (more than 500 G of maxRSS) when running some commands (especially window & tally). Detailed descriptions: My datasets
Dataset 2
MaxRSS captured by Slurm For Dataset 2 I was wondering if this behavior is normal and which aspect of my data caused such a large MEM consumption (I thought it would be the number of heterospecific sample pairs)? Thank you so much! Looking forward to your reply :^) Gratefully, |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Hi Meng, that is somewhat to be expected. The way things are being saved righht now is not yet "optimal" and we are thinking about improving the datastructure(s), both those kept in memory during computations and those saved to store, in future releases. In the meantime the easiest way to deal with this is making window-wise analysis on a scaffold/chromosome basis. So you make a new gimble store for each sequence by editing the genomefile to only include that sequence. Hope that helps, dom |
Beta Was this translation helpful? Give feedback.
Hi Meng,
that is somewhat to be expected. The way things are being saved righht now is not yet "optimal" and we are thinking about improving the datastructure(s), both those kept in memory during computations and those saved to store, in future releases.
In the meantime the easiest way to deal with this is making window-wise analysis on a scaffold/chromosome basis. So you make a new gimble store for each sequence by editing the genomefile to only include that sequence.
Hope that helps,
dom