-
Notifications
You must be signed in to change notification settings - Fork 215
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
missing tmp file during bamCoverage #546
Comments
How much space is free in |
I'm not sure this is the correct way of checking:
This is before and at the beginning of a run of the After some time, the available RAM is almost entirely used, no swapping observed, I spot the following lines in the output of
There are also 2 other recent files in
The bam file is 1.5G. |
To be noted: I see a I now see more
|
Update: About an hour after starting,
The only recent files left in
The I couldn't monitor precisely what happened during the last minute, but the last time I checked the recent files in
The |
I looked at the code where the error occurs:
I'm able to "manually" create the
And I now have |
I tried to edit the code as follows:
The following error shows that the sort command fails:
I should have taken the opportunity to insert more debugging information... |
With more debugging info:
The file |
Side question: What's the point of this piece of code? What I see is that the Isn't it possible to achieve the same sort within python? |
Possibly. This is pretty old code and predates me, but my understanding is that the goal was to produce consistent sorting across machines with different locale settings. That's possible with forcing |
Out of curiousity, what is the |
According to https://stackoverflow.com/questions/8776807/is-pythons-sort-function-the-same-as-linuxs-sort-with-lc-all-c, the correct sorting behaviour can be enforced the following way:
And then use |
I modified the code as follows in order to report the exit code and try using python sort:
Here is the output:
|
What does |
I tried to use the |
I had to do it differently because
I think the comparison function can be simplified as follows:
But this looks a bit "hackish". Probably better:
And use this directly as |
When I run the code using
|
The sorting issue appeared when the bigwig files were generated using a
system call, as bedGraphToBigWig required an specific sorting.
I wonder if this is now redundant.
Here are the messages that reported the problem:
https://groups.google.com/forum/#!searchin/deeptools/LC_ALL%7Csort:relevance/deeptools/BUBRjFV7hKc/pbW82duSFQAJ
…On Mon, Jul 3, 2017 at 4:46 PM, blaiseli ***@***.***> wrote:
When I run the code using sorted as described above (with the first
version of str_int_cmp), I get the following:
command "LC_ALL=C sort -k1,1 -k2,2n /tmp/tmpilsh0syc > /tmp/tmpilsh0syc.sorted" returned with -1 exit code
Trying python sort
I 15072434
II 15279421
III 13783801
IV 17493829
MtDNA 13794
V 20924180
X 17718942[bwClose] There was an error while finishing writing a bigWig file! The output is likely truncated.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
<#546 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AEu_1ebAOn4pOc4yL6L4KLxgGb0CEp36ks5sKP7fgaJpZM4OKhhC>
.
--
Fidel Ramirez
|
I just noticed this in my dmesg output:
Could be related to the |
I further edited
And it appears that this sort is actually also failing:
I don't know what a -1 error code is supposed to mean in this context. I suspect it is not the direct error code returned by I modified the code to have better diagnostics:
And I obtain the following result:
This is strange considering that I experience no swapping symptoms, so the RAM is not saturated. |
can you try
|
you can also try to set a fixed tmpdir in the deeptools code for sort, eg. |
Here are the output of
and
I'm suspecting the issue may be related to me having set I set back |
Some interesting reading that reinforces my suspicion about
I'm currently running the thing under It is quite surprizing that it takes so much time. Maybe this is valgrind massively slowing things [1]. There are no signs of swapping. [1](It does: "Massif runs programs about 20x slower than normal.") |
No crashes this time. The memory usage graph obtained using
So it actually never exceeds the available memory (around 10 GB), and the problem was likely this overcommit restriction I had set. I'll rerun the program without valgrind and back with That said, I believe it would be useful to use |
Thanks for the update. Don't worry about making a PR, I plan to just rewrite that bit of code completely for 2.6.0 (don't close this issue, I'm using it as the reminder to do that!). |
I've implemented the bigWig creation changes in the |
I'm trying to obtain bigwig files for some RNA-seq data, and it takes a lot of time and ends up crashing on a
FileNotFoundError
for a file in/tmp
I noticed that the process used quite a lot of RAM: it seemed to be using almost all what was available (around 10 GB), but I experienced no swap symptoms. I did not observe lot of IO going on with
iotop
.Is it possible that my workstation has not enough RAM for this and ends up loosing files in
/tmp
. Is it something that can happen?I updated to deeptools-2.5.1, and the error is still there.
(
${scale}
is a number close to 1, if that matters)The text was updated successfully, but these errors were encountered: