DiffBind "no lines available in input" error
Entering edit mode
igor ▴ 40
Last seen 9 months ago
United States

Sometimes my ChIP-seq samples do not get any significant peaks by MACS, so the output peaks file is empty. When I use that file in DiffBind sample sheet, I get an error:

Error in read.table(fn, blank.lines.skip = T, header = T) : 
  no lines available in input
Calls: dba ... pv.peakset -> pv.readPeaks -> pv.macs -> read.table

Is there a way to suppress that error?

I understand if all the samples have no peaks, there should definitely be an error, but sometimes you have one sample that just doesn't have strong peaks that you still want to include in the analysis.



diffbind chipseq • 1.6k views
Entering edit mode
Rory Stark ★ 4.1k
Last seen 11 days ago
CRUK, Cambridge, UK

Actually there is a workaround. If you leave the Peaks file empty (blank, no value) in the samplesheet for that sample, it should work! dba.count() will still count reads for this sample in the consensus peaks, and these reads can be used for a differential analysis.

We'll have a look at the error message so we can fix this in a future release.


Entering edit mode
Gord Brown ▴ 640
Last seen 9 months ago
United Kingdom

To the best of my knowledge there's no way to suppress that error.  I can see how it might be useful, though, if (for example) one condition completely eliminated binding.  I'll add it to the list of feature requests, but for now I can't see how to bypass it without hacking the source code.  It wouldn't surprise me if, even if you could read the file, other problems would arise downstream, since DiffBind hasn't been tested with empty peak files (that I know of).  Sorry I can't be more helpful.

Wait... just reading the manual page for "dba.peakset", it looks like you can add a peak set that is explicitly empty by setting "peaks=NA".  No idea what might happen downstream, though.


Login before adding your answer.

Traffic: 176 users visited in the last hour
Help About
Access RSS

Use of this site constitutes acceptance of our User Agreement and Privacy Policy.

Powered by the version 2.3.6