I have a sam file which contain alignments of about 30 million reads. I just would like to randomly sample e.g. 1 million alignments out of my sam file.
Whats the best way to do this? Anything already available in any kind of toolkit?
Or bash scripting? Or Perl?
Version is important.
I just found out -s option is not available on version 0.1.16
Also option is cryptic:
so I had to write it as
-s -1.5 ( to get 50% of reads)
-s -1.7 ( to get 70% of reads)
1 is the seed and .5/.7 the percent of reads for sub sampling
In version 0.1.19 if you extract two samples using the same random seed you always obtain the same reads.
So if you have several samples to extract, remember to change the random seed. As @sheetalshetty13 pointed out, the random seed is the part of the number before the point. So 0.5 will extract 50% with random seed "0" and 1.5 50% of the reads with random seed "1".
Interestingly, I noticed that if you use two digits random seeds you obtain different sample sizes (using the same probability).
I'm experiencing the same problem. You said you were not using samtools.
Do you use picard DownsampleSam or some custom script to downsample a sam/bam file?
Thank you!
ADD REPLY
• link
updated 2.7 years ago by
Ram
44k
•
written 9.9 years ago by
biola
▴
20
0
Entering edit mode
In that situation I used a completely different approach. Since the data were simulated I just simulated fastq files of different sizes (I had to align all of them, but better than having wrong results...).
DownsampleSam might be an option (I know that several people use it and they are happy with their results. I remember it being slow...)
The approaches suggested in the following answers are very good. I like the shuffle very much (you can remove the header and save it somewhere else and then add it back).
###remove header and save the sam file head
sed 1,23d file.sam > file_noHeader.sam
head -n 23 file.sam > head
###randomly select one Million reads and save them (I took the one liner from here: http://www.unix.com/shell-programming-scripting/68686-random-lines-selection-form-file.html)
awk 'BEGIN {srand()} {printf "%05.0f %s \n",rand()*9999999, $0; }' file_noHeader.sam | sort -n | head - 1000000| sed 's/^[0-9]* //' > randomReads.tmp
###join the header back to have randomly sampled million reads subset of original file
cat head randomReads.tmp > randomReads.sam
###remove the tmp files
rm file_noHeader.sam randomReads.tmp
Ofcourse it can be more efficient and automated using pipes.
Also, you can save the script in a file, and replace the file name with $1, like
SAM files have a header section, as well as a body section, thus simply shuffling the file would mess up the overall structure. You could break the file into parts, as Sukhdeep suggests, and then shuf the body part.
A simple workaround would be to sort your sam file by read names see the -n flag to samtools sort, then take the first, second etc 1 million to get a random sample without replacement.
In all honesty this approach most likely has some limitations since the read naming probably correlates to flow cell location that in turn may introduce some biases. I still think it could be useful and it is really easy to do.
One more possibility; assuming you have Illumina reads, use grep to grep out just the reads from a particular tile or two, preferably tiles not on the edge. That subset would have a slightly higher quality than the whole dataset, which includes reads on the edge of the flowcell, but that's probably a good thing.
Excellent! I wish I had known this before!
I could not find the
-s
option on samtools. When I tried using, gave me an error 'invalid option'.Would you know the reason. I am using samtools-0.1.16
Version is important. I just found out -s option is not available on version 0.1.16 Also option is cryptic: so I had to write it as -s -1.5 ( to get 50% of reads) -s -1.7 ( to get 70% of reads) 1 is the seed and .5/.7 the percent of reads for sub sampling
I think I found a problem with
samtools -s
In version 0.1.19 if you extract two samples using the same random seed you always obtain the same reads.
So if you have several samples to extract, remember to change the random seed. As @sheetalshetty13 pointed out, the random seed is the part of the number before the point. So 0.5 will extract 50% with random seed "0" and 1.5 50% of the reads with random seed "1".
Interestingly, I noticed that if you use two digits random seeds you obtain different sample sizes (using the same probability).
If you run samtools view -s on a file on which samtools view -s was already used you have unpredictable results
I am currently not using
samtools view -s
.Hi Fabio,
I'm experiencing the same problem. You said you were not using samtools.
Do you use picard DownsampleSam or some custom script to downsample a sam/bam file?
Thank you!
In that situation I used a completely different approach. Since the data were simulated I just simulated fastq files of different sizes (I had to align all of them, but better than having wrong results...).
DownsampleSam
might be an option (I know that several people use it and they are happy with their results. I remember it being slow...)The approaches suggested in the following answers are very good. I like the shuffle very much (you can remove the header and save it somewhere else and then add it back).
Hi! You said 1 is the seed, and what the seed means?