Data management duties have lapsed in my lab. I'm trying to identify files in our systems that have been published to SRA.
I have a hash value for every file in our system, if I can download the exact file that was upload to SRA, I can get the hash value and cross check for duplicates. However, when files are uploaded to SRA, they are transformed into SRA objects from which you get the sequence using SRA toolkits fastq-dump.
Downloading the fastq using this tool yields a file of a different size from the originally uploaded file.
Are there certain command line options for fastq-dump I can specify to regenerate the exact file that was uploaded?
Thanks.
I saw those (was using them for the size estimates mentioned). I didn't think they would be available for download via the s3 address, but I'll try it.
Thanks again.
Edit: following up
You have to use the SRA's cloud delivery service. You cannot download the original files from SRA's S3 bucket via AWS CLI.
Did you have to pay?
While technically you don't have to pay NCBI to get the data, you need to use the
cloudy delivery
as noted above. This requires you to create your own storage bucket in google cloud or amazon S3. Which mean you will end up with some cost since you have to pay for storage where the data will be delivered. This cost can vary where you are but is about 10 to 20 US cents per GB of data.