Hello,
I have been trying to run this line of code for the longest time:
STAR --runThreadN 20 --runMode genomeGenerate --genomeDir genomeDir/ --genomeFastaFiles Homo_sapiens.GRCh38.dna.toplevel.fa --sjdbGTFfile Homo_sapiens.GRCh38.110.chr.gtf
I first tried running it on my home terminal but then realized that that it would take several days so I switched to AWS. I have created an instance for EC2 and then a S2 bucket and then connected it all through IAM. But whenever I try and run it I keep on getting:
Oct 05 03:02:26 ..... started STAR run
Oct 05 03:02:26 ... starting to generate Genome files
terminate called after throwing an instance of 'std::bad_alloc'
what(): std::bad_alloc
Aborted (core dumped)
I then switched the code to
STAR --runThreadN 20 --runMode genomeGenerate --genomeDir genomeDir/ --genomeFastaFiles Homo_sapiens.GRCh38.dna.toplevel.fa --sjdbGTFfile Homo_sapiens.GRCh38.110.chr.gtf --limitGenomeGenerateRAM 30G --genomeSAsparseD 2
But it still doesn't work. My storage is a max of 30G. Are the files too big and should I be working with smaller ones to create the STAR genome index or is there a better way of dealing with this issue of running a genome index.
Thanks
Do you know the difference between memory and disk space? Please find out what your maximum allowed memory is.
I didn't mean storage but the instance that I am running the Linux on has a maximum of 30G.