Hi,
I made a blastdb using the makeblastdb command using 901 sequences. and I did not notice this at first but the file size for the .ntf and .ndb files was HUGE. Both of these were 300gb size each and it was taking the majority of the remaining space of my poor hard drive.
I deleted them and remade the database using -max_file_sz option set to 1GB. But it left me wondering, what is the use of these files, would having a larger version of this be benefitial in any way? was it an error in the first place that made them that large?
I hope someone can enlighten me on these questions.
Thanks, Julian
how long were the sequences?
At first glance, I can't see any reason for the databases to be that size. You can download the entirety of the
nr
database for less than about half that if memory serves (pun intended).Strange, must be something wrong with my blast installation I guess.
Hi Julian.dekker, I am having the same issues, in my case makeblastdb is creating huge 300GB pdb and ptf files from a small fasta file containing 425 sequences... However, setting a limit for the file size with -max_file_sz does not work in my case... so i was wondering whether you got to the root of the problem or if reinstallation of blast+ helped?
Kind regards, Joscha