batchtools_slurm job setting with seurat
1
0
Entering edit mode
7 weeks ago
marco.barr ▴ 150

Hi everyone, I am performing an analysis with Seurat on an object that is 2.4 GB in size. I am running my analysis on a Slurm cluster with physical limits of 64 GB of RAM and 10 cores. The HPC cluster administrator advises against using multiple node tasks but suggests utilizing only one node. Based on this, how can I make the most of these resources to avoid memory issues or interruptions? I am considering using future.batchtools with these options:

My sbatch script:

#!/bin/bash
#SBATCH --job-name=seurat_analysis
#SBATCH --output=seurat_analysis_%j.log
#SBATCH --error=seurat_analysis_%j.err
#SBATCH --time=168:00:00  
#SBATCH --ntasks=1
#SBATCH --cpus-per-task=10
#SBATCH --mem=64G

Rscript /home/scrna/script.R

My R script:

 library(future)
 library(batchtools)
 library(Seurat)

 plan(list(tweak(batchtools_slurm, resources = list(
  job.name = "seurat_analysis",
  log.file = "seurat_analysis.log",
  walltime = 168 * 60 * 60,
  memory = 64 * 1024,       
  ncpus = 10
   )),
 multiprocess
 ))

  gc() 

  Exp <- CreateSeuratObject() 
   # ......... Etc

Can you recommend anything? Thank you very much for your support.

seurat HPC Slurm • 307 views
ADD COMMENT
1
Entering edit mode
6 weeks ago
DBScan ▴ 440

future.batchtools only make sense when you want to submit SLURM directly from R. From your script it looks like you run R from a SLURM job, so I would just set the plan and that's it. So you could just do

plan("multiprocess", workers = Sys.getenv("SLURM_CPUS_PER_TASK"))

Seurat has a small section on parallelization here: https://satijalab.org/seurat/reference/prepsctfindmarkers#parallelization-with-future-1

ADD COMMENT

Login before adding your answer.

Traffic: 949 users visited in the last hour
Help About
FAQ
Access RSS
API
Stats

Use of this site constitutes acceptance of our User Agreement and Privacy Policy.

Powered by the version 2.3.6