-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathprophet.slurm
55 lines (41 loc) · 1.34 KB
/
prophet.slurm
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
#!/bin/bash
#############################
# les directives Slurm vont ici:
# Your job name (displayed by the queue)
#SBATCH -J prophet
# walltime (hh:mm::ss)
#SBATCH -t 02:00:00
# change working directory
# SBATCH --chdir=.
### In filenames, %j=jobid, %a=index in job array
#SBATCH -o /beegfs/tferte/std_out/%j_%a_%x.out # standard out goes to this file
#SBATCH -e /beegfs/tferte/std_err/%j_%a_%x.err # standard err goes to this file
#SBATCH --array 0-199
#SBATCH --ntasks 1
#SBATCH --cpus-per-task 5
#SBATCH --mem-per-cpu=2560
# fin des directives PBS
#############################
# useful informations to print
echo "#############################"
echo "User:" $USER
echo "Date:" `date`
echo "Host:" `hostname`
echo "Directory:" `pwd`
echo "SLURM_JOBID:" $SLURM_JOB_ID
echo "SLURM_ARRAY_JOB_ID:" $SLURM_ARRAY_JOB_ID
echo "SLURM_SUBMIT_DIR:" $SLURM_SUBMIT_DIR
echo "SLURM_JOB_NODELIST:" $SLURM_JOB_NODELIST
echo "SLURM_JOB_NAME" : $SLURM_JOB_NAME
echo "#############################"
#############################
export OPENBLAS_NUM_THREADS=1
export OMP_NUM_THREADS=1
export MKL_NUM_THREADS=1
# cd /gpfs/home/tferte910e/high_dimension_reservoir
# source /gpfs/home/tferte910e/predictcovid_api_python/testenv/bin/activate
module load build/conda/4.10
conda activate testenv
cd /home/tferte/high_dimension_reservoir
ulimit -c 0
python main_csv_evaluate.py