blob: 86da315fc7886ae28b2cd9e729bbe6e179d73d69 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
|
#!/bin/sh
# Split a large indexing job into many small tasks and submit using SLURM
# ./turbo-index my-files.lst label my.geom /location/for/streams
# Copyright © 2016-2020 Deutsches Elektronen-Synchrotron DESY,
# a research centre of the Helmholtz Association.
#
# Authors:
# 2016 Steve Aplin <steve.aplin@desy.de>
# 2016-2017 Thomas White <taw@physics.org>
SPLIT=1000 # Size of job chunks
MAIL=you@example.org # Email address for SLURM notifications
INPUT=$1
RUN=$2
GEOM=$3
STREAMDIR=$4
# Set up environment here if necessary
#source /path/to/crystfel/setup.sh
# Generate event list from file above
list_events -i $INPUT -g $GEOM -o events-${RUN}.lst
if [ $? != 0 ]; then
echo "list_events failed"
exit 1
fi
# If you are using single-event files instead of multi-event ("CXI") ones,
# comment out the above lines and uncomment the following one:
#cp $INPUT events-${RUN}.lst
# Count total number of events
wc -l events-${RUN}.lst
# Split the events up, will create files with $SPLIT lines
split -a 3 -d -l $SPLIT events-${RUN}.lst split-events-${RUN}.lst
# Clean up
rm -f events-${RUN}.lst
# Loop over the event list files, and submit a batch job for each of them
for FILE in split-events-${RUN}.lst*; do
# Stream file is the output of crystfel
STREAM=`echo $FILE | sed -e "s/split-events-${RUN}.lst/${RUN}.stream/"`
# Job name
NAME=`echo $FILE | sed -e "s/split-events-${RUN}.lst/${RUN}-/"`
# Job number
NUMBER=${NAME##$RUN-}
POS=`expr $NUMBER \* $SPLIT + 1`
echo "$NAME (serial start $POS): $FILE ---> $STREAM"
SLURMFILE="${NAME}.sh"
echo "#!/bin/sh" > $SLURMFILE
echo >> $SLURMFILE
echo "#SBATCH --partition=mypartition" >> $SLURMFILE # Set your partition here
echo "#SBATCH --time=01:00:00" >> $SLURMFILE
echo "#SBATCH --nodes=1" >> $SLURMFILE
# It may be polite to set the priority very low to allow other jobs through:
#echo "#SBATCH --nice=100" >> $SLURMFILE
echo >> $SLURMFILE
echo "#SBATCH --job-name $NAME" >> $SLURMFILE
echo "#SBATCH --output $NAME-%N-%j.out" >> $SLURMFILE
echo "#SBATCH --error $NAME-%N-%j.err" >> $SLURMFILE
echo "#SBATCH --mail-type END" >> $SLURMFILE
echo "#SBATCH --mail-user $MAIL" >> $SLURMFILE
echo >> $SLURMFILE
echo "#source /path/to/crystfel/setup.sh" >> $SLURMFILE # Set up environment here (again) if necessary
echo >> $SLURMFILE
command="indexamajig -i $FILE -o $STREAMDIR/$STREAM --serial-start=$POS"
command="$command -j \`nproc\` -g $GEOM"
#command="$command --peaks=zaef" # Indexing parameters here
echo $command >> $SLURMFILE
sbatch $SLURMFILE
done
|