File tree Expand file tree Collapse file tree 1 file changed +46
-0
lines changed Expand file tree Collapse file tree 1 file changed +46
-0
lines changed Original file line number Diff line number Diff line change 1+ #! /bin/bash
2+ # SBATCH -n 1
3+ # SBATCH -c 16
4+ # SBATCH --ntasks-per-node=1
5+ # SBATCH --gpus-per-task=1
6+ # SBATCH --mem=164000
7+ # SBATCH -t 23:59:59
8+ # SBATCH -A m2404
9+ # SBATCH -C gpu&hbm80g
10+ # SBATCH -q regular
11+ # SBATCH -J generate-codellama-70b-prompted
12+ # SBATCH -o generate-codellama-70b-prompted-%A.out
13+
14+ # settings
15+ MODEL=" codellama/CodeLlama-70b-hf"
16+ TEMP=0.2
17+ TOPP=0.95
18+ MAX_NEW_TKNS=1024
19+ SAMPLES_PER_PROMPT=20
20+ BATCH_SIZE=2
21+ hash=$( md5sum ../prompts/generation-prompts.json | cut -d' ' -f1)
22+ OUTPUT=" ../outputs/output_${hash: 0: 8} _${MODEL// \/ / --} _prompted_temp${TEMP} .json"
23+ CACHE=" ../outputs/cache/cache_${hash: 0: 8} _${MODEL// \/ / --} _prompted_temp${TEMP} .jsonl"
24+ echo " Writing to $OUTPUT "
25+ echo " model=$MODEL MAX_NEW_TKNS=$MAX_NEW_TKNS SAMPLES_PER_PROMPT=$SAMPLES_PER_PROMPT BATCH_SIZE=$BATCH_SIZE "
26+
27+ # setup
28+ # ml cuda/11.8.0
29+ source .env/bin/activate
30+ export HF_HOME=/pscratch/sd/d/dnicho/.cache/huggingface
31+ export OMP_NUM_THREADS=16
32+ export SLURM_CPU_BIND=" cores"
33+
34+ # generate
35+ srun python generate.py \
36+ --model $MODEL \
37+ --prompts ../prompts/generation-prompts.json \
38+ --cache $CACHE \
39+ --output $OUTPUT \
40+ --temperature $TEMP \
41+ --top_p $TOPP \
42+ --do_sample \
43+ --max_new_tokens $MAX_NEW_TKNS \
44+ --num_samples_per_prompt $SAMPLES_PER_PROMPT \
45+ --batch_size $BATCH_SIZE \
46+ --prompted
You can’t perform that action at this time.
0 commit comments