File size: 2,910 Bytes
ea31a58
7e4b855
 
 
 
 
 
 
 
ea31a58
7e4b855
ea31a58
7e4b855
ef678a2
7e4b855
ea31a58
15053a8
 
 
 
618dc77
15053a8
 
 
 
 
ea31a58
15053a8
 
 
 
 
 
 
 
 
ea31a58
ff2b0fa
 
 
 
 
ea31a58
c1ead4c
 
 
 
 
 
 
 
ea31a58
c1ead4c
ea31a58
c1ead4c
 
 
ea31a58
c1ead4c
 
 
 
618dc77
c1ead4c
 
 
 
 
ea31a58
c1ead4c
 
 
 
 
 
 
 
 
ea31a58
7e4b855
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
slurm submission log: 2024-05-29 11:13:48.705547
created following sbatch script: 

###############################

#!/bin/bash

#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7667678
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-2566646
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx1
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0

# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection

# cd to working directory
cd .

# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1,revision=main,dtype=float16,trust_remote_code=True --tasks piqa,arc_easy,xnli_en,xnli_fr,xnli_de,xnli_es,sciq,lambada --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/perf'

###############################

submission to slurm complete!


###############################
slurm submission output

Submitted batch job 7667679



###############################

slurm submission log: 2024-05-30 08:40:45.658941
created following sbatch script: 

###############################

#!/bin/bash

#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7670593
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-3481108
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx1
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0

# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection

# cd to working directory
cd .

# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1,revision=main,dtype=float16,trust_remote_code=True --tasks piqa,arc_easy,xnli_en,xnli_fr,xnli_de,xnli_es,sciq,lambada --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/perf'

###############################

submission to slurm complete!


###############################
slurm submission output

Submitted batch job 7670594



###############################