File size: 2,910 Bytes
ea31a58 7e4b855 ea31a58 7e4b855 ea31a58 7e4b855 ef678a2 7e4b855 ea31a58 15053a8 618dc77 15053a8 ea31a58 15053a8 ea31a58 ff2b0fa ea31a58 c1ead4c ea31a58 c1ead4c ea31a58 c1ead4c ea31a58 c1ead4c 618dc77 c1ead4c ea31a58 c1ead4c ea31a58 7e4b855 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 |
slurm submission log: 2024-05-29 11:13:48.705547 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --dependency=afterok:7667678 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-2566646 #SBATCH --mem=60G #SBATCH --nodelist=sphinx1 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/eval_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1,revision=main,dtype=float16,trust_remote_code=True --tasks piqa,arc_easy,xnli_en,xnli_fr,xnli_de,xnli_es,sciq,lambada --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/perf' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7667679 ############################### slurm submission log: 2024-05-30 08:40:45.658941 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --dependency=afterok:7670593 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-3481108 #SBATCH --mem=60G #SBATCH --nodelist=sphinx1 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/eval_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1,revision=main,dtype=float16,trust_remote_code=True --tasks piqa,arc_easy,xnli_en,xnli_fr,xnli_de,xnli_es,sciq,lambada --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_arc_easy_1/perf' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7670594 ############################### |