Tristan commited on
Commit
d740c4e
1 Parent(s): 3daa756

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "rotary_pct": 0.25,
24
  "tie_word_embeddings": false,
25
  "torch_dtype": "float32",
26
- "transformers_version": "4.39.3",
27
  "use_cache": true,
28
  "use_parallel_residual": true,
29
  "vocab_size": 50304
 
23
  "rotary_pct": 0.25,
24
  "tie_word_embeddings": false,
25
  "torch_dtype": "float32",
26
+ "transformers_version": "4.40.2",
27
  "use_cache": true,
28
  "use_parallel_residual": true,
29
  "vocab_size": 50304
eval_job_output.txt CHANGED
@@ -1,4 +1,4 @@
1
- slurm submission log: 2024-05-13 22:25:18.110347
2
  created following sbatch script:
3
 
4
  ###############################
@@ -7,24 +7,24 @@ created following sbatch script:
7
 
8
  #SBATCH --account=nlp
9
  #SBATCH --cpus-per-task=16
10
- #SBATCH --dependency=afterok:7608410
11
  #SBATCH --gres=gpu:1
12
- #SBATCH --job-name=tthrush-job-2687315
13
  #SBATCH --mem=60G
14
  #SBATCH --nodelist=sphinx2
15
  #SBATCH --open-mode=append
16
- #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/llms_basic/pythia-70m_arc_easy/eval_job_output.txt
17
  #SBATCH --partition=sphinx
18
  #SBATCH --time=14-0
19
 
20
  # activate your desired anaconda environment
21
- . /nlp/scr/tthrush/miniconda3/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
22
 
23
  # cd to working directory
24
  cd .
25
 
26
  # launch commands
27
- srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/llms_basic/pythia-70m_arc_easy,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/llms_basic/pythia-70m_arc_easy/perf'
28
 
29
  ###############################
30
 
@@ -34,7 +34,7 @@ submission to slurm complete!
34
  ###############################
35
  slurm submission output
36
 
37
- Submitted batch job 7608411
38
 
39
 
40
 
 
1
+ slurm submission log: 2024-05-17 11:42:21.569484
2
  created following sbatch script:
3
 
4
  ###############################
 
7
 
8
  #SBATCH --account=nlp
9
  #SBATCH --cpus-per-task=16
10
+ #SBATCH --dependency=afterok:7625435
11
  #SBATCH --gres=gpu:1
12
+ #SBATCH --job-name=tthrush-job-2029793
13
  #SBATCH --mem=60G
14
  #SBATCH --nodelist=sphinx2
15
  #SBATCH --open-mode=append
16
+ #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/llms_more_data/pythia-70m_arc_easy/eval_job_output.txt
17
  #SBATCH --partition=sphinx
18
  #SBATCH --time=14-0
19
 
20
  # activate your desired anaconda environment
21
+ . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
22
 
23
  # cd to working directory
24
  cd .
25
 
26
  # launch commands
27
+ srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/llms_more_data/pythia-70m_arc_easy,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/llms_more_data/pythia-70m_arc_easy/perf'
28
 
29
  ###############################
30
 
 
34
  ###############################
35
  slurm submission output
36
 
37
+ Submitted batch job 7625436
38
 
39
 
40
 
logs/events.out.tfevents.1716022652.sphinx2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a83d861b4fdb307118c91ce356ddd9c77af3b50f8d83c9a4145218c18dfa7ff4
3
+ size 95392
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f139b16d6b9703f2b3b5a7017d4472598e4b772444edbf3b07201a9017ba2758
3
  size 281715176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aac7930c164962b872a12a11bbc9fe9417fb0a96dcef35b9eb657b763e0d6afa
3
  size 281715176
train_job_output.txt CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e04c8621a1c09bf776f6dc0f916c6bd551d1ca6898a8c955ec19def07c49b351
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9660a724752170b6549cfdb0321416b703ab249763292c2c4c04b37e54e7435e
3
+ size 5176