SiMajid commited on
Commit
cbc6b33
1 Parent(s): 220571a

value-reward-model-opt-350m

Browse files
README.md CHANGED
@@ -1,10 +1,11 @@
1
  ---
2
  license: other
3
- base_model: facebook/opt-350m
4
  tags:
5
  - trl
6
  - reward-trainer
7
  - generated_from_trainer
 
8
  model-index:
9
  - name: facebook/opt-350m
10
  results: []
@@ -35,14 +36,12 @@ More information needed
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 1.41e-05
38
- - train_batch_size: 1
39
  - eval_batch_size: 8
40
  - seed: 42
41
- - gradient_accumulation_steps: 4
42
- - total_train_batch_size: 4
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
- - num_epochs: 5
46
 
47
  ### Training results
48
 
@@ -50,7 +49,8 @@ The following hyperparameters were used during training:
50
 
51
  ### Framework versions
52
 
 
53
  - Transformers 4.41.2
54
  - Pytorch 2.1.2
55
  - Datasets 2.19.2
56
- - Tokenizers 0.19.1
 
1
  ---
2
  license: other
3
+ library_name: peft
4
  tags:
5
  - trl
6
  - reward-trainer
7
  - generated_from_trainer
8
+ base_model: facebook/opt-350m
9
  model-index:
10
  - name: facebook/opt-350m
11
  results: []
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 1.41e-05
39
+ - train_batch_size: 8
40
  - eval_batch_size: 8
41
  - seed: 42
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
+ - num_epochs: 7
45
 
46
  ### Training results
47
 
 
49
 
50
  ### Framework versions
51
 
52
+ - PEFT 0.11.1
53
  - Transformers 4.41.2
54
  - Pytorch 2.1.2
55
  - Datasets 2.19.2
56
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -10,7 +10,7 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 32,
14
  "lora_dropout": 0.0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
@@ -20,19 +20,19 @@
20
  "score"
21
  ],
22
  "peft_type": "LORA",
23
- "r": 16,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "v_proj",
28
- "lm_head",
29
- "k_proj",
30
  "gate_proj",
31
- "down_proj",
32
- "embed_tokens",
33
- "o_proj",
34
  "up_proj",
35
- "q_proj"
 
 
 
 
 
36
  ],
37
  "task_type": "SEQ_CLS",
38
  "use_dora": false,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 64,
14
  "lora_dropout": 0.0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
 
20
  "score"
21
  ],
22
  "peft_type": "LORA",
23
+ "r": 64,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
 
27
  "gate_proj",
28
+ "lm_head",
 
 
29
  "up_proj",
30
+ "q_proj",
31
+ "embed_tokens",
32
+ "v_proj",
33
+ "down_proj",
34
+ "k_proj",
35
+ "o_proj"
36
  ],
37
  "task_type": "SEQ_CLS",
38
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbd784bcc1e082408be5fb467c5c57defb89d320179d97b14a82fccce9e08652
3
- size 115669872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8207f830136547a6ef273530b44ea0b3b39c306e3bfe7fa00721a9b6e654dfc7
3
+ size 153731952
runs/Jul21_14-23-35_690161aed707/events.out.tfevents.1721571815.690161aed707.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6eb1ea68eb6958a2c1ac0a87f929e33b70f17252bceb5e8d3f2dc1d5311b9cab
3
+ size 8657
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b09ebb86eed8455c666d7ff0f14b9e6c8b1be25436a1960cae882541494d8cbc
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdd9604c304beb8f6367010b6d205a120369030c57615a027844e910f7bb5140
3
  size 5112
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,35 +1,35 @@
1
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240720_144031-b29fhg07/logs/debug.log
11
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240720_144031-b29fhg07/logs/debug-internal.log
12
- 2024-07-20 14:40:31,636 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7f94eaea1870>
13
- 2024-07-20 14:40:31,637 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
14
- 2024-07-20 14:40:31,637 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
  config: {}
16
- 2024-07-20 14:40:31,637 INFO MainThread:34 [wandb_init.py:init():610] starting backend
17
- 2024-07-20 14:40:31,637 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
18
- 2024-07-20 14:40:31,639 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
- 2024-07-20 14:40:31,641 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
20
- 2024-07-20 14:40:31,654 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
- 2024-07-20 14:40:32,075 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
22
- 2024-07-20 14:40:32,078 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
- 2024-07-20 14:40:32,303 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
24
- 2024-07-20 14:40:32,353 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
- 2024-07-20 14:40:32,354 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
27
- 2024-07-20 14:40:48,447 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
28
- 2024-07-20 14:40:48,447 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
- 2024-07-20 14:40:48,447 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
- 2024-07-20 14:40:48,448 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
31
- 2024-07-20 14:40:48,451 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
32
- 2024-07-20 14:40:48,456 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 2048, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['OPTForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': '</s>', 'bos_token_id': 2, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/opt-350m', 'transformers_version': '4.41.2', 'activation_dropout': 0.0, 'model_type': 'opt', 'vocab_size': 50272, 'max_position_embeddings': 2048, 'num_attention_heads': 16, 'word_embed_proj_dim': 512, 'ffn_dim': 4096, 'hidden_size': 1024, 'num_hidden_layers': 24, 'dropout': 0.1, 'attention_dropout': 0.0, 'activation_function': 'relu', 'init_std': 0.02, 'layerdrop': 0.0, 'use_cache': True, 'do_layer_norm_before': False, 'enable_bias': True, 'layer_norm_elementwise_affine': True, '_remove_final_layer_norm': False, 'output_dir': './', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1.41e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Jul20_14-39-51_8fb1c041891b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'last_reward-train-facebook-opt350m_v5', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'no', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
- 2024-07-20 19:31:37,488 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
- 2024-07-20 19:31:37,493 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
35
- 2024-07-20 19:31:37,499 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
 
1
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240721_142400-5uukvmi3/logs/debug.log
11
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240721_142400-5uukvmi3/logs/debug-internal.log
12
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7cbeb281b040>
13
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
14
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
  config: {}
16
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():610] starting backend
17
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
18
+ 2024-07-21 14:24:00,535 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-07-21 14:24:00,538 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
20
+ 2024-07-21 14:24:00,551 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
+ 2024-07-21 14:24:01,312 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
22
+ 2024-07-21 14:24:01,315 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
+ 2024-07-21 14:24:01,795 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
24
+ 2024-07-21 14:24:01,880 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
+ 2024-07-21 14:24:01,881 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
27
+ 2024-07-21 14:24:17,909 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
28
+ 2024-07-21 14:24:17,910 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
+ 2024-07-21 14:24:17,910 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
+ 2024-07-21 14:24:17,910 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
31
+ 2024-07-21 14:24:17,915 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
32
+ 2024-07-21 14:24:17,920 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 2048, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['OPTForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': '</s>', 'bos_token_id': 2, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/opt-350m', 'transformers_version': '4.41.2', 'activation_dropout': 0.0, 'model_type': 'opt', 'vocab_size': 50272, 'max_position_embeddings': 2048, 'num_attention_heads': 16, 'word_embed_proj_dim': 512, 'ffn_dim': 4096, 'hidden_size': 1024, 'num_hidden_layers': 24, 'dropout': 0.1, 'attention_dropout': 0.0, 'activation_function': 'relu', 'init_std': 0.02, 'layerdrop': 0.0, 'use_cache': True, 'do_layer_norm_before': False, 'enable_bias': True, 'layer_norm_elementwise_affine': True, '_remove_final_layer_norm': False, 'output_dir': './', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1.41e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 7, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Jul21_14-23-35_690161aed707', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 250, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'value-reward-model-opt-350m', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'no', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
+ 2024-07-21 21:46:45,899 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-07-21 21:46:45,900 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
35
+ 2024-07-21 21:50:24,607 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
wandb/run-20240721_142400-5uukvmi3/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240721_142400-5uukvmi3/files/config.yaml ADDED
@@ -0,0 +1,693 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.17.0
8
+ framework: huggingface
9
+ huggingface_version: 4.41.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1721571840
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 13
47
+ - 23
48
+ - 62
49
+ - 66
50
+ 4: 3.10.13
51
+ 5: 0.17.0
52
+ 6: 4.41.2
53
+ 8:
54
+ - 1
55
+ - 2
56
+ - 5
57
+ 9:
58
+ 1: transformers_trainer
59
+ 13: linux-x86_64
60
+ m:
61
+ - 1: train/global_step
62
+ 6:
63
+ - 3
64
+ - 1: train/loss
65
+ 5: 1
66
+ 6:
67
+ - 1
68
+ - 1: train/grad_norm
69
+ 5: 1
70
+ 6:
71
+ - 1
72
+ - 1: train/learning_rate
73
+ 5: 1
74
+ 6:
75
+ - 1
76
+ - 1: train/epoch
77
+ 5: 1
78
+ 6:
79
+ - 1
80
+ return_dict:
81
+ desc: null
82
+ value: true
83
+ output_hidden_states:
84
+ desc: null
85
+ value: false
86
+ output_attentions:
87
+ desc: null
88
+ value: false
89
+ torchscript:
90
+ desc: null
91
+ value: false
92
+ torch_dtype:
93
+ desc: null
94
+ value: float16
95
+ use_bfloat16:
96
+ desc: null
97
+ value: false
98
+ tf_legacy_loss:
99
+ desc: null
100
+ value: false
101
+ pruned_heads:
102
+ desc: null
103
+ value: {}
104
+ tie_word_embeddings:
105
+ desc: null
106
+ value: true
107
+ chunk_size_feed_forward:
108
+ desc: null
109
+ value: 0
110
+ is_encoder_decoder:
111
+ desc: null
112
+ value: false
113
+ is_decoder:
114
+ desc: null
115
+ value: false
116
+ cross_attention_hidden_size:
117
+ desc: null
118
+ value: null
119
+ add_cross_attention:
120
+ desc: null
121
+ value: false
122
+ tie_encoder_decoder:
123
+ desc: null
124
+ value: false
125
+ max_length:
126
+ desc: null
127
+ value: 2048
128
+ min_length:
129
+ desc: null
130
+ value: 0
131
+ do_sample:
132
+ desc: null
133
+ value: false
134
+ early_stopping:
135
+ desc: null
136
+ value: false
137
+ num_beams:
138
+ desc: null
139
+ value: 1
140
+ num_beam_groups:
141
+ desc: null
142
+ value: 1
143
+ diversity_penalty:
144
+ desc: null
145
+ value: 0.0
146
+ temperature:
147
+ desc: null
148
+ value: 1.0
149
+ top_k:
150
+ desc: null
151
+ value: 50
152
+ top_p:
153
+ desc: null
154
+ value: 1.0
155
+ typical_p:
156
+ desc: null
157
+ value: 1.0
158
+ repetition_penalty:
159
+ desc: null
160
+ value: 1.0
161
+ length_penalty:
162
+ desc: null
163
+ value: 1.0
164
+ no_repeat_ngram_size:
165
+ desc: null
166
+ value: 0
167
+ encoder_no_repeat_ngram_size:
168
+ desc: null
169
+ value: 0
170
+ bad_words_ids:
171
+ desc: null
172
+ value: null
173
+ num_return_sequences:
174
+ desc: null
175
+ value: 1
176
+ output_scores:
177
+ desc: null
178
+ value: false
179
+ return_dict_in_generate:
180
+ desc: null
181
+ value: false
182
+ forced_bos_token_id:
183
+ desc: null
184
+ value: null
185
+ forced_eos_token_id:
186
+ desc: null
187
+ value: null
188
+ remove_invalid_values:
189
+ desc: null
190
+ value: false
191
+ exponential_decay_length_penalty:
192
+ desc: null
193
+ value: null
194
+ suppress_tokens:
195
+ desc: null
196
+ value: null
197
+ begin_suppress_tokens:
198
+ desc: null
199
+ value: null
200
+ architectures:
201
+ desc: null
202
+ value:
203
+ - OPTForCausalLM
204
+ finetuning_task:
205
+ desc: null
206
+ value: null
207
+ id2label:
208
+ desc: null
209
+ value:
210
+ '0': LABEL_0
211
+ '1': LABEL_1
212
+ label2id:
213
+ desc: null
214
+ value:
215
+ LABEL_0: 0
216
+ LABEL_1: 1
217
+ tokenizer_class:
218
+ desc: null
219
+ value: null
220
+ prefix:
221
+ desc: null
222
+ value: </s>
223
+ bos_token_id:
224
+ desc: null
225
+ value: 2
226
+ pad_token_id:
227
+ desc: null
228
+ value: 1
229
+ eos_token_id:
230
+ desc: null
231
+ value: 2
232
+ sep_token_id:
233
+ desc: null
234
+ value: null
235
+ decoder_start_token_id:
236
+ desc: null
237
+ value: null
238
+ task_specific_params:
239
+ desc: null
240
+ value: null
241
+ problem_type:
242
+ desc: null
243
+ value: null
244
+ _name_or_path:
245
+ desc: null
246
+ value: facebook/opt-350m
247
+ transformers_version:
248
+ desc: null
249
+ value: 4.41.2
250
+ activation_dropout:
251
+ desc: null
252
+ value: 0.0
253
+ model_type:
254
+ desc: null
255
+ value: opt
256
+ vocab_size:
257
+ desc: null
258
+ value: 50272
259
+ max_position_embeddings:
260
+ desc: null
261
+ value: 2048
262
+ num_attention_heads:
263
+ desc: null
264
+ value: 16
265
+ word_embed_proj_dim:
266
+ desc: null
267
+ value: 512
268
+ ffn_dim:
269
+ desc: null
270
+ value: 4096
271
+ hidden_size:
272
+ desc: null
273
+ value: 1024
274
+ num_hidden_layers:
275
+ desc: null
276
+ value: 24
277
+ dropout:
278
+ desc: null
279
+ value: 0.1
280
+ attention_dropout:
281
+ desc: null
282
+ value: 0.0
283
+ activation_function:
284
+ desc: null
285
+ value: relu
286
+ init_std:
287
+ desc: null
288
+ value: 0.02
289
+ layerdrop:
290
+ desc: null
291
+ value: 0.0
292
+ use_cache:
293
+ desc: null
294
+ value: true
295
+ do_layer_norm_before:
296
+ desc: null
297
+ value: false
298
+ enable_bias:
299
+ desc: null
300
+ value: true
301
+ layer_norm_elementwise_affine:
302
+ desc: null
303
+ value: true
304
+ _remove_final_layer_norm:
305
+ desc: null
306
+ value: false
307
+ output_dir:
308
+ desc: null
309
+ value: ./
310
+ overwrite_output_dir:
311
+ desc: null
312
+ value: false
313
+ do_train:
314
+ desc: null
315
+ value: false
316
+ do_eval:
317
+ desc: null
318
+ value: false
319
+ do_predict:
320
+ desc: null
321
+ value: false
322
+ eval_strategy:
323
+ desc: null
324
+ value: 'no'
325
+ prediction_loss_only:
326
+ desc: null
327
+ value: false
328
+ per_device_train_batch_size:
329
+ desc: null
330
+ value: 8
331
+ per_device_eval_batch_size:
332
+ desc: null
333
+ value: 8
334
+ per_gpu_train_batch_size:
335
+ desc: null
336
+ value: null
337
+ per_gpu_eval_batch_size:
338
+ desc: null
339
+ value: null
340
+ gradient_accumulation_steps:
341
+ desc: null
342
+ value: 1
343
+ eval_accumulation_steps:
344
+ desc: null
345
+ value: null
346
+ eval_delay:
347
+ desc: null
348
+ value: 0
349
+ learning_rate:
350
+ desc: null
351
+ value: 1.41e-05
352
+ weight_decay:
353
+ desc: null
354
+ value: 0.0
355
+ adam_beta1:
356
+ desc: null
357
+ value: 0.9
358
+ adam_beta2:
359
+ desc: null
360
+ value: 0.999
361
+ adam_epsilon:
362
+ desc: null
363
+ value: 1.0e-08
364
+ max_grad_norm:
365
+ desc: null
366
+ value: 1.0
367
+ num_train_epochs:
368
+ desc: null
369
+ value: 7
370
+ max_steps:
371
+ desc: null
372
+ value: -1
373
+ lr_scheduler_type:
374
+ desc: null
375
+ value: linear
376
+ lr_scheduler_kwargs:
377
+ desc: null
378
+ value: {}
379
+ warmup_ratio:
380
+ desc: null
381
+ value: 0.0
382
+ warmup_steps:
383
+ desc: null
384
+ value: 0
385
+ log_level:
386
+ desc: null
387
+ value: passive
388
+ log_level_replica:
389
+ desc: null
390
+ value: warning
391
+ log_on_each_node:
392
+ desc: null
393
+ value: true
394
+ logging_dir:
395
+ desc: null
396
+ value: ./runs/Jul21_14-23-35_690161aed707
397
+ logging_strategy:
398
+ desc: null
399
+ value: steps
400
+ logging_first_step:
401
+ desc: null
402
+ value: false
403
+ logging_steps:
404
+ desc: null
405
+ value: 250
406
+ logging_nan_inf_filter:
407
+ desc: null
408
+ value: true
409
+ save_strategy:
410
+ desc: null
411
+ value: steps
412
+ save_steps:
413
+ desc: null
414
+ value: 500
415
+ save_total_limit:
416
+ desc: null
417
+ value: null
418
+ save_safetensors:
419
+ desc: null
420
+ value: true
421
+ save_on_each_node:
422
+ desc: null
423
+ value: false
424
+ save_only_model:
425
+ desc: null
426
+ value: false
427
+ restore_callback_states_from_checkpoint:
428
+ desc: null
429
+ value: false
430
+ no_cuda:
431
+ desc: null
432
+ value: false
433
+ use_cpu:
434
+ desc: null
435
+ value: false
436
+ use_mps_device:
437
+ desc: null
438
+ value: false
439
+ seed:
440
+ desc: null
441
+ value: 42
442
+ data_seed:
443
+ desc: null
444
+ value: null
445
+ jit_mode_eval:
446
+ desc: null
447
+ value: false
448
+ use_ipex:
449
+ desc: null
450
+ value: false
451
+ bf16:
452
+ desc: null
453
+ value: false
454
+ fp16:
455
+ desc: null
456
+ value: false
457
+ fp16_opt_level:
458
+ desc: null
459
+ value: O1
460
+ half_precision_backend:
461
+ desc: null
462
+ value: auto
463
+ bf16_full_eval:
464
+ desc: null
465
+ value: false
466
+ fp16_full_eval:
467
+ desc: null
468
+ value: false
469
+ tf32:
470
+ desc: null
471
+ value: null
472
+ local_rank:
473
+ desc: null
474
+ value: 0
475
+ ddp_backend:
476
+ desc: null
477
+ value: null
478
+ tpu_num_cores:
479
+ desc: null
480
+ value: null
481
+ tpu_metrics_debug:
482
+ desc: null
483
+ value: false
484
+ debug:
485
+ desc: null
486
+ value: []
487
+ dataloader_drop_last:
488
+ desc: null
489
+ value: false
490
+ eval_steps:
491
+ desc: null
492
+ value: null
493
+ dataloader_num_workers:
494
+ desc: null
495
+ value: 0
496
+ dataloader_prefetch_factor:
497
+ desc: null
498
+ value: null
499
+ past_index:
500
+ desc: null
501
+ value: -1
502
+ run_name:
503
+ desc: null
504
+ value: value-reward-model-opt-350m
505
+ disable_tqdm:
506
+ desc: null
507
+ value: false
508
+ remove_unused_columns:
509
+ desc: null
510
+ value: false
511
+ label_names:
512
+ desc: null
513
+ value: null
514
+ load_best_model_at_end:
515
+ desc: null
516
+ value: false
517
+ metric_for_best_model:
518
+ desc: null
519
+ value: null
520
+ greater_is_better:
521
+ desc: null
522
+ value: null
523
+ ignore_data_skip:
524
+ desc: null
525
+ value: false
526
+ fsdp:
527
+ desc: null
528
+ value: []
529
+ fsdp_min_num_params:
530
+ desc: null
531
+ value: 0
532
+ fsdp_config:
533
+ desc: null
534
+ value:
535
+ min_num_params: 0
536
+ xla: false
537
+ xla_fsdp_v2: false
538
+ xla_fsdp_grad_ckpt: false
539
+ fsdp_transformer_layer_cls_to_wrap:
540
+ desc: null
541
+ value: null
542
+ accelerator_config:
543
+ desc: null
544
+ value:
545
+ split_batches: false
546
+ dispatch_batches: null
547
+ even_batches: true
548
+ use_seedable_sampler: true
549
+ non_blocking: false
550
+ gradient_accumulation_kwargs: null
551
+ deepspeed:
552
+ desc: null
553
+ value: null
554
+ label_smoothing_factor:
555
+ desc: null
556
+ value: 0.0
557
+ optim:
558
+ desc: null
559
+ value: adamw_torch
560
+ optim_args:
561
+ desc: null
562
+ value: null
563
+ adafactor:
564
+ desc: null
565
+ value: false
566
+ group_by_length:
567
+ desc: null
568
+ value: false
569
+ length_column_name:
570
+ desc: null
571
+ value: length
572
+ report_to:
573
+ desc: null
574
+ value:
575
+ - tensorboard
576
+ - wandb
577
+ ddp_find_unused_parameters:
578
+ desc: null
579
+ value: null
580
+ ddp_bucket_cap_mb:
581
+ desc: null
582
+ value: null
583
+ ddp_broadcast_buffers:
584
+ desc: null
585
+ value: null
586
+ dataloader_pin_memory:
587
+ desc: null
588
+ value: true
589
+ dataloader_persistent_workers:
590
+ desc: null
591
+ value: false
592
+ skip_memory_metrics:
593
+ desc: null
594
+ value: true
595
+ use_legacy_prediction_loop:
596
+ desc: null
597
+ value: false
598
+ push_to_hub:
599
+ desc: null
600
+ value: false
601
+ resume_from_checkpoint:
602
+ desc: null
603
+ value: null
604
+ hub_model_id:
605
+ desc: null
606
+ value: null
607
+ hub_strategy:
608
+ desc: null
609
+ value: every_save
610
+ hub_token:
611
+ desc: null
612
+ value: <HUB_TOKEN>
613
+ hub_private_repo:
614
+ desc: null
615
+ value: false
616
+ hub_always_push:
617
+ desc: null
618
+ value: false
619
+ gradient_checkpointing:
620
+ desc: null
621
+ value: true
622
+ gradient_checkpointing_kwargs:
623
+ desc: null
624
+ value: null
625
+ include_inputs_for_metrics:
626
+ desc: null
627
+ value: false
628
+ eval_do_concat_batches:
629
+ desc: null
630
+ value: true
631
+ fp16_backend:
632
+ desc: null
633
+ value: auto
634
+ evaluation_strategy:
635
+ desc: null
636
+ value: 'no'
637
+ push_to_hub_model_id:
638
+ desc: null
639
+ value: null
640
+ push_to_hub_organization:
641
+ desc: null
642
+ value: null
643
+ push_to_hub_token:
644
+ desc: null
645
+ value: <PUSH_TO_HUB_TOKEN>
646
+ mp_parameters:
647
+ desc: null
648
+ value: ''
649
+ auto_find_batch_size:
650
+ desc: null
651
+ value: false
652
+ full_determinism:
653
+ desc: null
654
+ value: false
655
+ torchdynamo:
656
+ desc: null
657
+ value: null
658
+ ray_scope:
659
+ desc: null
660
+ value: last
661
+ ddp_timeout:
662
+ desc: null
663
+ value: 1800
664
+ torch_compile:
665
+ desc: null
666
+ value: false
667
+ torch_compile_backend:
668
+ desc: null
669
+ value: null
670
+ torch_compile_mode:
671
+ desc: null
672
+ value: null
673
+ dispatch_batches:
674
+ desc: null
675
+ value: null
676
+ split_batches:
677
+ desc: null
678
+ value: null
679
+ include_tokens_per_second:
680
+ desc: null
681
+ value: false
682
+ include_num_input_tokens_seen:
683
+ desc: null
684
+ value: false
685
+ neftune_noise_alpha:
686
+ desc: null
687
+ value: null
688
+ optim_target_modules:
689
+ desc: null
690
+ value: null
691
+ batch_eval_metrics:
692
+ desc: null
693
+ value: false
wandb/run-20240721_142400-5uukvmi3/files/output.log ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ You're using a GPT2TokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
2
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
3
+ warnings.warn(
4
+ `use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`...
5
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
6
+ warnings.warn(
7
+ Could not estimate the number of tokens of the input, floating-point operations will not be computed
8
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
9
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
10
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
11
+ warnings.warn(
12
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
13
+ warnings.warn(
14
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
15
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
16
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
17
+ warnings.warn(
18
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
19
+ warnings.warn(
20
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
21
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
22
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
23
+ warnings.warn(
24
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
25
+ warnings.warn(
26
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
27
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
28
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
29
+ warnings.warn(
30
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
31
+ warnings.warn(
32
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
33
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
34
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
35
+ warnings.warn(
36
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
37
+ warnings.warn(
38
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
39
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
40
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
41
+ warnings.warn(
42
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
43
+ warnings.warn(
44
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
45
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
46
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
47
+ warnings.warn(
48
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
49
+ warnings.warn(
50
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
51
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
52
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
53
+ warnings.warn(
54
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
55
+ warnings.warn(
56
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:180: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
wandb/run-20240721_142400-5uukvmi3/files/requirements.txt ADDED
@@ -0,0 +1,872 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.1.0
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.23.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.3
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.5
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.3
44
+ absl-py==1.4.0
45
+ accelerate==0.30.1
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.13.0
49
+ aiofiles==22.1.0
50
+ aiohttp==3.9.1
51
+ aioitertools==0.11.0
52
+ aiorwlock==1.3.0
53
+ aiosignal==1.3.1
54
+ aiosqlite==0.19.0
55
+ albumentations==1.4.0
56
+ alembic==1.13.1
57
+ altair==5.3.0
58
+ annotated-types==0.6.0
59
+ annotated-types==0.7.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.18.0
71
+ astroid==3.2.2
72
+ astropy-iers-data==0.2024.6.3.0.31.14
73
+ astropy==6.1.0
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ blake3==0.2.1
86
+ bleach==6.1.0
87
+ blessed==1.20.0
88
+ blinker==1.8.2
89
+ blis==0.7.10
90
+ blosc2==2.6.2
91
+ bokeh==3.4.1
92
+ boltons==23.1.1
93
+ boto3==1.26.100
94
+ botocore==1.34.106
95
+ bq_helper==0.4.1
96
+ bqplot==0.12.43
97
+ branca==0.7.2
98
+ brewer2mpl==1.4.1
99
+ brotlipy==0.7.0
100
+ cached-property==1.5.2
101
+ cachetools==4.2.4
102
+ cachetools==5.3.2
103
+ catalogue==2.0.10
104
+ catalyst==22.4
105
+ catboost==1.2.5
106
+ category-encoders==2.6.3
107
+ certifi==2024.2.2
108
+ cesium==0.12.1
109
+ cffi==1.16.0
110
+ charset-normalizer==3.3.2
111
+ chex==0.1.86
112
+ cleverhans==4.0.0
113
+ click-plugins==1.1.1
114
+ click==8.1.7
115
+ cligj==0.7.2
116
+ cloud-tpu-client==0.10
117
+ cloud-tpu-profiler==2.4.0
118
+ cloudpathlib==0.16.0
119
+ cloudpickle==2.2.1
120
+ cloudpickle==3.0.0
121
+ cmdstanpy==1.2.3
122
+ colorama==0.4.6
123
+ colorcet==3.1.0
124
+ colorful==0.5.6
125
+ colorlog==6.8.2
126
+ colorlover==0.3.0
127
+ comm==0.2.1
128
+ conda-libmamba-solver==23.12.0
129
+ conda-package-handling==2.2.0
130
+ conda==24.5.0
131
+ conda_package_streaming==0.9.0
132
+ confection==0.1.4
133
+ contextily==1.6.0
134
+ contourpy==1.2.0
135
+ contourpy==1.2.1
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.5.0
140
+ cudf==24.4.1
141
+ cufflinks==0.17.3
142
+ cuml==24.4.0
143
+ cupy==13.1.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.4.0
148
+ daal==2024.4.0
149
+ dacite==1.8.1
150
+ dask-cuda==24.4.0
151
+ dask-cudf==24.4.1
152
+ dask-expr==1.1.2
153
+ dask==2024.5.2
154
+ dataclasses-json==0.6.6
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.19.2
157
+ datashader==0.16.2
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==7.0.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.8
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2024.1.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-to-markdown==0.15
177
+ docstring_parser==0.16
178
+ docutils==0.21.2
179
+ earthengine-api==0.1.405
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.12.1
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.15
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.43
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.31.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.4
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.53.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.4
220
+ frozenlist==1.4.1
221
+ fsspec==2024.3.1
222
+ fsspec==2024.5.0
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.3.1
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.4
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.12.0
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.6.4
241
+ google-api-core==2.11.1
242
+ google-api-core==2.19.0
243
+ google-api-python-client==2.131.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.6.0
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.59.3
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.2
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.23.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.5
312
+ imagecodecs==2024.6.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.3
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.19.1
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.11.1
331
+ ipyvuetify==1.9.4
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.2.0
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.26
341
+ jaxlib==0.4.26.dev20240504
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.4.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-leaflet==0.19.1
356
+ jupyter-lsp==1.5.1
357
+ jupyter-server-mathjax==0.2.6
358
+ jupyter-ydoc==0.2.5
359
+ jupyter_client==7.4.9
360
+ jupyter_client==8.6.0
361
+ jupyter_core==5.7.1
362
+ jupyter_server==2.12.5
363
+ jupyter_server_fileid==0.9.1
364
+ jupyter_server_proxy==4.1.0
365
+ jupyter_server_terminals==0.5.1
366
+ jupyter_server_ydoc==0.8.0
367
+ jupyterlab-lsp==5.1.0
368
+ jupyterlab-widgets==3.0.9
369
+ jupyterlab==4.2.1
370
+ jupyterlab_git==0.44.0
371
+ jupyterlab_pygments==0.3.0
372
+ jupyterlab_server==2.27.2
373
+ jupytext==1.16.0
374
+ kaggle-environments==1.14.11
375
+ kaggle==1.6.14
376
+ kagglehub==0.2.5
377
+ keras-cv==0.9.0
378
+ keras-nlp==0.12.1
379
+ keras-tuner==1.4.6
380
+ keras==3.3.3
381
+ kernels-mixer==0.0.7
382
+ keyring==24.3.0
383
+ keyrings.google-artifactregistry-auth==1.1.2
384
+ kfp-pipeline-spec==0.2.2
385
+ kfp-server-api==2.0.5
386
+ kfp==2.5.0
387
+ kiwisolver==1.4.5
388
+ kmapper==2.0.1
389
+ kmodes==0.12.2
390
+ korean-lunar-calendar==0.3.1
391
+ kornia==0.7.2
392
+ kornia_rs==0.1.3
393
+ kt-legacy==1.0.5
394
+ kubernetes==26.1.0
395
+ langcodes==3.4.0
396
+ langid==1.1.6
397
+ language_data==1.2.0
398
+ lazy_loader==0.3
399
+ learntools==0.3.4
400
+ leven==1.0.4
401
+ libclang==16.0.6
402
+ libmambapy==1.5.8
403
+ libpysal==4.9.2
404
+ librosa==0.10.2.post1
405
+ lightgbm==4.2.0
406
+ lightning-utilities==0.11.2
407
+ lime==0.2.0.1
408
+ line_profiler==4.1.3
409
+ linkify-it-py==2.0.3
410
+ llvmlite==0.41.1
411
+ llvmlite==0.42.0
412
+ lml==0.1.0
413
+ locket==1.0.0
414
+ loguru==0.7.2
415
+ lxml==5.2.2
416
+ lz4==4.3.3
417
+ mamba==1.5.8
418
+ mapclassify==2.6.1
419
+ marisa-trie==1.1.0
420
+ markdown-it-py==3.0.0
421
+ marshmallow==3.21.2
422
+ matplotlib-inline==0.1.6
423
+ matplotlib-venn==0.11.10
424
+ matplotlib==3.7.5
425
+ matplotlib==3.8.4
426
+ mccabe==0.7.0
427
+ mdit-py-plugins==0.4.0
428
+ mdurl==0.1.2
429
+ memory-profiler==0.61.0
430
+ menuinst==2.0.1
431
+ mercantile==1.2.1
432
+ mgwr==2.2.1
433
+ missingno==0.5.2
434
+ mistune==0.8.4
435
+ mizani==0.11.4
436
+ ml-dtypes==0.2.0
437
+ mlcrate==0.2.0
438
+ mlens==0.2.3
439
+ mlxtend==0.23.1
440
+ mne==1.7.0
441
+ mnist==0.2.2
442
+ momepy==0.7.0
443
+ more-itertools==10.2.0
444
+ mpld3==0.5.10
445
+ mpmath==1.3.0
446
+ msgpack==1.0.7
447
+ msgpack==1.0.8
448
+ multidict==6.0.4
449
+ multimethod==1.10
450
+ multipledispatch==1.0.0
451
+ multiprocess==0.70.16
452
+ munkres==1.1.4
453
+ murmurhash==1.0.10
454
+ mypy-extensions==1.0.0
455
+ namex==0.0.8
456
+ nb-conda-kernels==2.3.1
457
+ nb_conda==2.2.1
458
+ nbclassic==1.0.0
459
+ nbclient==0.5.13
460
+ nbclient==0.9.0
461
+ nbconvert==6.4.5
462
+ nbdime==3.2.0
463
+ nbformat==5.9.2
464
+ ndindex==1.8
465
+ nest-asyncio==1.5.8
466
+ networkx==3.2.1
467
+ nibabel==5.2.1
468
+ nilearn==0.10.4
469
+ ninja==1.11.1.1
470
+ nltk==3.2.4
471
+ nose==1.3.7
472
+ notebook==6.5.4
473
+ notebook==6.5.6
474
+ notebook_executor==0.2
475
+ notebook_shim==0.2.3
476
+ numba==0.58.1
477
+ numba==0.59.1
478
+ numexpr==2.10.0
479
+ numpy==1.26.4
480
+ nvidia-ml-py==11.495.46
481
+ nvtx==0.2.10
482
+ oauth2client==4.1.3
483
+ oauthlib==3.2.2
484
+ objsize==0.6.1
485
+ odfpy==1.4.1
486
+ olefile==0.47
487
+ onnx==1.16.1
488
+ opencensus-context==0.1.3
489
+ opencensus==0.11.4
490
+ opencv-contrib-python==4.10.0.82
491
+ opencv-python-headless==4.10.0.82
492
+ opencv-python==4.10.0.82
493
+ openpyxl==3.1.3
494
+ openslide-python==1.3.1
495
+ opentelemetry-api==1.22.0
496
+ opentelemetry-exporter-otlp-proto-common==1.22.0
497
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
498
+ opentelemetry-exporter-otlp-proto-http==1.22.0
499
+ opentelemetry-exporter-otlp==1.22.0
500
+ opentelemetry-proto==1.22.0
501
+ opentelemetry-sdk==1.22.0
502
+ opentelemetry-semantic-conventions==0.43b0
503
+ opt-einsum==3.3.0
504
+ optax==0.2.2
505
+ optree==0.11.0
506
+ optuna==3.6.1
507
+ orbax-checkpoint==0.5.15
508
+ ordered-set==4.1.0
509
+ orjson==3.9.10
510
+ ortools==9.4.1874
511
+ osmnx==1.9.3
512
+ overrides==7.4.0
513
+ packaging==21.3
514
+ pandas-datareader==0.10.0
515
+ pandas-profiling==3.6.6
516
+ pandas-summary==0.2.0
517
+ pandas==2.2.1
518
+ pandas==2.2.2
519
+ pandasql==0.7.3
520
+ pandocfilters==1.5.0
521
+ panel==1.4.4
522
+ papermill==2.5.0
523
+ param==2.1.0
524
+ parso==0.8.3
525
+ partd==1.4.2
526
+ path.py==12.5.0
527
+ path==16.14.0
528
+ pathos==0.3.2
529
+ pathy==0.10.3
530
+ patsy==0.5.6
531
+ pdf2image==1.17.0
532
+ peft==0.11.1
533
+ pettingzoo==1.24.0
534
+ pexpect==4.8.0
535
+ pexpect==4.9.0
536
+ phik==0.12.4
537
+ pickleshare==0.7.5
538
+ pillow==10.3.0
539
+ pip==23.3.2
540
+ pkgutil_resolve_name==1.3.10
541
+ platformdirs==3.11.0
542
+ platformdirs==4.1.0
543
+ plotly-express==0.4.1
544
+ plotly==5.18.0
545
+ plotnine==0.13.6
546
+ pluggy==1.5.0
547
+ pointpats==2.4.0
548
+ polars==0.20.31
549
+ polyglot==16.7.4
550
+ pooch==1.8.1
551
+ pox==0.3.4
552
+ ppca==0.0.4
553
+ ppft==1.7.6.8
554
+ preprocessing==0.1.13
555
+ preshed==3.0.9
556
+ prettytable==3.9.0
557
+ progressbar2==4.4.2
558
+ prometheus-client==0.19.0
559
+ promise==2.3
560
+ prompt-toolkit==3.0.42
561
+ prompt-toolkit==3.0.43
562
+ prophet==1.1.1
563
+ proto-plus==1.23.0
564
+ protobuf==3.20.3
565
+ protobuf==4.24.4
566
+ psutil==5.9.3
567
+ psutil==5.9.7
568
+ ptyprocess==0.7.0
569
+ pudb==2024.1
570
+ pure-eval==0.2.2
571
+ py-cpuinfo==9.0.0
572
+ py-spy==0.3.14
573
+ py4j==0.10.9.7
574
+ pyLDAvis==3.4.1
575
+ pyOpenSSL==23.3.0
576
+ pyaml==24.4.0
577
+ pyarrow-hotfix==0.6
578
+ pyarrow==14.0.2
579
+ pyasn1-modules==0.3.0
580
+ pyasn1==0.5.1
581
+ pybind11==2.12.0
582
+ pyclipper==1.3.0.post5
583
+ pycodestyle==2.11.1
584
+ pycosat==0.6.6
585
+ pycparser==2.21
586
+ pycryptodome==3.20.0
587
+ pyct==0.5.0
588
+ pycuda==2024.1
589
+ pydantic==2.5.3
590
+ pydantic==2.7.2
591
+ pydantic_core==2.14.6
592
+ pydantic_core==2.18.3
593
+ pydegensac==0.1.2
594
+ pydicom==2.4.4
595
+ pydocstyle==6.3.0
596
+ pydot==1.4.2
597
+ pydub==0.25.1
598
+ pyemd==1.0.0
599
+ pyerfa==2.0.1.4
600
+ pyexcel-io==0.6.6
601
+ pyexcel-ods==0.6.0
602
+ pyflakes==3.2.0
603
+ pygltflib==1.16.2
604
+ pykalman==0.9.7
605
+ pylibraft==24.4.0
606
+ pylint==3.2.2
607
+ pymc3==3.11.4
608
+ pymongo==3.13.0
609
+ pynndescent==0.5.12
610
+ pynvjitlink==0.2.3
611
+ pynvml==11.4.1
612
+ pynvrtc==9.2
613
+ pyparsing==3.1.1
614
+ pyparsing==3.1.2
615
+ pypdf==4.2.0
616
+ pyproj==3.6.1
617
+ pysal==24.1
618
+ pyshp==2.3.1
619
+ pytesseract==0.3.10
620
+ pytest==8.2.1
621
+ python-bidi==0.4.2
622
+ python-dateutil==2.9.0.post0
623
+ python-dotenv==1.0.0
624
+ python-json-logger==2.0.7
625
+ python-louvain==0.16
626
+ python-lsp-jsonrpc==1.1.2
627
+ python-lsp-server==1.11.0
628
+ python-slugify==8.0.4
629
+ python-utils==3.8.2
630
+ pythreejs==2.4.2
631
+ pytoolconfig==1.3.1
632
+ pytools==2024.1.3
633
+ pytorch-ignite==0.5.0.post2
634
+ pytorch-lightning==2.2.5
635
+ pytz==2023.3.post1
636
+ pytz==2024.1
637
+ pyu2f==0.1.5
638
+ pyviz_comms==3.0.2
639
+ pyzmq==24.0.1
640
+ pyzmq==25.1.2
641
+ qgrid==1.3.1
642
+ qtconsole==5.5.2
643
+ quantecon==0.7.2
644
+ qudida==0.0.4
645
+ raft-dask==24.4.0
646
+ rapids-dask-dependency==24.4.1a0
647
+ rasterio==1.3.10
648
+ rasterstats==0.19.0
649
+ ray-cpp==2.9.0
650
+ ray==2.9.0
651
+ referencing==0.32.1
652
+ regex==2023.12.25
653
+ requests-oauthlib==1.3.1
654
+ requests-toolbelt==0.10.1
655
+ requests==2.32.3
656
+ retrying==1.3.3
657
+ retrying==1.3.4
658
+ rfc3339-validator==0.1.4
659
+ rfc3986-validator==0.1.1
660
+ rgf-python==3.12.0
661
+ rich-click==1.8.2
662
+ rich==13.7.0
663
+ rich==13.7.1
664
+ rmm==24.4.0
665
+ rope==1.13.0
666
+ rpds-py==0.16.2
667
+ rsa==4.9
668
+ ruamel-yaml-conda==0.15.100
669
+ ruamel.yaml.clib==0.2.7
670
+ ruamel.yaml==0.18.5
671
+ s2sphere==0.2.5
672
+ s3fs==2024.3.1
673
+ s3transfer==0.6.2
674
+ safetensors==0.4.3
675
+ scattertext==0.1.19
676
+ scikit-image==0.22.0
677
+ scikit-learn-intelex==2024.4.0
678
+ scikit-learn==1.2.2
679
+ scikit-multilearn==0.2.0
680
+ scikit-optimize==0.10.1
681
+ scikit-plot==0.3.7
682
+ scikit-surprise==1.1.4
683
+ scipy==1.11.4
684
+ scipy==1.13.1
685
+ seaborn==0.12.2
686
+ segment_anything==1.0
687
+ segregation==2.5
688
+ semver==3.0.2
689
+ sentencepiece==0.2.0
690
+ sentry-sdk==2.3.1
691
+ setproctitle==1.3.3
692
+ setuptools-git==1.2
693
+ setuptools-scm==8.1.0
694
+ setuptools==69.0.3
695
+ shap==0.44.1
696
+ shapely==2.0.4
697
+ shellingham==1.5.4
698
+ shtab==1.7.1
699
+ simpervisor==1.0.0
700
+ simplejson==3.19.2
701
+ six==1.16.0
702
+ sklearn-pandas==2.2.0
703
+ slicer==0.0.7
704
+ smart-open==6.4.0
705
+ smmap==5.0.1
706
+ sniffio==1.3.0
707
+ snowballstemmer==2.2.0
708
+ snuggs==1.4.7
709
+ sortedcontainers==2.4.0
710
+ soundfile==0.12.1
711
+ soupsieve==2.5
712
+ soxr==0.3.7
713
+ spacy-legacy==3.0.12
714
+ spacy-loggers==1.0.5
715
+ spacy==3.7.3
716
+ spaghetti==1.7.5.post1
717
+ spectral==0.23.1
718
+ spglm==1.1.0
719
+ sphinx-rtd-theme==0.2.4
720
+ spint==1.0.7
721
+ splot==1.1.5.post1
722
+ spopt==0.6.0
723
+ spreg==1.4.2
724
+ spvcm==0.3.0
725
+ sqlparse==0.4.4
726
+ squarify==0.4.3
727
+ srsly==2.4.8
728
+ stable-baselines3==2.1.0
729
+ stack-data==0.6.2
730
+ stack-data==0.6.3
731
+ stanio==0.5.0
732
+ starlette==0.32.0.post1
733
+ statsmodels==0.14.1
734
+ stemming==1.0.1
735
+ stop-words==2018.7.23
736
+ stopit==1.1.2
737
+ stumpy==1.12.0
738
+ sympy==1.12.1
739
+ tables==3.9.2
740
+ tabulate==0.9.0
741
+ tangled-up-in-unicode==0.2.0
742
+ tbb==2021.12.0
743
+ tblib==3.0.0
744
+ tenacity==8.2.3
745
+ tensorboard-data-server==0.7.2
746
+ tensorboard-plugin-profile==2.15.0
747
+ tensorboard==2.15.1
748
+ tensorboardX==2.6.2.2
749
+ tensorflow-cloud==0.1.16
750
+ tensorflow-datasets==4.9.4
751
+ tensorflow-decision-forests==1.8.1
752
+ tensorflow-estimator==2.15.0
753
+ tensorflow-hub==0.16.1
754
+ tensorflow-io-gcs-filesystem==0.35.0
755
+ tensorflow-io==0.35.0
756
+ tensorflow-metadata==0.14.0
757
+ tensorflow-probability==0.23.0
758
+ tensorflow-serving-api==2.14.1
759
+ tensorflow-text==2.15.0
760
+ tensorflow-transform==0.14.0
761
+ tensorflow==2.15.0
762
+ tensorstore==0.1.60
763
+ termcolor==2.4.0
764
+ terminado==0.18.0
765
+ testpath==0.6.0
766
+ text-unidecode==1.3
767
+ textblob==0.18.0.post0
768
+ texttable==1.7.0
769
+ tf_keras==2.15.1
770
+ tfp-nightly==0.24.0.dev0
771
+ thinc==8.2.3
772
+ threadpoolctl==3.2.0
773
+ tifffile==2023.12.9
774
+ timm==1.0.3
775
+ tinycss2==1.2.1
776
+ tobler==0.11.2
777
+ tokenizers==0.19.1
778
+ toml==0.10.2
779
+ tomli==2.0.1
780
+ tomlkit==0.12.5
781
+ toolz==0.12.1
782
+ torch==2.1.2
783
+ torchaudio==2.1.2
784
+ torchdata==0.7.1
785
+ torchinfo==1.8.0
786
+ torchmetrics==1.4.0.post0
787
+ torchtext==0.16.2
788
+ torchvision==0.16.2
789
+ tornado==6.3.3
790
+ tqdm==4.66.4
791
+ traceml==1.0.8
792
+ traitlets==5.9.0
793
+ traittypes==0.2.1
794
+ transformers==4.41.2
795
+ treelite==4.1.2
796
+ trl==0.9.6
797
+ truststore==0.8.0
798
+ trx-python==0.2.9
799
+ tsfresh==0.20.2
800
+ typeguard==4.1.5
801
+ typer==0.9.0
802
+ typer==0.9.4
803
+ types-python-dateutil==2.8.19.20240106
804
+ typing-inspect==0.9.0
805
+ typing-utils==0.1.0
806
+ typing_extensions==4.9.0
807
+ tyro==0.8.5
808
+ tzdata==2023.4
809
+ tzdata==2024.1
810
+ uc-micro-py==1.0.3
811
+ ucx-py==0.37.0
812
+ ujson==5.10.0
813
+ umap-learn==0.5.6
814
+ unicodedata2==15.1.0
815
+ update-checker==0.18.0
816
+ uri-template==1.3.0
817
+ uritemplate==3.0.1
818
+ urllib3==1.26.18
819
+ urllib3==2.1.0
820
+ urwid==2.6.12
821
+ urwid_readline==0.14
822
+ uvicorn==0.25.0
823
+ uvloop==0.19.0
824
+ vaex-astro==0.9.3
825
+ vaex-core==4.17.1
826
+ vaex-hdf5==0.14.1
827
+ vaex-jupyter==0.8.2
828
+ vaex-ml==0.18.3
829
+ vaex-server==0.9.0
830
+ vaex-viz==0.5.4
831
+ vaex==4.17.0
832
+ vec_noise==1.1.4
833
+ vecstack==0.4.0
834
+ virtualenv==20.21.0
835
+ visions==0.7.5
836
+ vowpalwabbit==9.9.0
837
+ vtk==9.3.0
838
+ wandb==0.17.0
839
+ wasabi==1.1.2
840
+ watchfiles==0.21.0
841
+ wavio==0.0.9
842
+ wcwidth==0.2.13
843
+ weasel==0.3.4
844
+ webcolors==1.13
845
+ webencodings==0.5.1
846
+ websocket-client==1.7.0
847
+ websockets==12.0
848
+ wfdb==4.1.2
849
+ whatthepatch==1.0.5
850
+ wheel==0.42.0
851
+ widgetsnbextension==3.6.6
852
+ witwidget==1.8.1
853
+ woodwork==0.31.0
854
+ wordcloud==1.9.3
855
+ wordsegment==1.3.1
856
+ wrapt==1.14.1
857
+ xarray-einstats==0.7.0
858
+ xarray==2024.5.0
859
+ xgboost==2.0.3
860
+ xvfbwrapper==0.2.9
861
+ xxhash==3.4.1
862
+ xyzservices==2024.4.0
863
+ y-py==0.6.2
864
+ yapf==0.40.2
865
+ yarl==1.9.3
866
+ yarl==1.9.4
867
+ ydata-profiling==4.6.4
868
+ yellowbrick==1.5
869
+ ypy-websocket==0.8.4
870
+ zict==3.0.0
871
+ zipp==3.17.0
872
+ zstandard==0.19.0
wandb/run-20240721_142400-5uukvmi3/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.154+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-07-21T14:24:01.917969",
5
+ "startedAt": "2024-07-21T14:24:00.530778",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "690161aed707",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.178,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.178,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.178,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.178,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.178,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5772.35152053833
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357574462890625
61
+ }
62
+ }
wandb/run-20240721_142400-5uukvmi3/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.3331, "train/grad_norm": 35.485958099365234, "train/learning_rate": 4.8996138996139e-07, "train/epoch": 7.0, "train/global_step": 4144, "_timestamp": 1721598405.8942447, "_runtime": 26565.35582470894, "_step": 16, "train_runtime": 26589.9478, "train_samples_per_second": 1.246, "train_steps_per_second": 0.156, "total_flos": 0.0, "train_loss": 0.4988581812059557}
wandb/run-20240721_142400-5uukvmi3/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240721_142400-5uukvmi3/logs/debug.log ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-07-21 14:24:00,532 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240721_142400-5uukvmi3/logs/debug.log
11
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240721_142400-5uukvmi3/logs/debug-internal.log
12
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7cbeb281b040>
13
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
14
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():610] starting backend
17
+ 2024-07-21 14:24:00,533 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
18
+ 2024-07-21 14:24:00,535 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-07-21 14:24:00,538 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
20
+ 2024-07-21 14:24:00,551 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
+ 2024-07-21 14:24:01,312 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
22
+ 2024-07-21 14:24:01,315 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
+ 2024-07-21 14:24:01,795 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
24
+ 2024-07-21 14:24:01,880 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-07-21 14:24:01,881 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
27
+ 2024-07-21 14:24:17,909 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
28
+ 2024-07-21 14:24:17,910 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
+ 2024-07-21 14:24:17,910 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
+ 2024-07-21 14:24:17,910 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
31
+ 2024-07-21 14:24:17,915 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
32
+ 2024-07-21 14:24:17,920 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 2048, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['OPTForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': '</s>', 'bos_token_id': 2, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/opt-350m', 'transformers_version': '4.41.2', 'activation_dropout': 0.0, 'model_type': 'opt', 'vocab_size': 50272, 'max_position_embeddings': 2048, 'num_attention_heads': 16, 'word_embed_proj_dim': 512, 'ffn_dim': 4096, 'hidden_size': 1024, 'num_hidden_layers': 24, 'dropout': 0.1, 'attention_dropout': 0.0, 'activation_function': 'relu', 'init_std': 0.02, 'layerdrop': 0.0, 'use_cache': True, 'do_layer_norm_before': False, 'enable_bias': True, 'layer_norm_elementwise_affine': True, '_remove_final_layer_norm': False, 'output_dir': './', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1.41e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 7, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Jul21_14-23-35_690161aed707', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 250, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'value-reward-model-opt-350m', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'no', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
+ 2024-07-21 21:46:45,899 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-07-21 21:46:45,900 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
35
+ 2024-07-21 21:50:24,607 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
wandb/run-20240721_142400-5uukvmi3/run-5uukvmi3.wandb ADDED
Binary file (602 kB). View file