yassineafr commited on
Commit
38a97c5
1 Parent(s): 7dfb9bc

jaisBdarija

Browse files
Files changed (27) hide show
  1. README.md +55 -0
  2. adapter_config.json +28 -0
  3. adapter_model.safetensors +3 -0
  4. runs/May22_11-21-59_2c1b614ec68f/events.out.tfevents.1716376951.2c1b614ec68f.34.0 +3 -0
  5. runs/May22_11-21-59_2c1b614ec68f/events.out.tfevents.1716377045.2c1b614ec68f.34.1 +3 -0
  6. runs/May22_11-33-56_2c1b614ec68f/events.out.tfevents.1716377651.2c1b614ec68f.217.0 +3 -0
  7. training_args.bin +3 -0
  8. wandb/debug-internal.log +0 -0
  9. wandb/debug.log +54 -0
  10. wandb/run-20240522_112259-4b714brj/files/conda-environment.yaml +0 -0
  11. wandb/run-20240522_112259-4b714brj/files/config.yaml +737 -0
  12. wandb/run-20240522_112259-4b714brj/files/output.log +3 -0
  13. wandb/run-20240522_112259-4b714brj/files/requirements.txt +878 -0
  14. wandb/run-20240522_112259-4b714brj/files/wandb-metadata.json +62 -0
  15. wandb/run-20240522_112259-4b714brj/files/wandb-summary.json +1 -0
  16. wandb/run-20240522_112259-4b714brj/logs/debug-internal.log +308 -0
  17. wandb/run-20240522_112259-4b714brj/logs/debug.log +48 -0
  18. wandb/run-20240522_112259-4b714brj/run-4b714brj.wandb +0 -0
  19. wandb/run-20240522_113413-8mudzhjp/files/conda-environment.yaml +0 -0
  20. wandb/run-20240522_113413-8mudzhjp/files/config.yaml +754 -0
  21. wandb/run-20240522_113413-8mudzhjp/files/output.log +93 -0
  22. wandb/run-20240522_113413-8mudzhjp/files/requirements.txt +878 -0
  23. wandb/run-20240522_113413-8mudzhjp/files/wandb-metadata.json +62 -0
  24. wandb/run-20240522_113413-8mudzhjp/files/wandb-summary.json +1 -0
  25. wandb/run-20240522_113413-8mudzhjp/logs/debug-internal.log +0 -0
  26. wandb/run-20240522_113413-8mudzhjp/logs/debug.log +54 -0
  27. wandb/run-20240522_113413-8mudzhjp/run-8mudzhjp.wandb +0 -0
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - generated_from_trainer
6
+ base_model: core42/jais-13b
7
+ model-index:
8
+ - name: working
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/yassine-af/huggingface/runs/8mudzhjp)
16
+ # working
17
+
18
+ This model is a fine-tuned version of [core42/jais-13b](https://huggingface.co/core42/jais-13b) on an unknown dataset.
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 0.0002
38
+ - train_batch_size: 8
39
+ - eval_batch_size: 8
40
+ - seed: 42
41
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
+ - lr_scheduler_type: linear
43
+ - num_epochs: 2
44
+
45
+ ### Training results
46
+
47
+
48
+
49
+ ### Framework versions
50
+
51
+ - PEFT 0.11.1
52
+ - Transformers 4.41.0
53
+ - Pytorch 2.3.0+cu121
54
+ - Datasets 2.18.0
55
+ - Tokenizers 0.19.1
adapter_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "core42/jais-13b",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 16,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "c_attn"
24
+ ],
25
+ "task_type": "CAUSAL_LM",
26
+ "use_dora": false,
27
+ "use_rslora": false
28
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06eb94a68730c3df8c86971da90ff7ccdb9675cacf17d64680c2d56ad19ca84b
3
+ size 52439304
runs/May22_11-21-59_2c1b614ec68f/events.out.tfevents.1716376951.2c1b614ec68f.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77aac7af4f1322783dc01d0658d7058e73d34e75853655c8a7e4058af5ec5883
3
+ size 4184
runs/May22_11-21-59_2c1b614ec68f/events.out.tfevents.1716377045.2c1b614ec68f.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7ce9b9bce6054bed83d2e68701535d98230f470f01c89f9556a0294ba6b9182
3
+ size 4184
runs/May22_11-33-56_2c1b614ec68f/events.out.tfevents.1716377651.2c1b614ec68f.217.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba5e7576561ebc4f2e00618a7caa6fcca241fcf314ff5cb2df704c7d5bf07b77
3
+ size 71472
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84f44aec00f66301ea7ba9dc7fc89b45e0f501cd02ad5a27ad189e368306aa4a
3
+ size 5112
wandb/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-22 11:34:13,996 INFO MainThread:217 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
2
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Configure stats pid to 217
3
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240522_113413-8mudzhjp/logs/debug.log
10
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240522_113413-8mudzhjp/logs/debug-internal.log
11
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ef92390cee0>
12
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():561] calling init triggers
13
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():611] starting backend
16
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():615] setting up manager
17
+ 2024-05-22 11:34:14,000 INFO MainThread:217 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-05-22 11:34:14,002 INFO MainThread:217 [wandb_init.py:init():623] backend started and connected
19
+ 2024-05-22 11:34:14,016 INFO MainThread:217 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-05-22 11:34:14,540 INFO MainThread:217 [wandb_init.py:init():715] updated telemetry
21
+ 2024-05-22 11:34:14,544 INFO MainThread:217 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
22
+ 2024-05-22 11:34:14,778 INFO MainThread:217 [wandb_run.py:_on_init():2357] communicating current version
23
+ 2024-05-22 11:34:14,843 INFO MainThread:217 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-05-22 11:34:14,844 INFO MainThread:217 [wandb_init.py:init():799] starting run threads in backend
26
+ 2024-05-22 11:34:30,856 INFO MainThread:217 [wandb_run.py:_console_start():2335] atexit reg
27
+ 2024-05-22 11:34:30,857 INFO MainThread:217 [wandb_run.py:_redirect():2190] redirect: wrap_raw
28
+ 2024-05-22 11:34:30,857 INFO MainThread:217 [wandb_run.py:_redirect():2255] Wrapping output streams.
29
+ 2024-05-22 11:34:30,857 INFO MainThread:217 [wandb_run.py:_redirect():2280] Redirects installed.
30
+ 2024-05-22 11:34:30,858 INFO MainThread:217 [wandb_init.py:init():842] run started, returning control to user process
31
+ 2024-05-22 11:34:30,865 INFO MainThread:217 [wandb_run.py:_config_callback():1347] config_cb None None {'peft_config': {'default': {'peft_type': <PeftType.LORA: 'LORA'>, 'auto_mapping': None, 'base_model_name_or_path': 'core42/jais-13b', 'revision': None, 'task_type': 'CAUSAL_LM', 'inference_mode': False, 'r': 16, 'target_modules': {'c_attn'}, 'lora_alpha': 32, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': None, 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'loftq_config': {}, 'use_dora': False, 'layer_replication': None}}, 'vocab_size': 84992, 'n_positions': 2048, 'n_embd': 5120, 'n_layer': 40, 'n_head': 40, 'n_inner': 13653, 'activation_function': 'swiglu', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 0, 'eos_token_id': 0, 'position_embedding_type': 'alibi', 'width_scale': 0.11100000000000002, 'embeddings_scale': 14.6, 'scale_qk_dot_by_d': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['JAISLMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 0, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'core42/jais-13b', 'transformers_version': '4.41.0', 'auto_map': {'AutoConfig': 'core42/jais-13b--configuration_jais.JAISConfig', 'AutoModel': 'core42/jais-13b--modeling_jais.JAISModel', 'AutoModelForCausalLM': 'core42/jais-13b--modeling_jais.JAISLMHeadModel', 'AutoModelForQuestionAnswering': 'core42/jais-13b--modeling_jais.JAISForQuestionAnswering', 'AutoModelForSequenceClassification': 'core42/jais-13b--modeling_jais.JAISForSequenceClassification', 'AutoModelForTokenClassification': 'core42/jais-13b--modeling_jais.JAISForTokenClassification'}, 'model_type': 'jais', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'bfloat16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May22_11-33-56_2c1b614ec68f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
32
+ 2024-05-22 11:34:30,875 INFO MainThread:217 [wandb_config.py:__setitem__():151] config set model/num_parameters = 13033919160 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7ef9227a9060>>
33
+ 2024-05-22 11:34:30,876 INFO MainThread:217 [wandb_run.py:_config_callback():1347] config_cb model/num_parameters 13033919160 None
34
+ 2024-05-22 14:04:41,874 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
35
+ 2024-05-22 14:04:41,875 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
36
+ 2024-05-22 14:14:52,958 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
37
+ 2024-05-22 14:14:54,437 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
38
+ 2024-05-22 14:14:54,437 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
39
+ 2024-05-22 14:15:26,186 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
40
+ 2024-05-22 14:16:25,347 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
41
+ 2024-05-22 14:16:25,348 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
42
+ 2024-05-22 14:16:29,691 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
43
+ 2024-05-22 14:16:44,749 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
44
+ 2024-05-22 14:16:44,749 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
45
+ 2024-05-22 14:23:14,136 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
46
+ 2024-05-22 14:23:16,353 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
47
+ 2024-05-22 14:23:16,353 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
48
+ 2024-05-22 14:26:18,732 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
49
+ 2024-05-22 14:26:19,623 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
50
+ 2024-05-22 14:26:19,624 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
51
+ 2024-05-22 14:34:00,493 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
52
+ 2024-05-22 14:34:00,984 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
53
+ 2024-05-22 14:34:00,984 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
54
+ 2024-05-22 14:34:16,410 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
wandb/run-20240522_112259-4b714brj/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240522_112259-4b714brj/files/config.yaml ADDED
@@ -0,0 +1,737 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.6
8
+ framework: huggingface
9
+ huggingface_version: 4.41.0
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1716376979.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ - 105
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 98
41
+ - 105
42
+ 3:
43
+ - 7
44
+ - 13
45
+ - 19
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.6
49
+ 6: 4.41.0
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ peft_config:
62
+ desc: null
63
+ value:
64
+ default:
65
+ peft_type: LORA
66
+ auto_mapping: null
67
+ base_model_name_or_path: core42/jais-13b
68
+ revision: null
69
+ task_type: CAUSAL_LM
70
+ inference_mode: false
71
+ r: 16
72
+ target_modules:
73
+ - c_attn
74
+ lora_alpha: 32
75
+ lora_dropout: 0.05
76
+ fan_in_fan_out: false
77
+ bias: none
78
+ use_rslora: false
79
+ modules_to_save: null
80
+ init_lora_weights: true
81
+ layers_to_transform: null
82
+ layers_pattern: null
83
+ rank_pattern: {}
84
+ alpha_pattern: {}
85
+ megatron_config: null
86
+ megatron_core: megatron.core
87
+ loftq_config: {}
88
+ use_dora: false
89
+ layer_replication: null
90
+ vocab_size:
91
+ desc: null
92
+ value: 84992
93
+ n_positions:
94
+ desc: null
95
+ value: 2048
96
+ n_embd:
97
+ desc: null
98
+ value: 5120
99
+ n_layer:
100
+ desc: null
101
+ value: 40
102
+ n_head:
103
+ desc: null
104
+ value: 40
105
+ n_inner:
106
+ desc: null
107
+ value: 13653
108
+ activation_function:
109
+ desc: null
110
+ value: swiglu
111
+ resid_pdrop:
112
+ desc: null
113
+ value: 0.0
114
+ embd_pdrop:
115
+ desc: null
116
+ value: 0.0
117
+ attn_pdrop:
118
+ desc: null
119
+ value: 0.0
120
+ layer_norm_epsilon:
121
+ desc: null
122
+ value: 1.0e-05
123
+ initializer_range:
124
+ desc: null
125
+ value: 0.02
126
+ scale_attn_weights:
127
+ desc: null
128
+ value: true
129
+ use_cache:
130
+ desc: null
131
+ value: false
132
+ scale_attn_by_inverse_layer_idx:
133
+ desc: null
134
+ value: false
135
+ reorder_and_upcast_attn:
136
+ desc: null
137
+ value: false
138
+ bos_token_id:
139
+ desc: null
140
+ value: 0
141
+ eos_token_id:
142
+ desc: null
143
+ value: 0
144
+ position_embedding_type:
145
+ desc: null
146
+ value: alibi
147
+ width_scale:
148
+ desc: null
149
+ value: 0.11100000000000002
150
+ embeddings_scale:
151
+ desc: null
152
+ value: 14.6
153
+ scale_qk_dot_by_d:
154
+ desc: null
155
+ value: true
156
+ return_dict:
157
+ desc: null
158
+ value: true
159
+ output_hidden_states:
160
+ desc: null
161
+ value: false
162
+ output_attentions:
163
+ desc: null
164
+ value: false
165
+ torchscript:
166
+ desc: null
167
+ value: false
168
+ torch_dtype:
169
+ desc: null
170
+ value: float32
171
+ use_bfloat16:
172
+ desc: null
173
+ value: false
174
+ tf_legacy_loss:
175
+ desc: null
176
+ value: false
177
+ pruned_heads:
178
+ desc: null
179
+ value: {}
180
+ tie_word_embeddings:
181
+ desc: null
182
+ value: true
183
+ chunk_size_feed_forward:
184
+ desc: null
185
+ value: 0
186
+ is_encoder_decoder:
187
+ desc: null
188
+ value: false
189
+ is_decoder:
190
+ desc: null
191
+ value: false
192
+ cross_attention_hidden_size:
193
+ desc: null
194
+ value: null
195
+ add_cross_attention:
196
+ desc: null
197
+ value: false
198
+ tie_encoder_decoder:
199
+ desc: null
200
+ value: false
201
+ max_length:
202
+ desc: null
203
+ value: 20
204
+ min_length:
205
+ desc: null
206
+ value: 0
207
+ do_sample:
208
+ desc: null
209
+ value: false
210
+ early_stopping:
211
+ desc: null
212
+ value: false
213
+ num_beams:
214
+ desc: null
215
+ value: 1
216
+ num_beam_groups:
217
+ desc: null
218
+ value: 1
219
+ diversity_penalty:
220
+ desc: null
221
+ value: 0.0
222
+ temperature:
223
+ desc: null
224
+ value: 1.0
225
+ top_k:
226
+ desc: null
227
+ value: 50
228
+ top_p:
229
+ desc: null
230
+ value: 1.0
231
+ typical_p:
232
+ desc: null
233
+ value: 1.0
234
+ repetition_penalty:
235
+ desc: null
236
+ value: 1.0
237
+ length_penalty:
238
+ desc: null
239
+ value: 1.0
240
+ no_repeat_ngram_size:
241
+ desc: null
242
+ value: 0
243
+ encoder_no_repeat_ngram_size:
244
+ desc: null
245
+ value: 0
246
+ bad_words_ids:
247
+ desc: null
248
+ value: null
249
+ num_return_sequences:
250
+ desc: null
251
+ value: 1
252
+ output_scores:
253
+ desc: null
254
+ value: false
255
+ return_dict_in_generate:
256
+ desc: null
257
+ value: false
258
+ forced_bos_token_id:
259
+ desc: null
260
+ value: null
261
+ forced_eos_token_id:
262
+ desc: null
263
+ value: null
264
+ remove_invalid_values:
265
+ desc: null
266
+ value: false
267
+ exponential_decay_length_penalty:
268
+ desc: null
269
+ value: null
270
+ suppress_tokens:
271
+ desc: null
272
+ value: null
273
+ begin_suppress_tokens:
274
+ desc: null
275
+ value: null
276
+ architectures:
277
+ desc: null
278
+ value:
279
+ - JAISLMHeadModel
280
+ finetuning_task:
281
+ desc: null
282
+ value: null
283
+ id2label:
284
+ desc: null
285
+ value:
286
+ '0': LABEL_0
287
+ '1': LABEL_1
288
+ label2id:
289
+ desc: null
290
+ value:
291
+ LABEL_0: 0
292
+ LABEL_1: 1
293
+ tokenizer_class:
294
+ desc: null
295
+ value: null
296
+ prefix:
297
+ desc: null
298
+ value: null
299
+ pad_token_id:
300
+ desc: null
301
+ value: 0
302
+ sep_token_id:
303
+ desc: null
304
+ value: null
305
+ decoder_start_token_id:
306
+ desc: null
307
+ value: null
308
+ task_specific_params:
309
+ desc: null
310
+ value: null
311
+ problem_type:
312
+ desc: null
313
+ value: null
314
+ _name_or_path:
315
+ desc: null
316
+ value: core42/jais-13b
317
+ transformers_version:
318
+ desc: null
319
+ value: 4.41.0
320
+ auto_map:
321
+ desc: null
322
+ value:
323
+ AutoConfig: core42/jais-13b--configuration_jais.JAISConfig
324
+ AutoModel: core42/jais-13b--modeling_jais.JAISModel
325
+ AutoModelForCausalLM: core42/jais-13b--modeling_jais.JAISLMHeadModel
326
+ AutoModelForQuestionAnswering: core42/jais-13b--modeling_jais.JAISForQuestionAnswering
327
+ AutoModelForSequenceClassification: core42/jais-13b--modeling_jais.JAISForSequenceClassification
328
+ AutoModelForTokenClassification: core42/jais-13b--modeling_jais.JAISForTokenClassification
329
+ model_type:
330
+ desc: null
331
+ value: jais
332
+ quantization_config:
333
+ desc: null
334
+ value:
335
+ quant_method: QuantizationMethod.BITS_AND_BYTES
336
+ _load_in_8bit: false
337
+ _load_in_4bit: true
338
+ llm_int8_threshold: 6.0
339
+ llm_int8_skip_modules: null
340
+ llm_int8_enable_fp32_cpu_offload: false
341
+ llm_int8_has_fp16_weight: false
342
+ bnb_4bit_quant_type: nf4
343
+ bnb_4bit_use_double_quant: false
344
+ bnb_4bit_compute_dtype: bfloat16
345
+ bnb_4bit_quant_storage: uint8
346
+ load_in_4bit: true
347
+ load_in_8bit: false
348
+ output_dir:
349
+ desc: null
350
+ value: /kaggle/working/
351
+ overwrite_output_dir:
352
+ desc: null
353
+ value: false
354
+ do_train:
355
+ desc: null
356
+ value: false
357
+ do_eval:
358
+ desc: null
359
+ value: false
360
+ do_predict:
361
+ desc: null
362
+ value: false
363
+ eval_strategy:
364
+ desc: null
365
+ value: 'no'
366
+ prediction_loss_only:
367
+ desc: null
368
+ value: false
369
+ per_device_train_batch_size:
370
+ desc: null
371
+ value: 8
372
+ per_device_eval_batch_size:
373
+ desc: null
374
+ value: 8
375
+ per_gpu_train_batch_size:
376
+ desc: null
377
+ value: null
378
+ per_gpu_eval_batch_size:
379
+ desc: null
380
+ value: null
381
+ gradient_accumulation_steps:
382
+ desc: null
383
+ value: 1
384
+ eval_accumulation_steps:
385
+ desc: null
386
+ value: null
387
+ eval_delay:
388
+ desc: null
389
+ value: 0
390
+ learning_rate:
391
+ desc: null
392
+ value: 0.0002
393
+ weight_decay:
394
+ desc: null
395
+ value: 0.0
396
+ adam_beta1:
397
+ desc: null
398
+ value: 0.9
399
+ adam_beta2:
400
+ desc: null
401
+ value: 0.999
402
+ adam_epsilon:
403
+ desc: null
404
+ value: 1.0e-08
405
+ max_grad_norm:
406
+ desc: null
407
+ value: 1.0
408
+ num_train_epochs:
409
+ desc: null
410
+ value: 2
411
+ max_steps:
412
+ desc: null
413
+ value: -1
414
+ lr_scheduler_type:
415
+ desc: null
416
+ value: linear
417
+ lr_scheduler_kwargs:
418
+ desc: null
419
+ value: {}
420
+ warmup_ratio:
421
+ desc: null
422
+ value: 0.0
423
+ warmup_steps:
424
+ desc: null
425
+ value: 0
426
+ log_level:
427
+ desc: null
428
+ value: passive
429
+ log_level_replica:
430
+ desc: null
431
+ value: warning
432
+ log_on_each_node:
433
+ desc: null
434
+ value: true
435
+ logging_dir:
436
+ desc: null
437
+ value: /kaggle/working/runs/May22_11-21-59_2c1b614ec68f
438
+ logging_strategy:
439
+ desc: null
440
+ value: steps
441
+ logging_first_step:
442
+ desc: null
443
+ value: false
444
+ logging_steps:
445
+ desc: null
446
+ value: 10
447
+ logging_nan_inf_filter:
448
+ desc: null
449
+ value: true
450
+ save_strategy:
451
+ desc: null
452
+ value: epoch
453
+ save_steps:
454
+ desc: null
455
+ value: 500
456
+ save_total_limit:
457
+ desc: null
458
+ value: 4
459
+ save_safetensors:
460
+ desc: null
461
+ value: true
462
+ save_on_each_node:
463
+ desc: null
464
+ value: false
465
+ save_only_model:
466
+ desc: null
467
+ value: false
468
+ restore_callback_states_from_checkpoint:
469
+ desc: null
470
+ value: false
471
+ no_cuda:
472
+ desc: null
473
+ value: false
474
+ use_cpu:
475
+ desc: null
476
+ value: false
477
+ use_mps_device:
478
+ desc: null
479
+ value: false
480
+ seed:
481
+ desc: null
482
+ value: 42
483
+ data_seed:
484
+ desc: null
485
+ value: null
486
+ jit_mode_eval:
487
+ desc: null
488
+ value: false
489
+ use_ipex:
490
+ desc: null
491
+ value: false
492
+ bf16:
493
+ desc: null
494
+ value: true
495
+ fp16:
496
+ desc: null
497
+ value: false
498
+ fp16_opt_level:
499
+ desc: null
500
+ value: O1
501
+ half_precision_backend:
502
+ desc: null
503
+ value: auto
504
+ bf16_full_eval:
505
+ desc: null
506
+ value: false
507
+ fp16_full_eval:
508
+ desc: null
509
+ value: false
510
+ tf32:
511
+ desc: null
512
+ value: null
513
+ local_rank:
514
+ desc: null
515
+ value: 0
516
+ ddp_backend:
517
+ desc: null
518
+ value: null
519
+ tpu_num_cores:
520
+ desc: null
521
+ value: null
522
+ tpu_metrics_debug:
523
+ desc: null
524
+ value: false
525
+ debug:
526
+ desc: null
527
+ value: []
528
+ dataloader_drop_last:
529
+ desc: null
530
+ value: false
531
+ eval_steps:
532
+ desc: null
533
+ value: null
534
+ dataloader_num_workers:
535
+ desc: null
536
+ value: 0
537
+ dataloader_prefetch_factor:
538
+ desc: null
539
+ value: null
540
+ past_index:
541
+ desc: null
542
+ value: -1
543
+ run_name:
544
+ desc: null
545
+ value: /kaggle/working/
546
+ disable_tqdm:
547
+ desc: null
548
+ value: false
549
+ remove_unused_columns:
550
+ desc: null
551
+ value: true
552
+ label_names:
553
+ desc: null
554
+ value: null
555
+ load_best_model_at_end:
556
+ desc: null
557
+ value: false
558
+ metric_for_best_model:
559
+ desc: null
560
+ value: null
561
+ greater_is_better:
562
+ desc: null
563
+ value: null
564
+ ignore_data_skip:
565
+ desc: null
566
+ value: false
567
+ fsdp:
568
+ desc: null
569
+ value: []
570
+ fsdp_min_num_params:
571
+ desc: null
572
+ value: 0
573
+ fsdp_config:
574
+ desc: null
575
+ value:
576
+ min_num_params: 0
577
+ xla: false
578
+ xla_fsdp_v2: false
579
+ xla_fsdp_grad_ckpt: false
580
+ fsdp_transformer_layer_cls_to_wrap:
581
+ desc: null
582
+ value: null
583
+ accelerator_config:
584
+ desc: null
585
+ value:
586
+ split_batches: false
587
+ dispatch_batches: null
588
+ even_batches: true
589
+ use_seedable_sampler: true
590
+ non_blocking: false
591
+ gradient_accumulation_kwargs: null
592
+ deepspeed:
593
+ desc: null
594
+ value: null
595
+ label_smoothing_factor:
596
+ desc: null
597
+ value: 0.0
598
+ optim:
599
+ desc: null
600
+ value: adamw_torch
601
+ optim_args:
602
+ desc: null
603
+ value: null
604
+ adafactor:
605
+ desc: null
606
+ value: false
607
+ group_by_length:
608
+ desc: null
609
+ value: false
610
+ length_column_name:
611
+ desc: null
612
+ value: length
613
+ report_to:
614
+ desc: null
615
+ value:
616
+ - tensorboard
617
+ - wandb
618
+ ddp_find_unused_parameters:
619
+ desc: null
620
+ value: null
621
+ ddp_bucket_cap_mb:
622
+ desc: null
623
+ value: null
624
+ ddp_broadcast_buffers:
625
+ desc: null
626
+ value: null
627
+ dataloader_pin_memory:
628
+ desc: null
629
+ value: true
630
+ dataloader_persistent_workers:
631
+ desc: null
632
+ value: false
633
+ skip_memory_metrics:
634
+ desc: null
635
+ value: true
636
+ use_legacy_prediction_loop:
637
+ desc: null
638
+ value: false
639
+ push_to_hub:
640
+ desc: null
641
+ value: false
642
+ resume_from_checkpoint:
643
+ desc: null
644
+ value: null
645
+ hub_model_id:
646
+ desc: null
647
+ value: null
648
+ hub_strategy:
649
+ desc: null
650
+ value: every_save
651
+ hub_token:
652
+ desc: null
653
+ value: <HUB_TOKEN>
654
+ hub_private_repo:
655
+ desc: null
656
+ value: false
657
+ hub_always_push:
658
+ desc: null
659
+ value: false
660
+ gradient_checkpointing:
661
+ desc: null
662
+ value: false
663
+ gradient_checkpointing_kwargs:
664
+ desc: null
665
+ value: null
666
+ include_inputs_for_metrics:
667
+ desc: null
668
+ value: false
669
+ eval_do_concat_batches:
670
+ desc: null
671
+ value: true
672
+ fp16_backend:
673
+ desc: null
674
+ value: auto
675
+ evaluation_strategy:
676
+ desc: null
677
+ value: null
678
+ push_to_hub_model_id:
679
+ desc: null
680
+ value: null
681
+ push_to_hub_organization:
682
+ desc: null
683
+ value: null
684
+ push_to_hub_token:
685
+ desc: null
686
+ value: <PUSH_TO_HUB_TOKEN>
687
+ mp_parameters:
688
+ desc: null
689
+ value: ''
690
+ auto_find_batch_size:
691
+ desc: null
692
+ value: true
693
+ full_determinism:
694
+ desc: null
695
+ value: false
696
+ torchdynamo:
697
+ desc: null
698
+ value: null
699
+ ray_scope:
700
+ desc: null
701
+ value: last
702
+ ddp_timeout:
703
+ desc: null
704
+ value: 1800
705
+ torch_compile:
706
+ desc: null
707
+ value: false
708
+ torch_compile_backend:
709
+ desc: null
710
+ value: null
711
+ torch_compile_mode:
712
+ desc: null
713
+ value: null
714
+ dispatch_batches:
715
+ desc: null
716
+ value: null
717
+ split_batches:
718
+ desc: null
719
+ value: null
720
+ include_tokens_per_second:
721
+ desc: null
722
+ value: false
723
+ include_num_input_tokens_seen:
724
+ desc: null
725
+ value: false
726
+ neftune_noise_alpha:
727
+ desc: null
728
+ value: null
729
+ optim_target_modules:
730
+ desc: null
731
+ value: null
732
+ batch_eval_metrics:
733
+ desc: null
734
+ value: false
735
+ model/num_parameters:
736
+ desc: null
737
+ value: 13033919160
wandb/run-20240522_112259-4b714brj/files/output.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
2
+ warnings.warn(
3
+ Overwriting /opt/conda/lib/python3.10/site-packages/torch/utils/data/_utils/fetch.py
wandb/run-20240522_112259-4b714brj/files/requirements.txt ADDED
@@ -0,0 +1,878 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.23.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.3
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.3
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.30.1
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.3
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.18.0
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.15.2.45.49
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.1
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.2
92
+ bokeh==3.4.1
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.69
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ contourpy==1.2.1
137
+ convertdate==2.4.0
138
+ crcmod==1.7
139
+ cryptography==41.0.7
140
+ cuda-python==12.4.0
141
+ cudf==23.8.0
142
+ cufflinks==0.17.3
143
+ cuml==23.8.0
144
+ cupy==13.0.0
145
+ cycler==0.12.1
146
+ cymem==2.0.8
147
+ cytoolz==0.12.3
148
+ daal4py==2024.3.0
149
+ daal==2024.3.0
150
+ dacite==1.8.1
151
+ dask-cuda==23.8.0
152
+ dask-cudf==23.8.0
153
+ dask-expr==1.0.11
154
+ dask==2024.4.1
155
+ dataclasses-json==0.6.4
156
+ dataproc_jupyter_plugin==0.1.66
157
+ datasets==2.18.0
158
+ datashader==0.16.0
159
+ datatile==1.0.3
160
+ db-dtypes==1.2.0
161
+ deap==1.4.1
162
+ debugpy==1.8.0
163
+ decorator==5.1.1
164
+ deepdiff==7.0.1
165
+ defusedxml==0.7.1
166
+ deprecation==2.1.0
167
+ descartes==1.1.0
168
+ dill==0.3.8
169
+ dipy==1.9.0
170
+ distlib==0.3.8
171
+ distributed==2023.7.1
172
+ distro==1.9.0
173
+ dm-tree==0.1.8
174
+ docker-pycreds==0.4.0
175
+ docker==7.0.0
176
+ docopt==0.6.2
177
+ docstring-parser==0.15
178
+ docstring-to-markdown==0.15
179
+ docutils==0.21.1
180
+ earthengine-api==0.1.399
181
+ easydict==1.13
182
+ easyocr==1.7.1
183
+ ecos==2.0.13
184
+ einops==0.8.0
185
+ eli5==0.13.0
186
+ emoji==2.11.0
187
+ en-core-web-lg==3.7.1
188
+ en-core-web-sm==3.7.1
189
+ entrypoints==0.4
190
+ ephem==4.1.5
191
+ esda==2.5.1
192
+ essentia==2.1b6.dev1110
193
+ et-xmlfile==1.1.0
194
+ etils==1.6.0
195
+ exceptiongroup==1.2.0
196
+ executing==2.0.1
197
+ explainable-ai-sdk==1.3.3
198
+ fastai==2.7.14
199
+ fastapi==0.108.0
200
+ fastavro==1.9.3
201
+ fastcore==1.5.29
202
+ fastdownload==0.0.7
203
+ fasteners==0.19
204
+ fastjsonschema==2.19.1
205
+ fastprogress==1.0.3
206
+ fastrlock==0.8.2
207
+ fasttext==0.9.2
208
+ feather-format==0.4.1
209
+ featuretools==1.30.0
210
+ filelock==3.13.1
211
+ fiona==1.9.6
212
+ fitter==1.7.0
213
+ flake8==7.0.0
214
+ flashtext==2.7
215
+ flatbuffers==23.5.26
216
+ flax==0.8.2
217
+ folium==0.16.0
218
+ fonttools==4.47.0
219
+ fonttools==4.51.0
220
+ fqdn==1.5.1
221
+ frozendict==2.4.2
222
+ frozenlist==1.4.1
223
+ fsspec==2024.2.0
224
+ fsspec==2024.3.1
225
+ funcy==2.0
226
+ fury==0.10.0
227
+ future==1.0.0
228
+ fuzzywuzzy==0.18.0
229
+ gast==0.5.4
230
+ gatspy==0.3
231
+ gcsfs==2024.2.0
232
+ gensim==4.3.2
233
+ geographiclib==2.0
234
+ geojson==3.1.0
235
+ geopandas==0.14.3
236
+ geoplot==0.5.1
237
+ geopy==2.4.1
238
+ geoviews==1.12.0
239
+ ggplot==0.11.5
240
+ giddy==2.3.5
241
+ gitdb==4.0.11
242
+ google-ai-generativelanguage==0.6.2
243
+ google-api-core==2.11.1
244
+ google-api-core==2.18.0
245
+ google-api-python-client==2.126.0
246
+ google-apitools==0.5.31
247
+ google-auth-httplib2==0.2.0
248
+ google-auth-oauthlib==1.2.0
249
+ google-auth==2.26.1
250
+ google-cloud-aiplatform==0.6.0a1
251
+ google-cloud-artifact-registry==1.10.0
252
+ google-cloud-automl==1.0.1
253
+ google-cloud-bigquery==2.34.4
254
+ google-cloud-bigtable==1.7.3
255
+ google-cloud-core==2.4.1
256
+ google-cloud-datastore==2.19.0
257
+ google-cloud-dlp==3.14.0
258
+ google-cloud-jupyter-config==0.0.5
259
+ google-cloud-language==2.13.3
260
+ google-cloud-monitoring==2.18.0
261
+ google-cloud-pubsub==2.19.0
262
+ google-cloud-pubsublite==1.9.0
263
+ google-cloud-recommendations-ai==0.7.1
264
+ google-cloud-resource-manager==1.11.0
265
+ google-cloud-spanner==3.40.1
266
+ google-cloud-storage==1.44.0
267
+ google-cloud-translate==3.12.1
268
+ google-cloud-videointelligence==2.13.3
269
+ google-cloud-vision==2.8.0
270
+ google-crc32c==1.5.0
271
+ google-generativeai==0.5.1
272
+ google-pasta==0.2.0
273
+ google-resumable-media==2.7.0
274
+ googleapis-common-protos==1.62.0
275
+ gplearn==0.4.2
276
+ gpustat==1.0.0
277
+ gpxpy==1.6.2
278
+ graphviz==0.20.3
279
+ greenlet==3.0.3
280
+ grpc-google-iam-v1==0.12.7
281
+ grpcio-status==1.48.1
282
+ grpcio-status==1.48.2
283
+ grpcio==1.51.1
284
+ grpcio==1.60.0
285
+ gviz-api==1.10.0
286
+ gym-notices==0.0.8
287
+ gym==0.26.2
288
+ gymnasium==0.29.0
289
+ h11==0.14.0
290
+ h2o==3.46.0.1
291
+ h5netcdf==1.3.0
292
+ h5py==3.10.0
293
+ haversine==2.8.1
294
+ hdfs==2.7.3
295
+ hep-ml==0.7.2
296
+ hijri-converter==2.3.1
297
+ hmmlearn==0.3.2
298
+ holidays==0.24
299
+ holoviews==1.18.3
300
+ hpsklearn==0.1.0
301
+ html5lib==1.1
302
+ htmlmin==0.1.12
303
+ httpcore==1.0.5
304
+ httplib2==0.21.0
305
+ httptools==0.6.1
306
+ httpx==0.27.0
307
+ huggingface-hub==0.23.1
308
+ hunspell==0.5.5
309
+ hydra-slayer==0.5.0
310
+ hyperopt==0.2.7
311
+ hypertools==0.8.0
312
+ idna==3.6
313
+ igraph==0.11.4
314
+ imagecodecs==2024.1.1
315
+ imageio==2.33.1
316
+ imbalanced-learn==0.12.2
317
+ imgaug==0.4.0
318
+ importlib-metadata==6.11.0
319
+ importlib-metadata==7.0.1
320
+ importlib-resources==6.1.1
321
+ inequality==1.0.1
322
+ iniconfig==2.0.0
323
+ ipydatawidgets==4.3.5
324
+ ipykernel==6.28.0
325
+ ipyleaflet==0.18.2
326
+ ipympl==0.7.0
327
+ ipython-genutils==0.2.0
328
+ ipython-genutils==0.2.0
329
+ ipython-sql==0.5.0
330
+ ipython==8.20.0
331
+ ipyvolume==0.6.3
332
+ ipyvue==1.11.0
333
+ ipyvuetify==1.9.4
334
+ ipywebrtc==0.6.0
335
+ ipywidgets==7.7.1
336
+ isoduration==20.11.0
337
+ isort==5.13.2
338
+ isoweek==1.3.3
339
+ itsdangerous==2.2.0
340
+ jaraco.classes==3.3.0
341
+ jax-jumpy==1.0.0
342
+ jax==0.4.23
343
+ jaxlib==0.4.23.dev20240116
344
+ jedi==0.19.1
345
+ jeepney==0.8.0
346
+ jieba==0.42.1
347
+ jmespath==1.0.1
348
+ joblib==1.4.0
349
+ json5==0.9.14
350
+ jsonpatch==1.33
351
+ jsonpointer==2.4
352
+ jsonschema-specifications==2023.12.1
353
+ jsonschema==4.20.0
354
+ jupyter-console==6.6.3
355
+ jupyter-events==0.9.0
356
+ jupyter-http-over-ws==0.0.8
357
+ jupyter-lsp==1.5.1
358
+ jupyter-server-mathjax==0.2.6
359
+ jupyter-ydoc==0.2.5
360
+ jupyter_client==7.4.9
361
+ jupyter_client==8.6.0
362
+ jupyter_core==5.7.1
363
+ jupyter_server==2.12.5
364
+ jupyter_server_fileid==0.9.1
365
+ jupyter_server_proxy==4.1.0
366
+ jupyter_server_terminals==0.5.1
367
+ jupyter_server_ydoc==0.8.0
368
+ jupyterlab-lsp==5.1.0
369
+ jupyterlab-widgets==3.0.9
370
+ jupyterlab==4.1.6
371
+ jupyterlab_git==0.44.0
372
+ jupyterlab_pygments==0.3.0
373
+ jupyterlab_server==2.25.2
374
+ jupytext==1.16.0
375
+ kaggle-environments==1.14.3
376
+ kaggle==1.6.12
377
+ kagglehub==0.2.3
378
+ keras-cv==0.8.2
379
+ keras-nlp==0.9.3
380
+ keras-tuner==1.4.6
381
+ keras==3.2.1
382
+ kernels-mixer==0.0.7
383
+ keyring==24.3.0
384
+ keyrings.google-artifactregistry-auth==1.1.2
385
+ kfp-pipeline-spec==0.2.2
386
+ kfp-server-api==2.0.5
387
+ kfp==2.5.0
388
+ kiwisolver==1.4.5
389
+ kmapper==2.0.1
390
+ kmodes==0.12.2
391
+ korean-lunar-calendar==0.3.1
392
+ kornia==0.7.2
393
+ kornia_rs==0.1.3
394
+ kt-legacy==1.0.5
395
+ kubernetes==26.1.0
396
+ langcodes==3.3.0
397
+ langid==1.1.6
398
+ lazy_loader==0.3
399
+ learntools==0.3.4
400
+ leven==1.0.4
401
+ libclang==16.0.6
402
+ libmambapy==1.5.0
403
+ libpysal==4.9.2
404
+ librosa==0.10.1
405
+ lightgbm==4.2.0
406
+ lightning-utilities==0.11.2
407
+ lime==0.2.0.1
408
+ line-profiler==4.1.2
409
+ linkify-it-py==2.0.3
410
+ llvmlite==0.41.1
411
+ llvmlite==0.42.0
412
+ lml==0.1.0
413
+ locket==1.0.0
414
+ loguru==0.7.2
415
+ loralib==0.1.2
416
+ lxml==5.2.1
417
+ lz4==4.3.3
418
+ mamba==1.5.0
419
+ mapclassify==2.6.1
420
+ markdown-it-py==3.0.0
421
+ marshmallow==3.21.1
422
+ matplotlib-inline==0.1.6
423
+ matplotlib-venn==0.11.10
424
+ matplotlib==3.7.5
425
+ matplotlib==3.8.4
426
+ mccabe==0.7.0
427
+ mdit-py-plugins==0.4.0
428
+ mdurl==0.1.2
429
+ memory-profiler==0.61.0
430
+ menuinst==2.0.1
431
+ mercantile==1.2.1
432
+ mgwr==2.2.1
433
+ missingno==0.5.2
434
+ mistune==0.8.4
435
+ mizani==0.11.1
436
+ ml-dtypes==0.2.0
437
+ mlcrate==0.2.0
438
+ mlens==0.2.3
439
+ mlxtend==0.23.1
440
+ mne==1.6.1
441
+ mnist==0.2.2
442
+ momepy==0.7.0
443
+ more-itertools==10.2.0
444
+ mpld3==0.5.10
445
+ mpmath==1.3.0
446
+ msgpack==1.0.7
447
+ multidict==6.0.4
448
+ multimethod==1.10
449
+ multipledispatch==1.0.0
450
+ multiprocess==0.70.16
451
+ munkres==1.1.4
452
+ murmurhash==1.0.10
453
+ mypy-extensions==1.0.0
454
+ namex==0.0.8
455
+ nb-conda-kernels==2.3.1
456
+ nb_conda==2.2.1
457
+ nbclassic==1.0.0
458
+ nbclient==0.5.13
459
+ nbconvert==6.4.5
460
+ nbdime==3.2.0
461
+ nbformat==5.9.2
462
+ ndindex==1.8
463
+ nest-asyncio==1.5.8
464
+ networkx==3.2.1
465
+ nibabel==5.2.1
466
+ nilearn==0.10.4
467
+ ninja==1.11.1.1
468
+ nltk==3.2.4
469
+ nose==1.3.7
470
+ notebook==6.5.4
471
+ notebook==6.5.6
472
+ notebook_executor==0.2
473
+ notebook_shim==0.2.3
474
+ numba==0.58.1
475
+ numba==0.59.1
476
+ numexpr==2.10.0
477
+ numpy==1.26.4
478
+ nvidia-cublas-cu12==12.1.3.1
479
+ nvidia-cuda-cupti-cu12==12.1.105
480
+ nvidia-cuda-nvrtc-cu12==12.1.105
481
+ nvidia-cuda-runtime-cu12==12.1.105
482
+ nvidia-cudnn-cu12==8.9.2.26
483
+ nvidia-cufft-cu12==11.0.2.54
484
+ nvidia-curand-cu12==10.3.2.106
485
+ nvidia-cusolver-cu12==11.4.5.107
486
+ nvidia-cusparse-cu12==12.1.0.106
487
+ nvidia-ml-py==11.495.46
488
+ nvidia-nccl-cu12==2.20.5
489
+ nvidia-nvjitlink-cu12==12.5.40
490
+ nvidia-nvtx-cu12==12.1.105
491
+ nvtx==0.2.10
492
+ oauth2client==4.1.3
493
+ oauthlib==3.2.2
494
+ objsize==0.6.1
495
+ odfpy==1.4.1
496
+ olefile==0.47
497
+ onnx==1.16.0
498
+ opencensus-context==0.1.3
499
+ opencensus==0.11.4
500
+ opencv-contrib-python==4.9.0.80
501
+ opencv-python-headless==4.9.0.80
502
+ opencv-python==4.9.0.80
503
+ openpyxl==3.1.2
504
+ openslide-python==1.3.1
505
+ opentelemetry-api==1.22.0
506
+ opentelemetry-exporter-otlp-proto-common==1.22.0
507
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
508
+ opentelemetry-exporter-otlp-proto-http==1.22.0
509
+ opentelemetry-exporter-otlp==1.22.0
510
+ opentelemetry-proto==1.22.0
511
+ opentelemetry-sdk==1.22.0
512
+ opentelemetry-semantic-conventions==0.43b0
513
+ opt-einsum==3.3.0
514
+ optax==0.2.2
515
+ optree==0.11.0
516
+ optuna==3.6.1
517
+ orbax-checkpoint==0.5.9
518
+ ordered-set==4.1.0
519
+ orjson==3.9.10
520
+ ortools==9.4.1874
521
+ osmnx==1.9.2
522
+ overrides==7.4.0
523
+ packaging==21.3
524
+ pandas-datareader==0.10.0
525
+ pandas-profiling==3.6.6
526
+ pandas-summary==0.2.0
527
+ pandas==2.1.4
528
+ pandas==2.2.2
529
+ pandasql==0.7.3
530
+ pandocfilters==1.5.0
531
+ panel==1.4.1
532
+ papermill==2.5.0
533
+ param==2.1.0
534
+ parso==0.8.3
535
+ partd==1.4.1
536
+ path.py==12.5.0
537
+ path==16.14.0
538
+ pathos==0.3.2
539
+ pathy==0.10.3
540
+ patsy==0.5.6
541
+ pdf2image==1.17.0
542
+ peft==0.11.1
543
+ pettingzoo==1.24.0
544
+ pexpect==4.8.0
545
+ pexpect==4.9.0
546
+ phik==0.12.4
547
+ pickleshare==0.7.5
548
+ pillow==10.3.0
549
+ pip==23.3.2
550
+ pkgutil_resolve_name==1.3.10
551
+ platformdirs==4.2.0
552
+ plotly-express==0.4.1
553
+ plotly==5.18.0
554
+ plotnine==0.13.4
555
+ pluggy==1.4.0
556
+ pointpats==2.4.0
557
+ polars==0.20.21
558
+ polyglot==16.7.4
559
+ pooch==1.8.1
560
+ pox==0.3.4
561
+ ppca==0.0.4
562
+ ppft==1.7.6.8
563
+ preprocessing==0.1.13
564
+ preshed==3.0.9
565
+ prettytable==3.9.0
566
+ progressbar2==4.4.2
567
+ prometheus-client==0.19.0
568
+ promise==2.3
569
+ prompt-toolkit==3.0.42
570
+ prompt-toolkit==3.0.43
571
+ prophet==1.1.1
572
+ proto-plus==1.23.0
573
+ protobuf==3.20.3
574
+ protobuf==4.21.12
575
+ psutil==5.9.3
576
+ psutil==5.9.7
577
+ ptyprocess==0.7.0
578
+ pudb==2024.1
579
+ pure-eval==0.2.2
580
+ py-cpuinfo==9.0.0
581
+ py-spy==0.3.14
582
+ py4j==0.10.9.7
583
+ pyLDAvis==3.4.1
584
+ pyOpenSSL==23.3.0
585
+ pyaml==23.12.0
586
+ pyarrow-hotfix==0.6
587
+ pyarrow==15.0.2
588
+ pyasn1-modules==0.3.0
589
+ pyasn1==0.5.1
590
+ pybind11==2.12.0
591
+ pyclipper==1.3.0.post5
592
+ pycodestyle==2.11.1
593
+ pycosat==0.6.6
594
+ pycparser==2.21
595
+ pycryptodome==3.20.0
596
+ pyct==0.5.0
597
+ pycuda==2024.1
598
+ pydantic==2.5.3
599
+ pydantic==2.7.0
600
+ pydantic_core==2.14.6
601
+ pydantic_core==2.18.1
602
+ pydegensac==0.1.2
603
+ pydicom==2.4.4
604
+ pydocstyle==6.3.0
605
+ pydot==1.4.2
606
+ pydub==0.25.1
607
+ pyemd==1.0.0
608
+ pyerfa==2.0.1.4
609
+ pyexcel-io==0.6.6
610
+ pyexcel-ods==0.6.0
611
+ pyflakes==3.2.0
612
+ pygltflib==1.16.2
613
+ pykalman==0.9.7
614
+ pylibraft==23.8.0
615
+ pylint==3.1.0
616
+ pymc3==3.11.4
617
+ pymongo==3.13.0
618
+ pynndescent==0.5.12
619
+ pynvml==11.4.1
620
+ pynvrtc==9.2
621
+ pyparsing==3.1.1
622
+ pyparsing==3.1.2
623
+ pypdf==4.2.0
624
+ pyproj==3.6.1
625
+ pysal==24.1
626
+ pyshp==2.3.1
627
+ pytesseract==0.3.10
628
+ pytest==8.1.1
629
+ python-bidi==0.4.2
630
+ python-dateutil==2.9.0.post0
631
+ python-dotenv==1.0.0
632
+ python-json-logger==2.0.7
633
+ python-louvain==0.16
634
+ python-lsp-jsonrpc==1.1.2
635
+ python-lsp-server==1.11.0
636
+ python-slugify==8.0.4
637
+ python-utils==3.8.2
638
+ pythreejs==2.4.2
639
+ pytoolconfig==1.3.1
640
+ pytools==2024.1.1
641
+ pytorch-ignite==0.5.0.post2
642
+ pytorch-lightning==2.2.2
643
+ pytz==2023.3.post1
644
+ pytz==2024.1
645
+ pyu2f==0.1.5
646
+ pyviz_comms==3.0.2
647
+ pyzmq==24.0.1
648
+ pyzmq==25.1.2
649
+ qgrid==1.3.1
650
+ qtconsole==5.5.1
651
+ quantecon==0.7.2
652
+ qudida==0.0.4
653
+ raft-dask==23.8.0
654
+ rasterio==1.3.10
655
+ rasterstats==0.19.0
656
+ ray-cpp==2.9.0
657
+ ray==2.9.0
658
+ referencing==0.32.1
659
+ regex==2023.12.25
660
+ requests-oauthlib==1.3.1
661
+ requests-toolbelt==0.10.1
662
+ requests==2.31.0
663
+ retrying==1.3.3
664
+ retrying==1.3.4
665
+ rfc3339-validator==0.1.4
666
+ rfc3986-validator==0.1.1
667
+ rgf-python==3.12.0
668
+ rich-click==1.7.4
669
+ rich==13.7.0
670
+ rich==13.7.1
671
+ rmm==23.8.0
672
+ rope==1.13.0
673
+ rpds-py==0.16.2
674
+ rsa==4.9
675
+ ruamel-yaml-conda==0.15.100
676
+ ruamel.yaml.clib==0.2.7
677
+ ruamel.yaml==0.17.40
678
+ s2sphere==0.2.5
679
+ s3fs==2024.2.0
680
+ s3transfer==0.6.2
681
+ safetensors==0.4.3
682
+ scattertext==0.1.19
683
+ scikit-image==0.22.0
684
+ scikit-learn-intelex==2024.3.0
685
+ scikit-learn==1.2.2
686
+ scikit-multilearn==0.2.0
687
+ scikit-optimize==0.10.1
688
+ scikit-plot==0.3.7
689
+ scikit-surprise==1.1.3
690
+ scipy==1.11.4
691
+ scipy==1.13.0
692
+ seaborn==0.12.2
693
+ segment_anything==1.0
694
+ segregation==2.5
695
+ semver==3.0.2
696
+ sentencepiece==0.2.0
697
+ sentry-sdk==1.45.0
698
+ setproctitle==1.3.3
699
+ setuptools-git==1.2
700
+ setuptools-scm==8.0.4
701
+ setuptools==69.0.3
702
+ shap==0.44.1
703
+ shapely==2.0.4
704
+ shellingham==1.5.4
705
+ simpervisor==1.0.0
706
+ simplejson==3.19.2
707
+ six==1.16.0
708
+ sklearn-pandas==2.2.0
709
+ slicer==0.0.7
710
+ smart-open==6.4.0
711
+ smmap==5.0.1
712
+ sniffio==1.3.0
713
+ snowballstemmer==2.2.0
714
+ snuggs==1.4.7
715
+ sortedcontainers==2.4.0
716
+ soundfile==0.12.1
717
+ soupsieve==2.5
718
+ soxr==0.3.7
719
+ spacy-legacy==3.0.12
720
+ spacy-loggers==1.0.5
721
+ spacy==3.7.3
722
+ spaghetti==1.7.5.post1
723
+ spectral==0.23.1
724
+ spglm==1.1.0
725
+ sphinx-rtd-theme==0.2.4
726
+ spint==1.0.7
727
+ splot==1.1.5.post1
728
+ spopt==0.6.0
729
+ spreg==1.4.2
730
+ spvcm==0.3.0
731
+ sqlparse==0.4.4
732
+ squarify==0.4.3
733
+ srsly==2.4.8
734
+ stable-baselines3==2.1.0
735
+ stack-data==0.6.2
736
+ stack-data==0.6.3
737
+ stanio==0.5.0
738
+ starlette==0.32.0.post1
739
+ statsmodels==0.14.1
740
+ stemming==1.0.1
741
+ stop-words==2018.7.23
742
+ stopit==1.1.2
743
+ stumpy==1.12.0
744
+ sympy==1.12
745
+ tables==3.9.2
746
+ tabulate==0.9.0
747
+ tangled-up-in-unicode==0.2.0
748
+ tbb==2021.12.0
749
+ tblib==3.0.0
750
+ tenacity==8.2.3
751
+ tensorboard-data-server==0.7.2
752
+ tensorboard-plugin-profile==2.15.0
753
+ tensorboard==2.15.1
754
+ tensorboardX==2.6.2.2
755
+ tensorflow-cloud==0.1.16
756
+ tensorflow-datasets==4.9.4
757
+ tensorflow-decision-forests==1.8.1
758
+ tensorflow-estimator==2.15.0
759
+ tensorflow-hub==0.16.1
760
+ tensorflow-io-gcs-filesystem==0.35.0
761
+ tensorflow-io==0.35.0
762
+ tensorflow-metadata==0.14.0
763
+ tensorflow-probability==0.23.0
764
+ tensorflow-serving-api==2.14.1
765
+ tensorflow-text==2.15.0
766
+ tensorflow-transform==0.14.0
767
+ tensorflow==2.15.0
768
+ tensorstore==0.1.56
769
+ termcolor==2.4.0
770
+ terminado==0.18.0
771
+ testpath==0.6.0
772
+ text-unidecode==1.3
773
+ textblob==0.18.0.post0
774
+ texttable==1.7.0
775
+ tf_keras==2.15.1
776
+ tfp-nightly==0.24.0.dev0
777
+ thinc==8.2.2
778
+ threadpoolctl==3.2.0
779
+ tifffile==2023.12.9
780
+ timm==0.9.16
781
+ tinycss2==1.2.1
782
+ tobler==0.11.2
783
+ tokenizers==0.19.1
784
+ toml==0.10.2
785
+ tomli==2.0.1
786
+ tomlkit==0.12.4
787
+ toolz==0.12.1
788
+ torch==2.3.0
789
+ torchaudio==2.1.2
790
+ torchdata==0.7.1
791
+ torchinfo==1.8.0
792
+ torchmetrics==1.3.2
793
+ torchtext==0.16.2
794
+ torchvision==0.16.2
795
+ tornado==6.3.3
796
+ tqdm==4.66.1
797
+ traceml==1.0.8
798
+ traitlets==5.9.0
799
+ traittypes==0.2.1
800
+ transformers==4.41.0
801
+ treelite-runtime==3.2.0
802
+ treelite==3.2.0
803
+ triton==2.3.0
804
+ truststore==0.8.0
805
+ trx-python==0.2.9
806
+ tsfresh==0.20.2
807
+ typeguard==4.1.5
808
+ typer==0.9.0
809
+ typer==0.9.4
810
+ types-python-dateutil==2.8.19.20240106
811
+ typing-inspect==0.9.0
812
+ typing-utils==0.1.0
813
+ typing_extensions==4.9.0
814
+ tzdata==2023.4
815
+ uc-micro-py==1.0.3
816
+ ucx-py==0.33.0
817
+ ujson==5.9.0
818
+ umap-learn==0.5.6
819
+ unicodedata2==15.1.0
820
+ update-checker==0.18.0
821
+ uri-template==1.3.0
822
+ uritemplate==3.0.1
823
+ urllib3==1.26.18
824
+ urllib3==2.1.0
825
+ urwid==2.6.10
826
+ urwid_readline==0.14
827
+ uvicorn==0.25.0
828
+ uvloop==0.19.0
829
+ vaex-astro==0.9.3
830
+ vaex-core==4.17.1
831
+ vaex-hdf5==0.14.1
832
+ vaex-jupyter==0.8.2
833
+ vaex-ml==0.18.3
834
+ vaex-server==0.9.0
835
+ vaex-viz==0.5.4
836
+ vaex==4.17.0
837
+ vec_noise==1.1.4
838
+ vecstack==0.4.0
839
+ virtualenv==20.21.0
840
+ visions==0.7.5
841
+ vowpalwabbit==9.9.0
842
+ vtk==9.3.0
843
+ wandb==0.16.6
844
+ wasabi==1.1.2
845
+ watchfiles==0.21.0
846
+ wavio==0.0.8
847
+ wcwidth==0.2.13
848
+ weasel==0.3.4
849
+ webcolors==1.13
850
+ webencodings==0.5.1
851
+ websocket-client==1.7.0
852
+ websockets==12.0
853
+ wfdb==4.1.2
854
+ whatthepatch==1.0.5
855
+ wheel==0.42.0
856
+ widgetsnbextension==3.6.6
857
+ witwidget==1.8.1
858
+ woodwork==0.30.0
859
+ wordcloud==1.9.3
860
+ wordsegment==1.3.1
861
+ wrapt==1.14.1
862
+ xarray-einstats==0.7.0
863
+ xarray==2024.3.0
864
+ xformers==0.0.26.post1
865
+ xgboost==2.0.3
866
+ xvfbwrapper==0.2.9
867
+ xxhash==3.4.1
868
+ xyzservices==2024.4.0
869
+ y-py==0.6.2
870
+ yapf==0.40.2
871
+ yarl==1.9.3
872
+ yarl==1.9.4
873
+ ydata-profiling==4.6.4
874
+ yellowbrick==1.5
875
+ ypy-websocket==0.8.4
876
+ zict==3.0.0
877
+ zipp==3.17.0
878
+ zstandard==0.22.0
wandb/run-20240522_112259-4b714brj/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-05-22T11:23:00.864653",
5
+ "startedAt": "2024-05-22T11:22:59.891144",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "2c1b614ec68f",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.144,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.144,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.144,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.144,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.144,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5656.321590423584
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357563018798828
61
+ }
62
+ }
wandb/run-20240522_112259-4b714brj/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 23}}
wandb/run-20240522_112259-4b714brj/logs/debug-internal.log ADDED
@@ -0,0 +1,308 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-22 11:22:59,898 INFO StreamThr :151 [internal.py:wandb_internal():86] W&B internal server running at pid: 151, started at: 2024-05-22 11:22:59.898127
2
+ 2024-05-22 11:22:59,900 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status
3
+ 2024-05-22 11:23:00,555 INFO WriterThread:151 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240522_112259-4b714brj/run-4b714brj.wandb
4
+ 2024-05-22 11:23:00,555 DEBUG SenderThread:151 [sender.py:send():379] send: header
5
+ 2024-05-22 11:23:00,561 DEBUG SenderThread:151 [sender.py:send():379] send: run
6
+ 2024-05-22 11:23:00,763 INFO SenderThread:151 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240522_112259-4b714brj/files
7
+ 2024-05-22 11:23:00,764 INFO SenderThread:151 [sender.py:_start_run_threads():1124] run started: 4b714brj with start time 1716376979.899342
8
+ 2024-05-22 11:23:00,767 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-05-22 11:23:00,767 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: check_version
10
+ 2024-05-22 11:23:00,842 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-05-22 11:23:00,854 DEBUG HandlerThread:151 [system_info.py:__init__():26] System info init
12
+ 2024-05-22 11:23:00,854 DEBUG HandlerThread:151 [system_info.py:__init__():41] System info init done
13
+ 2024-05-22 11:23:00,854 INFO HandlerThread:151 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-05-22 11:23:00,854 INFO SystemMonitor:151 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-05-22 11:23:00,854 INFO HandlerThread:151 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-05-22 11:23:00,855 INFO SystemMonitor:151 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-05-22 11:23:00,855 INFO SystemMonitor:151 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-05-22 11:23:00,856 INFO SystemMonitor:151 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-05-22 11:23:00,858 INFO SystemMonitor:151 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-05-22 11:23:00,858 INFO SystemMonitor:151 [interfaces.py:start():190] Started network monitoring
21
+ 2024-05-22 11:23:00,864 DEBUG HandlerThread:151 [system_info.py:probe():150] Probing system
22
+ 2024-05-22 11:23:00,867 DEBUG HandlerThread:151 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-05-22 11:23:00,867 DEBUG HandlerThread:151 [system_info.py:probe():198] Probing system done
24
+ 2024-05-22 11:23:00,867 DEBUG HandlerThread:151 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-05-22T11:23:00.864653', 'startedAt': '2024-05-22T11:22:59.891144', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '2c1b614ec68f', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.144, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.144, 'min': 0.0, 'max': 0.0}, {'current': 2000.144, 'min': 0.0, 'max': 0.0}, {'current': 2000.144, 'min': 0.0, 'max': 0.0}, {'current': 2000.144, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5656.321590423584}}, 'gpu': 'Tesla P100-PCIE-16GB', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla P100-PCIE-16GB', 'memory_total': 17179869184}], 'memory': {'total': 31.357563018798828}}
25
+ 2024-05-22 11:23:00,867 INFO HandlerThread:151 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-05-22 11:23:00,867 INFO HandlerThread:151 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-05-22 11:23:00,867 DEBUG HandlerThread:151 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-05-22 11:23:01,766 INFO Thread-12 :151 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/conda-environment.yaml
29
+ 2024-05-22 11:23:15,885 ERROR HandlerThread:151 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-05-22 11:23:15,887 DEBUG HandlerThread:151 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-05-22 11:23:15,888 INFO HandlerThread:151 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-05-22 11:23:15,896 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-05-22 11:23:15,896 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-05-22 11:23:15,896 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-05-22 11:23:15,896 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-05-22 11:23:15,896 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-05-22 11:23:15,896 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-05-22 11:23:15,897 DEBUG SenderThread:151 [sender.py:send():379] send: files
49
+ 2024-05-22 11:23:15,897 INFO SenderThread:151 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-05-22 11:23:16,228 INFO wandb-upload_0:151 [upload_job.py:push():131] Uploaded file /tmp/tmpbupdg2uiwandb/jmuw7ldk-wandb-metadata.json
51
+ 2024-05-22 11:23:16,769 INFO Thread-12 :151 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/wandb-metadata.json
52
+ 2024-05-22 11:23:16,882 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-05-22 11:23:16,882 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-05-22 11:23:16,885 DEBUG SenderThread:151 [sender.py:send():379] send: telemetry
55
+ 2024-05-22 11:23:16,896 DEBUG SenderThread:151 [sender.py:send():379] send: config
56
+ 2024-05-22 11:23:16,898 DEBUG SenderThread:151 [sender.py:send():379] send: metric
57
+ 2024-05-22 11:23:16,899 DEBUG SenderThread:151 [sender.py:send():379] send: telemetry
58
+ 2024-05-22 11:23:16,899 DEBUG SenderThread:151 [sender.py:send():379] send: metric
59
+ 2024-05-22 11:23:16,899 WARNING SenderThread:151 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-05-22 11:23:16,899 DEBUG SenderThread:151 [sender.py:send():379] send: telemetry
61
+ 2024-05-22 11:23:16,903 DEBUG SenderThread:151 [sender.py:send():379] send: telemetry
62
+ 2024-05-22 11:23:16,903 DEBUG SenderThread:151 [sender.py:send():379] send: config
63
+ 2024-05-22 11:23:16,904 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: stop_status
64
+ 2024-05-22 11:23:16,904 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: stop_status
65
+ 2024-05-22 11:23:16,907 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: internal_messages
66
+ 2024-05-22 11:23:17,769 INFO Thread-12 :151 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/output.log
67
+ 2024-05-22 11:23:17,770 INFO Thread-12 :151 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/requirements.txt
68
+ 2024-05-22 11:23:17,858 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: log_artifact
69
+ 2024-05-22 11:23:17,858 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: log_artifact
70
+ 2024-05-22 11:23:18,625 INFO wandb-upload_1:151 [upload_job.py:push():89] Uploaded file /tmp/tmpfgt5iaju/adapter_config.json
71
+ 2024-05-22 11:23:18,645 INFO wandb-upload_3:151 [upload_job.py:push():89] Uploaded file /tmp/tmpfgt5iaju/model_architecture.txt
72
+ 2024-05-22 11:23:18,661 INFO wandb-upload_0:151 [upload_job.py:push():89] Uploaded file /tmp/tmpfgt5iaju/README.md
73
+ 2024-05-22 11:23:19,067 INFO wandb-upload_2:151 [upload_job.py:push():89] Uploaded file /tmp/tmpfgt5iaju/adapter_model.safetensors
74
+ 2024-05-22 11:23:19,580 INFO SenderThread:151 [sender.py:send_request_log_artifact():1456] logged artifact model-4b714brj - {'id': 'QXJ0aWZhY3Q6ODQ2MDMyMjU0', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3Nzk2OTgzMA==', 'latestArtifact': None}}
75
+ 2024-05-22 11:23:19,770 INFO Thread-12 :151 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/output.log
76
+ 2024-05-22 11:23:20,547 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: pause
77
+ 2024-05-22 11:23:20,547 INFO HandlerThread:151 [handler.py:handle_request_pause():708] stopping system metrics thread
78
+ 2024-05-22 11:23:20,547 INFO HandlerThread:151 [system_monitor.py:finish():203] Stopping system monitor
79
+ 2024-05-22 11:23:20,548 DEBUG SystemMonitor:151 [system_monitor.py:_start():172] Starting system metrics aggregation loop
80
+ 2024-05-22 11:23:20,548 DEBUG SystemMonitor:151 [system_monitor.py:_start():179] Finished system metrics aggregation loop
81
+ 2024-05-22 11:23:20,548 DEBUG SystemMonitor:151 [system_monitor.py:_start():183] Publishing last batch of metrics
82
+ 2024-05-22 11:23:20,548 INFO HandlerThread:151 [interfaces.py:finish():202] Joined cpu monitor
83
+ 2024-05-22 11:23:20,549 INFO HandlerThread:151 [interfaces.py:finish():202] Joined disk monitor
84
+ 2024-05-22 11:23:20,555 INFO HandlerThread:151 [interfaces.py:finish():202] Joined gpu monitor
85
+ 2024-05-22 11:23:20,555 INFO HandlerThread:151 [interfaces.py:finish():202] Joined memory monitor
86
+ 2024-05-22 11:23:20,555 INFO HandlerThread:151 [interfaces.py:finish():202] Joined network monitor
87
+ 2024-05-22 11:23:21,581 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-05-22 11:23:26,583 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
89
+ 2024-05-22 11:23:31,589 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
90
+ 2024-05-22 11:23:31,775 INFO Thread-12 :151 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/config.yaml
91
+ 2024-05-22 11:23:31,883 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: stop_status
92
+ 2024-05-22 11:23:31,883 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: stop_status
93
+ 2024-05-22 11:23:31,886 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: internal_messages
94
+ 2024-05-22 11:23:36,999 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
95
+ 2024-05-22 11:23:42,000 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
96
+ 2024-05-22 11:23:45,679 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: resume
97
+ 2024-05-22 11:23:45,680 INFO HandlerThread:151 [handler.py:handle_request_resume():699] starting system metrics thread
98
+ 2024-05-22 11:23:45,680 INFO HandlerThread:151 [system_monitor.py:start():194] Starting system monitor
99
+ 2024-05-22 11:23:45,680 INFO SystemMonitor:151 [system_monitor.py:_start():158] Starting system asset monitoring threads
100
+ 2024-05-22 11:23:45,681 INFO SystemMonitor:151 [interfaces.py:start():190] Started cpu monitoring
101
+ 2024-05-22 11:23:45,681 INFO SystemMonitor:151 [interfaces.py:start():190] Started disk monitoring
102
+ 2024-05-22 11:23:45,682 INFO SystemMonitor:151 [interfaces.py:start():190] Started gpu monitoring
103
+ 2024-05-22 11:23:45,685 INFO SystemMonitor:151 [interfaces.py:start():190] Started memory monitoring
104
+ 2024-05-22 11:23:45,685 INFO SystemMonitor:151 [interfaces.py:start():190] Started network monitoring
105
+ 2024-05-22 11:23:45,720 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: pause
106
+ 2024-05-22 11:23:45,720 INFO HandlerThread:151 [handler.py:handle_request_pause():708] stopping system metrics thread
107
+ 2024-05-22 11:23:45,720 INFO HandlerThread:151 [system_monitor.py:finish():203] Stopping system monitor
108
+ 2024-05-22 11:23:45,721 DEBUG SystemMonitor:151 [system_monitor.py:_start():172] Starting system metrics aggregation loop
109
+ 2024-05-22 11:23:45,721 DEBUG SystemMonitor:151 [system_monitor.py:_start():179] Finished system metrics aggregation loop
110
+ 2024-05-22 11:23:45,721 DEBUG SystemMonitor:151 [system_monitor.py:_start():183] Publishing last batch of metrics
111
+ 2024-05-22 11:23:45,722 INFO HandlerThread:151 [interfaces.py:finish():202] Joined cpu monitor
112
+ 2024-05-22 11:23:45,722 INFO HandlerThread:151 [interfaces.py:finish():202] Joined disk monitor
113
+ 2024-05-22 11:23:45,727 INFO HandlerThread:151 [interfaces.py:finish():202] Joined gpu monitor
114
+ 2024-05-22 11:23:45,727 INFO HandlerThread:151 [interfaces.py:finish():202] Joined memory monitor
115
+ 2024-05-22 11:23:45,728 INFO HandlerThread:151 [interfaces.py:finish():202] Joined network monitor
116
+ 2024-05-22 11:23:45,728 DEBUG SenderThread:151 [sender.py:send():379] send: stats
117
+ 2024-05-22 11:23:46,883 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: stop_status
118
+ 2024-05-22 11:23:46,883 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: internal_messages
119
+ 2024-05-22 11:23:46,884 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: stop_status
120
+ 2024-05-22 11:23:47,780 INFO Thread-12 :151 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/output.log
121
+ 2024-05-22 11:23:47,995 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-05-22 11:23:52,144 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: resume
123
+ 2024-05-22 11:23:52,144 INFO HandlerThread:151 [handler.py:handle_request_resume():699] starting system metrics thread
124
+ 2024-05-22 11:23:52,144 INFO HandlerThread:151 [system_monitor.py:start():194] Starting system monitor
125
+ 2024-05-22 11:23:52,144 INFO SystemMonitor:151 [system_monitor.py:_start():158] Starting system asset monitoring threads
126
+ 2024-05-22 11:23:52,145 INFO SystemMonitor:151 [interfaces.py:start():190] Started cpu monitoring
127
+ 2024-05-22 11:23:52,145 INFO SystemMonitor:151 [interfaces.py:start():190] Started disk monitoring
128
+ 2024-05-22 11:23:52,147 INFO SystemMonitor:151 [interfaces.py:start():190] Started gpu monitoring
129
+ 2024-05-22 11:23:52,147 INFO SystemMonitor:151 [interfaces.py:start():190] Started memory monitoring
130
+ 2024-05-22 11:23:52,148 INFO SystemMonitor:151 [interfaces.py:start():190] Started network monitoring
131
+ 2024-05-22 11:23:52,185 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: pause
132
+ 2024-05-22 11:23:52,185 INFO HandlerThread:151 [handler.py:handle_request_pause():708] stopping system metrics thread
133
+ 2024-05-22 11:23:52,185 INFO HandlerThread:151 [system_monitor.py:finish():203] Stopping system monitor
134
+ 2024-05-22 11:23:52,185 INFO HandlerThread:151 [interfaces.py:finish():202] Joined cpu monitor
135
+ 2024-05-22 11:23:52,185 DEBUG SystemMonitor:151 [system_monitor.py:_start():172] Starting system metrics aggregation loop
136
+ 2024-05-22 11:23:52,186 INFO HandlerThread:151 [interfaces.py:finish():202] Joined disk monitor
137
+ 2024-05-22 11:23:52,186 DEBUG SystemMonitor:151 [system_monitor.py:_start():179] Finished system metrics aggregation loop
138
+ 2024-05-22 11:23:52,186 DEBUG SystemMonitor:151 [system_monitor.py:_start():183] Publishing last batch of metrics
139
+ 2024-05-22 11:23:52,191 INFO HandlerThread:151 [interfaces.py:finish():202] Joined gpu monitor
140
+ 2024-05-22 11:23:52,191 INFO HandlerThread:151 [interfaces.py:finish():202] Joined memory monitor
141
+ 2024-05-22 11:23:52,192 INFO HandlerThread:151 [interfaces.py:finish():202] Joined network monitor
142
+ 2024-05-22 11:23:52,192 DEBUG SenderThread:151 [sender.py:send():379] send: stats
143
+ 2024-05-22 11:23:53,193 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
144
+ 2024-05-22 11:23:58,194 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
145
+ 2024-05-22 11:24:01,883 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: stop_status
146
+ 2024-05-22 11:24:01,883 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: stop_status
147
+ 2024-05-22 11:24:01,885 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: internal_messages
148
+ 2024-05-22 11:24:02,823 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: resume
149
+ 2024-05-22 11:24:02,823 INFO HandlerThread:151 [handler.py:handle_request_resume():699] starting system metrics thread
150
+ 2024-05-22 11:24:02,823 INFO HandlerThread:151 [system_monitor.py:start():194] Starting system monitor
151
+ 2024-05-22 11:24:02,824 INFO SystemMonitor:151 [system_monitor.py:_start():158] Starting system asset monitoring threads
152
+ 2024-05-22 11:24:02,828 INFO SystemMonitor:151 [interfaces.py:start():190] Started cpu monitoring
153
+ 2024-05-22 11:24:02,829 INFO SystemMonitor:151 [interfaces.py:start():190] Started disk monitoring
154
+ 2024-05-22 11:24:02,831 INFO SystemMonitor:151 [interfaces.py:start():190] Started gpu monitoring
155
+ 2024-05-22 11:24:02,835 INFO SystemMonitor:151 [interfaces.py:start():190] Started memory monitoring
156
+ 2024-05-22 11:24:02,836 INFO SystemMonitor:151 [interfaces.py:start():190] Started network monitoring
157
+ 2024-05-22 11:24:04,031 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
158
+ 2024-05-22 11:24:05,597 DEBUG SenderThread:151 [sender.py:send():379] send: config
159
+ 2024-05-22 11:24:05,599 DEBUG SenderThread:151 [sender.py:send():379] send: metric
160
+ 2024-05-22 11:24:05,599 DEBUG SenderThread:151 [sender.py:send():379] send: metric
161
+ 2024-05-22 11:24:05,599 WARNING SenderThread:151 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
162
+ 2024-05-22 11:24:05,602 DEBUG SenderThread:151 [sender.py:send():379] send: config
163
+ 2024-05-22 11:24:06,395 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: log_artifact
164
+ 2024-05-22 11:24:06,395 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: log_artifact
165
+ 2024-05-22 11:24:06,657 INFO SenderThread:151 [sender.py:send_request_log_artifact():1456] logged artifact model-4b714brj - {'id': 'QXJ0aWZhY3Q6ODQ2MDMyMjU0', 'state': 'COMMITTED', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3Nzk2OTgzMA==', 'latestArtifact': {'id': 'QXJ0aWZhY3Q6ODQ2MDMyMjU0', 'versionIndex': 0}}}
166
+ 2024-05-22 11:24:06,748 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: pause
167
+ 2024-05-22 11:24:06,748 INFO HandlerThread:151 [handler.py:handle_request_pause():708] stopping system metrics thread
168
+ 2024-05-22 11:24:06,748 INFO HandlerThread:151 [system_monitor.py:finish():203] Stopping system monitor
169
+ 2024-05-22 11:24:06,748 DEBUG SystemMonitor:151 [system_monitor.py:_start():172] Starting system metrics aggregation loop
170
+ 2024-05-22 11:24:06,749 DEBUG SystemMonitor:151 [system_monitor.py:_start():179] Finished system metrics aggregation loop
171
+ 2024-05-22 11:24:06,749 DEBUG SystemMonitor:151 [system_monitor.py:_start():183] Publishing last batch of metrics
172
+ 2024-05-22 11:24:06,750 INFO HandlerThread:151 [interfaces.py:finish():202] Joined cpu monitor
173
+ 2024-05-22 11:24:06,750 INFO HandlerThread:151 [interfaces.py:finish():202] Joined disk monitor
174
+ 2024-05-22 11:24:06,755 INFO HandlerThread:151 [interfaces.py:finish():202] Joined gpu monitor
175
+ 2024-05-22 11:24:06,756 INFO HandlerThread:151 [interfaces.py:finish():202] Joined memory monitor
176
+ 2024-05-22 11:24:06,756 INFO HandlerThread:151 [interfaces.py:finish():202] Joined network monitor
177
+ 2024-05-22 11:24:06,756 DEBUG SenderThread:151 [sender.py:send():379] send: stats
178
+ 2024-05-22 11:24:09,757 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
179
+ 2024-05-22 11:24:14,758 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
180
+ 2024-05-22 11:24:16,883 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: stop_status
181
+ 2024-05-22 11:24:16,884 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: internal_messages
182
+ 2024-05-22 11:24:16,884 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: stop_status
183
+ 2024-05-22 11:24:20,045 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
184
+ 2024-05-22 11:24:21,664 DEBUG SenderThread:151 [sender.py:send():379] send: exit
185
+ 2024-05-22 11:24:21,664 INFO SenderThread:151 [sender.py:send_exit():586] handling exit code: 0
186
+ 2024-05-22 11:24:21,664 INFO SenderThread:151 [sender.py:send_exit():588] handling runtime: 23
187
+ 2024-05-22 11:24:21,666 INFO SenderThread:151 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
188
+ 2024-05-22 11:24:21,666 INFO SenderThread:151 [sender.py:send_exit():594] send defer
189
+ 2024-05-22 11:24:21,666 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
190
+ 2024-05-22 11:24:21,667 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 0
191
+ 2024-05-22 11:24:21,667 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
192
+ 2024-05-22 11:24:21,667 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 0
193
+ 2024-05-22 11:24:21,667 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 1
194
+ 2024-05-22 11:24:21,667 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
195
+ 2024-05-22 11:24:21,667 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 1
196
+ 2024-05-22 11:24:21,668 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
197
+ 2024-05-22 11:24:21,668 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 1
198
+ 2024-05-22 11:24:21,668 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 2
199
+ 2024-05-22 11:24:21,668 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
200
+ 2024-05-22 11:24:21,668 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 2
201
+ 2024-05-22 11:24:21,668 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
202
+ 2024-05-22 11:24:21,668 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 2
203
+ 2024-05-22 11:24:21,668 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 3
204
+ 2024-05-22 11:24:21,668 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
205
+ 2024-05-22 11:24:21,668 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 3
206
+ 2024-05-22 11:24:21,669 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
207
+ 2024-05-22 11:24:21,669 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 3
208
+ 2024-05-22 11:24:21,669 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 4
209
+ 2024-05-22 11:24:21,669 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
210
+ 2024-05-22 11:24:21,669 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 4
211
+ 2024-05-22 11:24:21,669 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
212
+ 2024-05-22 11:24:21,669 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 4
213
+ 2024-05-22 11:24:21,669 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 5
214
+ 2024-05-22 11:24:21,669 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
215
+ 2024-05-22 11:24:21,669 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 5
216
+ 2024-05-22 11:24:21,670 DEBUG SenderThread:151 [sender.py:send():379] send: summary
217
+ 2024-05-22 11:24:21,670 INFO SenderThread:151 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
218
+ 2024-05-22 11:24:21,670 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
219
+ 2024-05-22 11:24:21,670 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 5
220
+ 2024-05-22 11:24:21,670 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 6
221
+ 2024-05-22 11:24:21,670 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
222
+ 2024-05-22 11:24:21,670 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 6
223
+ 2024-05-22 11:24:21,671 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
224
+ 2024-05-22 11:24:21,671 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 6
225
+ 2024-05-22 11:24:21,676 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: status_report
226
+ 2024-05-22 11:24:21,793 INFO Thread-12 :151 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/wandb-summary.json
227
+ 2024-05-22 11:24:21,885 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 7
228
+ 2024-05-22 11:24:21,885 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
229
+ 2024-05-22 11:24:21,885 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 7
230
+ 2024-05-22 11:24:21,885 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
231
+ 2024-05-22 11:24:21,885 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 7
232
+ 2024-05-22 11:24:22,664 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: poll_exit
233
+ 2024-05-22 11:24:22,794 INFO Thread-12 :151 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/config.yaml
234
+ 2024-05-22 11:24:23,767 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 8
235
+ 2024-05-22 11:24:23,768 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: poll_exit
236
+ 2024-05-22 11:24:23,768 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
237
+ 2024-05-22 11:24:23,768 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 8
238
+ 2024-05-22 11:24:23,769 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
239
+ 2024-05-22 11:24:23,769 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 8
240
+ 2024-05-22 11:24:23,769 INFO SenderThread:151 [job_builder.py:build():318] Attempting to build job artifact
241
+ 2024-05-22 11:24:23,772 INFO SenderThread:151 [job_builder.py:_get_source_type():466] no source found
242
+ 2024-05-22 11:24:23,772 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 9
243
+ 2024-05-22 11:24:23,773 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
244
+ 2024-05-22 11:24:23,773 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 9
245
+ 2024-05-22 11:24:23,773 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
246
+ 2024-05-22 11:24:23,773 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 9
247
+ 2024-05-22 11:24:23,773 INFO SenderThread:151 [dir_watcher.py:finish():358] shutting down directory watcher
248
+ 2024-05-22 11:24:23,794 INFO Thread-12 :151 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/output.log
249
+ 2024-05-22 11:24:23,795 INFO SenderThread:151 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240522_112259-4b714brj/files
250
+ 2024-05-22 11:24:23,795 INFO SenderThread:151 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/output.log output.log
251
+ 2024-05-22 11:24:23,795 INFO SenderThread:151 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/config.yaml config.yaml
252
+ 2024-05-22 11:24:23,800 INFO SenderThread:151 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/wandb-metadata.json wandb-metadata.json
253
+ 2024-05-22 11:24:23,800 INFO SenderThread:151 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/conda-environment.yaml conda-environment.yaml
254
+ 2024-05-22 11:24:23,801 INFO SenderThread:151 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/wandb-summary.json wandb-summary.json
255
+ 2024-05-22 11:24:23,803 INFO SenderThread:151 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240522_112259-4b714brj/files/requirements.txt requirements.txt
256
+ 2024-05-22 11:24:23,807 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 10
257
+ 2024-05-22 11:24:23,807 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
258
+ 2024-05-22 11:24:23,807 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 10
259
+ 2024-05-22 11:24:23,811 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
260
+ 2024-05-22 11:24:23,811 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 10
261
+ 2024-05-22 11:24:23,811 INFO SenderThread:151 [file_pusher.py:finish():172] shutting down file pusher
262
+ 2024-05-22 11:24:24,020 INFO wandb-upload_1:151 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240522_112259-4b714brj/files/output.log
263
+ 2024-05-22 11:24:24,062 INFO wandb-upload_3:151 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240522_112259-4b714brj/files/wandb-summary.json
264
+ 2024-05-22 11:24:24,063 INFO wandb-upload_0:151 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240522_112259-4b714brj/files/config.yaml
265
+ 2024-05-22 11:24:24,065 INFO wandb-upload_2:151 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240522_112259-4b714brj/files/requirements.txt
266
+ 2024-05-22 11:24:24,265 INFO Thread-11 (_thread_body):151 [sender.py:transition_state():614] send defer: 11
267
+ 2024-05-22 11:24:24,266 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
268
+ 2024-05-22 11:24:24,266 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 11
269
+ 2024-05-22 11:24:24,267 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
270
+ 2024-05-22 11:24:24,267 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 11
271
+ 2024-05-22 11:24:24,267 INFO SenderThread:151 [file_pusher.py:join():178] waiting for file pusher
272
+ 2024-05-22 11:24:24,267 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 12
273
+ 2024-05-22 11:24:24,267 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
274
+ 2024-05-22 11:24:24,267 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 12
275
+ 2024-05-22 11:24:24,267 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
276
+ 2024-05-22 11:24:24,268 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 12
277
+ 2024-05-22 11:24:24,268 INFO SenderThread:151 [file_stream.py:finish():614] file stream finish called
278
+ 2024-05-22 11:24:24,475 INFO SenderThread:151 [file_stream.py:finish():618] file stream finish is done
279
+ 2024-05-22 11:24:24,475 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 13
280
+ 2024-05-22 11:24:24,475 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
281
+ 2024-05-22 11:24:24,475 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 13
282
+ 2024-05-22 11:24:24,475 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
283
+ 2024-05-22 11:24:24,476 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 13
284
+ 2024-05-22 11:24:24,476 INFO SenderThread:151 [sender.py:transition_state():614] send defer: 14
285
+ 2024-05-22 11:24:24,476 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: defer
286
+ 2024-05-22 11:24:24,476 INFO HandlerThread:151 [handler.py:handle_request_defer():172] handle defer: 14
287
+ 2024-05-22 11:24:24,476 DEBUG SenderThread:151 [sender.py:send():379] send: final
288
+ 2024-05-22 11:24:24,477 DEBUG SenderThread:151 [sender.py:send():379] send: footer
289
+ 2024-05-22 11:24:24,477 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: defer
290
+ 2024-05-22 11:24:24,477 INFO SenderThread:151 [sender.py:send_request_defer():610] handle sender defer: 14
291
+ 2024-05-22 11:24:24,478 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: poll_exit
292
+ 2024-05-22 11:24:24,479 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: poll_exit
293
+ 2024-05-22 11:24:24,479 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: poll_exit
294
+ 2024-05-22 11:24:24,479 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: poll_exit
295
+ 2024-05-22 11:24:24,480 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: server_info
296
+ 2024-05-22 11:24:24,480 DEBUG SenderThread:151 [sender.py:send_request():406] send_request: server_info
297
+ 2024-05-22 11:24:24,483 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: get_summary
298
+ 2024-05-22 11:24:24,484 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: sampled_history
299
+ 2024-05-22 11:24:24,484 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: internal_messages
300
+ 2024-05-22 11:24:24,548 INFO MainThread:151 [wandb_run.py:_footer_history_summary_info():3936] rendering history
301
+ 2024-05-22 11:24:24,548 INFO MainThread:151 [wandb_run.py:_footer_history_summary_info():3968] rendering summary
302
+ 2024-05-22 11:24:24,549 INFO MainThread:151 [wandb_run.py:_footer_sync_info():3895] logging synced files
303
+ 2024-05-22 11:24:24,549 DEBUG HandlerThread:151 [handler.py:handle_request():146] handle_request: shutdown
304
+ 2024-05-22 11:24:24,549 INFO HandlerThread:151 [handler.py:finish():866] shutting down handler
305
+ 2024-05-22 11:24:25,480 INFO WriterThread:151 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240522_112259-4b714brj/run-4b714brj.wandb
306
+ 2024-05-22 11:24:25,548 INFO SenderThread:151 [sender.py:finish():1546] shutting down sender
307
+ 2024-05-22 11:24:25,548 INFO SenderThread:151 [file_pusher.py:finish():172] shutting down file pusher
308
+ 2024-05-22 11:24:25,549 INFO SenderThread:151 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240522_112259-4b714brj/logs/debug.log ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
2
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240522_112259-4b714brj/logs/debug.log
11
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240522_112259-4b714brj/logs/debug-internal.log
12
+ 2024-05-22 11:22:59,893 INFO MainThread:34 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78d7123ff250>
13
+ 2024-05-22 11:22:59,894 INFO MainThread:34 [wandb_init.py:init():561] calling init triggers
14
+ 2024-05-22 11:22:59,894 INFO MainThread:34 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-05-22 11:22:59,894 INFO MainThread:34 [wandb_init.py:init():611] starting backend
17
+ 2024-05-22 11:22:59,894 INFO MainThread:34 [wandb_init.py:init():615] setting up manager
18
+ 2024-05-22 11:22:59,896 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-05-22 11:22:59,899 INFO MainThread:34 [wandb_init.py:init():623] backend started and connected
20
+ 2024-05-22 11:22:59,913 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-05-22 11:23:00,554 INFO MainThread:34 [wandb_init.py:init():715] updated telemetry
22
+ 2024-05-22 11:23:00,559 INFO MainThread:34 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
23
+ 2024-05-22 11:23:00,766 INFO MainThread:34 [wandb_run.py:_on_init():2357] communicating current version
24
+ 2024-05-22 11:23:00,835 INFO MainThread:34 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-05-22 11:23:00,836 INFO MainThread:34 [wandb_init.py:init():799] starting run threads in backend
27
+ 2024-05-22 11:23:16,882 INFO MainThread:34 [wandb_run.py:_console_start():2335] atexit reg
28
+ 2024-05-22 11:23:16,883 INFO MainThread:34 [wandb_run.py:_redirect():2190] redirect: wrap_raw
29
+ 2024-05-22 11:23:16,883 INFO MainThread:34 [wandb_run.py:_redirect():2255] Wrapping output streams.
30
+ 2024-05-22 11:23:16,883 INFO MainThread:34 [wandb_run.py:_redirect():2280] Redirects installed.
31
+ 2024-05-22 11:23:16,885 INFO MainThread:34 [wandb_init.py:init():842] run started, returning control to user process
32
+ 2024-05-22 11:23:16,891 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'peft_config': {'default': {'peft_type': <PeftType.LORA: 'LORA'>, 'auto_mapping': None, 'base_model_name_or_path': 'core42/jais-13b', 'revision': None, 'task_type': 'CAUSAL_LM', 'inference_mode': False, 'r': 16, 'target_modules': {'c_attn'}, 'lora_alpha': 32, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': None, 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'loftq_config': {}, 'use_dora': False, 'layer_replication': None}}, 'vocab_size': 84992, 'n_positions': 2048, 'n_embd': 5120, 'n_layer': 40, 'n_head': 40, 'n_inner': 13653, 'activation_function': 'swiglu', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 0, 'eos_token_id': 0, 'position_embedding_type': 'alibi', 'width_scale': 0.11100000000000002, 'embeddings_scale': 14.6, 'scale_qk_dot_by_d': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['JAISLMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 0, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'core42/jais-13b', 'transformers_version': '4.41.0', 'auto_map': {'AutoConfig': 'core42/jais-13b--configuration_jais.JAISConfig', 'AutoModel': 'core42/jais-13b--modeling_jais.JAISModel', 'AutoModelForCausalLM': 'core42/jais-13b--modeling_jais.JAISLMHeadModel', 'AutoModelForQuestionAnswering': 'core42/jais-13b--modeling_jais.JAISForQuestionAnswering', 'AutoModelForSequenceClassification': 'core42/jais-13b--modeling_jais.JAISForSequenceClassification', 'AutoModelForTokenClassification': 'core42/jais-13b--modeling_jais.JAISForTokenClassification'}, 'model_type': 'jais', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'bfloat16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May22_11-21-59_2c1b614ec68f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
+ 2024-05-22 11:23:16,902 INFO MainThread:34 [wandb_config.py:__setitem__():151] config set model/num_parameters = 13033919160 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x78d6d5269d80>>
34
+ 2024-05-22 11:23:16,903 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb model/num_parameters 13033919160 None
35
+ 2024-05-22 11:23:20,546 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-05-22 11:23:20,547 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
37
+ 2024-05-22 11:23:45,679 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
38
+ 2024-05-22 11:23:45,683 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
+ 2024-05-22 11:23:45,683 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
40
+ 2024-05-22 11:23:52,143 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
41
+ 2024-05-22 11:23:52,145 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
42
+ 2024-05-22 11:23:52,145 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
43
+ 2024-05-22 11:24:02,823 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
44
+ 2024-05-22 11:24:05,592 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'peft_config': {'default': {'peft_type': <PeftType.LORA: 'LORA'>, 'auto_mapping': None, 'base_model_name_or_path': 'core42/jais-13b', 'revision': None, 'task_type': 'CAUSAL_LM', 'inference_mode': False, 'r': 16, 'target_modules': {'c_attn'}, 'lora_alpha': 32, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': None, 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'loftq_config': {}, 'use_dora': False, 'layer_replication': None}}, 'vocab_size': 84992, 'n_positions': 2048, 'n_embd': 5120, 'n_layer': 40, 'n_head': 40, 'n_inner': 13653, 'activation_function': 'swiglu', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 0, 'eos_token_id': 0, 'position_embedding_type': 'alibi', 'width_scale': 0.11100000000000002, 'embeddings_scale': 14.6, 'scale_qk_dot_by_d': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['JAISLMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 0, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'core42/jais-13b', 'transformers_version': '4.41.0', 'auto_map': {'AutoConfig': 'core42/jais-13b--configuration_jais.JAISConfig', 'AutoModel': 'core42/jais-13b--modeling_jais.JAISModel', 'AutoModelForCausalLM': 'core42/jais-13b--modeling_jais.JAISLMHeadModel', 'AutoModelForQuestionAnswering': 'core42/jais-13b--modeling_jais.JAISForQuestionAnswering', 'AutoModelForSequenceClassification': 'core42/jais-13b--modeling_jais.JAISForSequenceClassification', 'AutoModelForTokenClassification': 'core42/jais-13b--modeling_jais.JAISForTokenClassification'}, 'model_type': 'jais', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'bfloat16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May22_11-21-59_2c1b614ec68f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
45
+ 2024-05-22 11:24:05,601 INFO MainThread:34 [wandb_config.py:__setitem__():151] config set model/num_parameters = 13033919160 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x78d6d5269d80>>
46
+ 2024-05-22 11:24:05,602 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb model/num_parameters 13033919160 None
47
+ 2024-05-22 11:24:06,747 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
48
+ 2024-05-22 11:24:06,748 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
wandb/run-20240522_112259-4b714brj/run-4b714brj.wandb ADDED
Binary file (18.9 kB). View file
 
wandb/run-20240522_113413-8mudzhjp/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240522_113413-8mudzhjp/files/config.yaml ADDED
@@ -0,0 +1,754 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.6
8
+ framework: huggingface
9
+ huggingface_version: 4.41.0
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1716377654.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ - 105
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 98
41
+ - 105
42
+ 3:
43
+ - 7
44
+ - 13
45
+ - 19
46
+ - 23
47
+ - 62
48
+ 4: 3.10.13
49
+ 5: 0.16.6
50
+ 6: 4.41.0
51
+ 8:
52
+ - 1
53
+ - 2
54
+ - 5
55
+ 9:
56
+ 1: transformers_trainer
57
+ 13: linux-x86_64
58
+ m:
59
+ - 1: train/global_step
60
+ 6:
61
+ - 3
62
+ - 1: train/loss
63
+ 5: 1
64
+ 6:
65
+ - 1
66
+ - 1: train/grad_norm
67
+ 5: 1
68
+ 6:
69
+ - 1
70
+ - 1: train/learning_rate
71
+ 5: 1
72
+ 6:
73
+ - 1
74
+ - 1: train/epoch
75
+ 5: 1
76
+ 6:
77
+ - 1
78
+ peft_config:
79
+ desc: null
80
+ value:
81
+ default:
82
+ peft_type: LORA
83
+ auto_mapping: null
84
+ base_model_name_or_path: core42/jais-13b
85
+ revision: null
86
+ task_type: CAUSAL_LM
87
+ inference_mode: false
88
+ r: 16
89
+ target_modules:
90
+ - c_attn
91
+ lora_alpha: 32
92
+ lora_dropout: 0.05
93
+ fan_in_fan_out: false
94
+ bias: none
95
+ use_rslora: false
96
+ modules_to_save: null
97
+ init_lora_weights: true
98
+ layers_to_transform: null
99
+ layers_pattern: null
100
+ rank_pattern: {}
101
+ alpha_pattern: {}
102
+ megatron_config: null
103
+ megatron_core: megatron.core
104
+ loftq_config: {}
105
+ use_dora: false
106
+ layer_replication: null
107
+ vocab_size:
108
+ desc: null
109
+ value: 84992
110
+ n_positions:
111
+ desc: null
112
+ value: 2048
113
+ n_embd:
114
+ desc: null
115
+ value: 5120
116
+ n_layer:
117
+ desc: null
118
+ value: 40
119
+ n_head:
120
+ desc: null
121
+ value: 40
122
+ n_inner:
123
+ desc: null
124
+ value: 13653
125
+ activation_function:
126
+ desc: null
127
+ value: swiglu
128
+ resid_pdrop:
129
+ desc: null
130
+ value: 0.0
131
+ embd_pdrop:
132
+ desc: null
133
+ value: 0.0
134
+ attn_pdrop:
135
+ desc: null
136
+ value: 0.0
137
+ layer_norm_epsilon:
138
+ desc: null
139
+ value: 1.0e-05
140
+ initializer_range:
141
+ desc: null
142
+ value: 0.02
143
+ scale_attn_weights:
144
+ desc: null
145
+ value: true
146
+ use_cache:
147
+ desc: null
148
+ value: false
149
+ scale_attn_by_inverse_layer_idx:
150
+ desc: null
151
+ value: false
152
+ reorder_and_upcast_attn:
153
+ desc: null
154
+ value: false
155
+ bos_token_id:
156
+ desc: null
157
+ value: 0
158
+ eos_token_id:
159
+ desc: null
160
+ value: 0
161
+ position_embedding_type:
162
+ desc: null
163
+ value: alibi
164
+ width_scale:
165
+ desc: null
166
+ value: 0.11100000000000002
167
+ embeddings_scale:
168
+ desc: null
169
+ value: 14.6
170
+ scale_qk_dot_by_d:
171
+ desc: null
172
+ value: true
173
+ return_dict:
174
+ desc: null
175
+ value: true
176
+ output_hidden_states:
177
+ desc: null
178
+ value: false
179
+ output_attentions:
180
+ desc: null
181
+ value: false
182
+ torchscript:
183
+ desc: null
184
+ value: false
185
+ torch_dtype:
186
+ desc: null
187
+ value: float32
188
+ use_bfloat16:
189
+ desc: null
190
+ value: false
191
+ tf_legacy_loss:
192
+ desc: null
193
+ value: false
194
+ pruned_heads:
195
+ desc: null
196
+ value: {}
197
+ tie_word_embeddings:
198
+ desc: null
199
+ value: true
200
+ chunk_size_feed_forward:
201
+ desc: null
202
+ value: 0
203
+ is_encoder_decoder:
204
+ desc: null
205
+ value: false
206
+ is_decoder:
207
+ desc: null
208
+ value: false
209
+ cross_attention_hidden_size:
210
+ desc: null
211
+ value: null
212
+ add_cross_attention:
213
+ desc: null
214
+ value: false
215
+ tie_encoder_decoder:
216
+ desc: null
217
+ value: false
218
+ max_length:
219
+ desc: null
220
+ value: 20
221
+ min_length:
222
+ desc: null
223
+ value: 0
224
+ do_sample:
225
+ desc: null
226
+ value: false
227
+ early_stopping:
228
+ desc: null
229
+ value: false
230
+ num_beams:
231
+ desc: null
232
+ value: 1
233
+ num_beam_groups:
234
+ desc: null
235
+ value: 1
236
+ diversity_penalty:
237
+ desc: null
238
+ value: 0.0
239
+ temperature:
240
+ desc: null
241
+ value: 1.0
242
+ top_k:
243
+ desc: null
244
+ value: 50
245
+ top_p:
246
+ desc: null
247
+ value: 1.0
248
+ typical_p:
249
+ desc: null
250
+ value: 1.0
251
+ repetition_penalty:
252
+ desc: null
253
+ value: 1.0
254
+ length_penalty:
255
+ desc: null
256
+ value: 1.0
257
+ no_repeat_ngram_size:
258
+ desc: null
259
+ value: 0
260
+ encoder_no_repeat_ngram_size:
261
+ desc: null
262
+ value: 0
263
+ bad_words_ids:
264
+ desc: null
265
+ value: null
266
+ num_return_sequences:
267
+ desc: null
268
+ value: 1
269
+ output_scores:
270
+ desc: null
271
+ value: false
272
+ return_dict_in_generate:
273
+ desc: null
274
+ value: false
275
+ forced_bos_token_id:
276
+ desc: null
277
+ value: null
278
+ forced_eos_token_id:
279
+ desc: null
280
+ value: null
281
+ remove_invalid_values:
282
+ desc: null
283
+ value: false
284
+ exponential_decay_length_penalty:
285
+ desc: null
286
+ value: null
287
+ suppress_tokens:
288
+ desc: null
289
+ value: null
290
+ begin_suppress_tokens:
291
+ desc: null
292
+ value: null
293
+ architectures:
294
+ desc: null
295
+ value:
296
+ - JAISLMHeadModel
297
+ finetuning_task:
298
+ desc: null
299
+ value: null
300
+ id2label:
301
+ desc: null
302
+ value:
303
+ '0': LABEL_0
304
+ '1': LABEL_1
305
+ label2id:
306
+ desc: null
307
+ value:
308
+ LABEL_0: 0
309
+ LABEL_1: 1
310
+ tokenizer_class:
311
+ desc: null
312
+ value: null
313
+ prefix:
314
+ desc: null
315
+ value: null
316
+ pad_token_id:
317
+ desc: null
318
+ value: 0
319
+ sep_token_id:
320
+ desc: null
321
+ value: null
322
+ decoder_start_token_id:
323
+ desc: null
324
+ value: null
325
+ task_specific_params:
326
+ desc: null
327
+ value: null
328
+ problem_type:
329
+ desc: null
330
+ value: null
331
+ _name_or_path:
332
+ desc: null
333
+ value: core42/jais-13b
334
+ transformers_version:
335
+ desc: null
336
+ value: 4.41.0
337
+ auto_map:
338
+ desc: null
339
+ value:
340
+ AutoConfig: core42/jais-13b--configuration_jais.JAISConfig
341
+ AutoModel: core42/jais-13b--modeling_jais.JAISModel
342
+ AutoModelForCausalLM: core42/jais-13b--modeling_jais.JAISLMHeadModel
343
+ AutoModelForQuestionAnswering: core42/jais-13b--modeling_jais.JAISForQuestionAnswering
344
+ AutoModelForSequenceClassification: core42/jais-13b--modeling_jais.JAISForSequenceClassification
345
+ AutoModelForTokenClassification: core42/jais-13b--modeling_jais.JAISForTokenClassification
346
+ model_type:
347
+ desc: null
348
+ value: jais
349
+ quantization_config:
350
+ desc: null
351
+ value:
352
+ quant_method: QuantizationMethod.BITS_AND_BYTES
353
+ _load_in_8bit: false
354
+ _load_in_4bit: true
355
+ llm_int8_threshold: 6.0
356
+ llm_int8_skip_modules: null
357
+ llm_int8_enable_fp32_cpu_offload: false
358
+ llm_int8_has_fp16_weight: false
359
+ bnb_4bit_quant_type: nf4
360
+ bnb_4bit_use_double_quant: false
361
+ bnb_4bit_compute_dtype: bfloat16
362
+ bnb_4bit_quant_storage: uint8
363
+ load_in_4bit: true
364
+ load_in_8bit: false
365
+ output_dir:
366
+ desc: null
367
+ value: /kaggle/working/
368
+ overwrite_output_dir:
369
+ desc: null
370
+ value: false
371
+ do_train:
372
+ desc: null
373
+ value: false
374
+ do_eval:
375
+ desc: null
376
+ value: false
377
+ do_predict:
378
+ desc: null
379
+ value: false
380
+ eval_strategy:
381
+ desc: null
382
+ value: 'no'
383
+ prediction_loss_only:
384
+ desc: null
385
+ value: false
386
+ per_device_train_batch_size:
387
+ desc: null
388
+ value: 8
389
+ per_device_eval_batch_size:
390
+ desc: null
391
+ value: 8
392
+ per_gpu_train_batch_size:
393
+ desc: null
394
+ value: null
395
+ per_gpu_eval_batch_size:
396
+ desc: null
397
+ value: null
398
+ gradient_accumulation_steps:
399
+ desc: null
400
+ value: 1
401
+ eval_accumulation_steps:
402
+ desc: null
403
+ value: null
404
+ eval_delay:
405
+ desc: null
406
+ value: 0
407
+ learning_rate:
408
+ desc: null
409
+ value: 0.0002
410
+ weight_decay:
411
+ desc: null
412
+ value: 0.0
413
+ adam_beta1:
414
+ desc: null
415
+ value: 0.9
416
+ adam_beta2:
417
+ desc: null
418
+ value: 0.999
419
+ adam_epsilon:
420
+ desc: null
421
+ value: 1.0e-08
422
+ max_grad_norm:
423
+ desc: null
424
+ value: 1.0
425
+ num_train_epochs:
426
+ desc: null
427
+ value: 2
428
+ max_steps:
429
+ desc: null
430
+ value: -1
431
+ lr_scheduler_type:
432
+ desc: null
433
+ value: linear
434
+ lr_scheduler_kwargs:
435
+ desc: null
436
+ value: {}
437
+ warmup_ratio:
438
+ desc: null
439
+ value: 0.0
440
+ warmup_steps:
441
+ desc: null
442
+ value: 0
443
+ log_level:
444
+ desc: null
445
+ value: passive
446
+ log_level_replica:
447
+ desc: null
448
+ value: warning
449
+ log_on_each_node:
450
+ desc: null
451
+ value: true
452
+ logging_dir:
453
+ desc: null
454
+ value: /kaggle/working/runs/May22_11-33-56_2c1b614ec68f
455
+ logging_strategy:
456
+ desc: null
457
+ value: steps
458
+ logging_first_step:
459
+ desc: null
460
+ value: false
461
+ logging_steps:
462
+ desc: null
463
+ value: 10
464
+ logging_nan_inf_filter:
465
+ desc: null
466
+ value: true
467
+ save_strategy:
468
+ desc: null
469
+ value: epoch
470
+ save_steps:
471
+ desc: null
472
+ value: 500
473
+ save_total_limit:
474
+ desc: null
475
+ value: 4
476
+ save_safetensors:
477
+ desc: null
478
+ value: true
479
+ save_on_each_node:
480
+ desc: null
481
+ value: false
482
+ save_only_model:
483
+ desc: null
484
+ value: false
485
+ restore_callback_states_from_checkpoint:
486
+ desc: null
487
+ value: false
488
+ no_cuda:
489
+ desc: null
490
+ value: false
491
+ use_cpu:
492
+ desc: null
493
+ value: false
494
+ use_mps_device:
495
+ desc: null
496
+ value: false
497
+ seed:
498
+ desc: null
499
+ value: 42
500
+ data_seed:
501
+ desc: null
502
+ value: null
503
+ jit_mode_eval:
504
+ desc: null
505
+ value: false
506
+ use_ipex:
507
+ desc: null
508
+ value: false
509
+ bf16:
510
+ desc: null
511
+ value: true
512
+ fp16:
513
+ desc: null
514
+ value: false
515
+ fp16_opt_level:
516
+ desc: null
517
+ value: O1
518
+ half_precision_backend:
519
+ desc: null
520
+ value: auto
521
+ bf16_full_eval:
522
+ desc: null
523
+ value: false
524
+ fp16_full_eval:
525
+ desc: null
526
+ value: false
527
+ tf32:
528
+ desc: null
529
+ value: null
530
+ local_rank:
531
+ desc: null
532
+ value: 0
533
+ ddp_backend:
534
+ desc: null
535
+ value: null
536
+ tpu_num_cores:
537
+ desc: null
538
+ value: null
539
+ tpu_metrics_debug:
540
+ desc: null
541
+ value: false
542
+ debug:
543
+ desc: null
544
+ value: []
545
+ dataloader_drop_last:
546
+ desc: null
547
+ value: false
548
+ eval_steps:
549
+ desc: null
550
+ value: null
551
+ dataloader_num_workers:
552
+ desc: null
553
+ value: 0
554
+ dataloader_prefetch_factor:
555
+ desc: null
556
+ value: null
557
+ past_index:
558
+ desc: null
559
+ value: -1
560
+ run_name:
561
+ desc: null
562
+ value: /kaggle/working/
563
+ disable_tqdm:
564
+ desc: null
565
+ value: false
566
+ remove_unused_columns:
567
+ desc: null
568
+ value: true
569
+ label_names:
570
+ desc: null
571
+ value: null
572
+ load_best_model_at_end:
573
+ desc: null
574
+ value: false
575
+ metric_for_best_model:
576
+ desc: null
577
+ value: null
578
+ greater_is_better:
579
+ desc: null
580
+ value: null
581
+ ignore_data_skip:
582
+ desc: null
583
+ value: false
584
+ fsdp:
585
+ desc: null
586
+ value: []
587
+ fsdp_min_num_params:
588
+ desc: null
589
+ value: 0
590
+ fsdp_config:
591
+ desc: null
592
+ value:
593
+ min_num_params: 0
594
+ xla: false
595
+ xla_fsdp_v2: false
596
+ xla_fsdp_grad_ckpt: false
597
+ fsdp_transformer_layer_cls_to_wrap:
598
+ desc: null
599
+ value: null
600
+ accelerator_config:
601
+ desc: null
602
+ value:
603
+ split_batches: false
604
+ dispatch_batches: null
605
+ even_batches: true
606
+ use_seedable_sampler: true
607
+ non_blocking: false
608
+ gradient_accumulation_kwargs: null
609
+ deepspeed:
610
+ desc: null
611
+ value: null
612
+ label_smoothing_factor:
613
+ desc: null
614
+ value: 0.0
615
+ optim:
616
+ desc: null
617
+ value: adamw_torch
618
+ optim_args:
619
+ desc: null
620
+ value: null
621
+ adafactor:
622
+ desc: null
623
+ value: false
624
+ group_by_length:
625
+ desc: null
626
+ value: false
627
+ length_column_name:
628
+ desc: null
629
+ value: length
630
+ report_to:
631
+ desc: null
632
+ value:
633
+ - tensorboard
634
+ - wandb
635
+ ddp_find_unused_parameters:
636
+ desc: null
637
+ value: null
638
+ ddp_bucket_cap_mb:
639
+ desc: null
640
+ value: null
641
+ ddp_broadcast_buffers:
642
+ desc: null
643
+ value: null
644
+ dataloader_pin_memory:
645
+ desc: null
646
+ value: true
647
+ dataloader_persistent_workers:
648
+ desc: null
649
+ value: false
650
+ skip_memory_metrics:
651
+ desc: null
652
+ value: true
653
+ use_legacy_prediction_loop:
654
+ desc: null
655
+ value: false
656
+ push_to_hub:
657
+ desc: null
658
+ value: false
659
+ resume_from_checkpoint:
660
+ desc: null
661
+ value: null
662
+ hub_model_id:
663
+ desc: null
664
+ value: null
665
+ hub_strategy:
666
+ desc: null
667
+ value: every_save
668
+ hub_token:
669
+ desc: null
670
+ value: <HUB_TOKEN>
671
+ hub_private_repo:
672
+ desc: null
673
+ value: false
674
+ hub_always_push:
675
+ desc: null
676
+ value: false
677
+ gradient_checkpointing:
678
+ desc: null
679
+ value: false
680
+ gradient_checkpointing_kwargs:
681
+ desc: null
682
+ value: null
683
+ include_inputs_for_metrics:
684
+ desc: null
685
+ value: false
686
+ eval_do_concat_batches:
687
+ desc: null
688
+ value: true
689
+ fp16_backend:
690
+ desc: null
691
+ value: auto
692
+ evaluation_strategy:
693
+ desc: null
694
+ value: null
695
+ push_to_hub_model_id:
696
+ desc: null
697
+ value: null
698
+ push_to_hub_organization:
699
+ desc: null
700
+ value: null
701
+ push_to_hub_token:
702
+ desc: null
703
+ value: <PUSH_TO_HUB_TOKEN>
704
+ mp_parameters:
705
+ desc: null
706
+ value: ''
707
+ auto_find_batch_size:
708
+ desc: null
709
+ value: true
710
+ full_determinism:
711
+ desc: null
712
+ value: false
713
+ torchdynamo:
714
+ desc: null
715
+ value: null
716
+ ray_scope:
717
+ desc: null
718
+ value: last
719
+ ddp_timeout:
720
+ desc: null
721
+ value: 1800
722
+ torch_compile:
723
+ desc: null
724
+ value: false
725
+ torch_compile_backend:
726
+ desc: null
727
+ value: null
728
+ torch_compile_mode:
729
+ desc: null
730
+ value: null
731
+ dispatch_batches:
732
+ desc: null
733
+ value: null
734
+ split_batches:
735
+ desc: null
736
+ value: null
737
+ include_tokens_per_second:
738
+ desc: null
739
+ value: false
740
+ include_num_input_tokens_seen:
741
+ desc: null
742
+ value: false
743
+ neftune_noise_alpha:
744
+ desc: null
745
+ value: null
746
+ optim_target_modules:
747
+ desc: null
748
+ value: null
749
+ batch_eval_metrics:
750
+ desc: null
751
+ value: false
752
+ model/num_parameters:
753
+ desc: null
754
+ value: 13033919160
wandb/run-20240522_113413-8mudzhjp/files/output.log ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
2
+ warnings.warn(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:464: UserWarning: torch.utils.checkpoint: the use_reentrant parameter should be passed explicitly. In version 2.4 we will raise an exception if use_reentrant is not passed. use_reentrant=False is recommended, but if you need to preserve the current default behavior, you can pass use_reentrant=True. Refer to docs for more details on the differences between the two variants.
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:464: UserWarning: torch.utils.checkpoint: the use_reentrant parameter should be passed explicitly. In version 2.4 we will raise an exception if use_reentrant is not passed. use_reentrant=False is recommended, but if you need to preserve the current default behavior, you can pass use_reentrant=True. Refer to docs for more details on the differences between the two variants.
6
+ warnings.warn(
7
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
8
+ warnings.warn(
9
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
10
+ To disable this warning, you can either:
11
+ - Avoid using `tokenizers` before the fork if possible
12
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
13
+ usage: huggingface-cli <command> [<args>]
14
+ positional arguments:
15
+ {env,login,whoami,logout,repo,upload,download,lfs-enable-largefiles,lfs-multipart-upload,scan-cache,delete-cache,tag}
16
+ huggingface-cli command helpers
17
+ env Print information about the environment.
18
+ login Log in using a token from
19
+ huggingface.co/settings/tokens
20
+ whoami Find out which huggingface.co account you are logged
21
+ in as.
22
+ logout Log out
23
+ repo {create} Commands to interact with your huggingface.co
24
+ repos.
25
+ upload Upload a file or a folder to a repo on the Hub
26
+ download Download files from the Hub
27
+ lfs-enable-largefiles
28
+ Configure your repository to enable upload of files >
29
+ 5GB.
30
+ scan-cache Scan cache directory.
31
+ delete-cache Delete revisions from the cache directory.
32
+ tag (create, list, delete) tags for a repo in the hub
33
+ options:
34
+ -h, --help show this help message and exit
35
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
36
+ To disable this warning, you can either:
37
+ - Avoid using `tokenizers` before the fork if possible
38
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
39
+ _| _| _| _| _|_|_| _|_|_| _|_|_| _| _| _|_|_| _|_|_|_| _|_| _|_|_| _|_|_|_|
40
+ _| _| _| _| _| _| _| _|_| _| _| _| _| _| _| _|
41
+ _|_|_|_| _| _| _| _|_| _| _|_| _| _| _| _| _| _|_| _|_|_| _|_|_|_| _| _|_|_|
42
+ _| _| _| _| _| _| _| _| _| _| _|_| _| _| _| _| _| _| _|
43
+ _| _| _|_| _|_|_| _|_|_| _|_|_| _| _| _|_|_| _| _| _| _|_|_| _|_|_|_|
44
+ To login, `huggingface_hub` requires a token generated from https://huggingface.co/settings/tokens .
45
+ Enter your token (input will not be visible): Traceback (most recent call last):
46
+ File "/opt/conda/bin/huggingface-cli", line 8, in <module>
47
+ sys.exit(main())
48
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/commands/huggingface_cli.py", line 51, in main
49
+ service.run()
50
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/commands/user.py", line 98, in run
51
+ login(token=self.args.token, add_to_git_credential=self.args.add_to_git_credential)
52
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/_login.py", line 115, in login
53
+ interpreter_login(new_session=new_session, write_permission=write_permission)
54
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/_login.py", line 191, in interpreter_login
55
+ token = getpass("Enter your token (input will not be visible): ")
56
+ File "/opt/conda/lib/python3.10/getpass.py", line 77, in unix_getpass
57
+ passwd = _raw_input(prompt, stream, input=input)
58
+ File "/opt/conda/lib/python3.10/getpass.py", line 146, in _raw_input
59
+ line = input.readline()
60
+ File "/opt/conda/lib/python3.10/codecs.py", line 319, in decode
61
+ def decode(self, input, final=False):
62
+ KeyboardInterrupt
63
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
64
+ To disable this warning, you can either:
65
+ - Avoid using `tokenizers` before the fork if possible
66
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
67
+ _| _| _| _| _|_|_| _|_|_| _|_|_| _| _| _|_|_| _|_|_|_| _|_| _|_|_| _|_|_|_|
68
+ _| _| _| _| _| _| _| _|_| _| _| _| _| _| _| _|
69
+ _|_|_|_| _| _| _| _|_| _| _|_| _| _| _| _| _| _|_| _|_|_| _|_|_|_| _| _|_|_|
70
+ _| _| _| _| _| _| _| _| _| _| _|_| _| _| _| _| _| _| _|
71
+ _| _| _|_| _|_|_| _|_|_| _|_|_| _| _| _|_|_| _| _| _| _|_|_| _|_|_|_|
72
+ To login, `huggingface_hub` requires a token generated from https://huggingface.co/settings/tokens .
73
+ Enter your token (input will not be visible): Traceback (most recent call last):
74
+ File "/opt/conda/bin/huggingface-cli", line 8, in <module>
75
+ sys.exit(main())
76
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/commands/huggingface_cli.py", line 51, in main
77
+ service.run()
78
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/commands/user.py", line 98, in run
79
+ login(token=self.args.token, add_to_git_credential=self.args.add_to_git_credential)
80
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/_login.py", line 115, in login
81
+ interpreter_login(new_session=new_session, write_permission=write_permission)
82
+ File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/_login.py", line 191, in interpreter_login
83
+ token = getpass("Enter your token (input will not be visible): ")
84
+ File "/opt/conda/lib/python3.10/getpass.py", line 77, in unix_getpass
85
+ passwd = _raw_input(prompt, stream, input=input)
86
+ File "/opt/conda/lib/python3.10/getpass.py", line 146, in _raw_input
87
+ line = input.readline()
88
+ File "/opt/conda/lib/python3.10/codecs.py", line 319, in decode
89
+ def decode(self, input, final=False):
90
+ KeyboardInterrupt
91
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
92
+ warnings.warn(
93
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
wandb/run-20240522_113413-8mudzhjp/files/requirements.txt ADDED
@@ -0,0 +1,878 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.23.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.3
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.3
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.30.1
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.3
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.18.0
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.15.2.45.49
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.1
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.2
92
+ bokeh==3.4.1
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.69
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ contourpy==1.2.1
137
+ convertdate==2.4.0
138
+ crcmod==1.7
139
+ cryptography==41.0.7
140
+ cuda-python==12.4.0
141
+ cudf==23.8.0
142
+ cufflinks==0.17.3
143
+ cuml==23.8.0
144
+ cupy==13.0.0
145
+ cycler==0.12.1
146
+ cymem==2.0.8
147
+ cytoolz==0.12.3
148
+ daal4py==2024.3.0
149
+ daal==2024.3.0
150
+ dacite==1.8.1
151
+ dask-cuda==23.8.0
152
+ dask-cudf==23.8.0
153
+ dask-expr==1.0.11
154
+ dask==2024.4.1
155
+ dataclasses-json==0.6.4
156
+ dataproc_jupyter_plugin==0.1.66
157
+ datasets==2.18.0
158
+ datashader==0.16.0
159
+ datatile==1.0.3
160
+ db-dtypes==1.2.0
161
+ deap==1.4.1
162
+ debugpy==1.8.0
163
+ decorator==5.1.1
164
+ deepdiff==7.0.1
165
+ defusedxml==0.7.1
166
+ deprecation==2.1.0
167
+ descartes==1.1.0
168
+ dill==0.3.8
169
+ dipy==1.9.0
170
+ distlib==0.3.8
171
+ distributed==2023.7.1
172
+ distro==1.9.0
173
+ dm-tree==0.1.8
174
+ docker-pycreds==0.4.0
175
+ docker==7.0.0
176
+ docopt==0.6.2
177
+ docstring-parser==0.15
178
+ docstring-to-markdown==0.15
179
+ docutils==0.21.1
180
+ earthengine-api==0.1.399
181
+ easydict==1.13
182
+ easyocr==1.7.1
183
+ ecos==2.0.13
184
+ einops==0.8.0
185
+ eli5==0.13.0
186
+ emoji==2.11.0
187
+ en-core-web-lg==3.7.1
188
+ en-core-web-sm==3.7.1
189
+ entrypoints==0.4
190
+ ephem==4.1.5
191
+ esda==2.5.1
192
+ essentia==2.1b6.dev1110
193
+ et-xmlfile==1.1.0
194
+ etils==1.6.0
195
+ exceptiongroup==1.2.0
196
+ executing==2.0.1
197
+ explainable-ai-sdk==1.3.3
198
+ fastai==2.7.14
199
+ fastapi==0.108.0
200
+ fastavro==1.9.3
201
+ fastcore==1.5.29
202
+ fastdownload==0.0.7
203
+ fasteners==0.19
204
+ fastjsonschema==2.19.1
205
+ fastprogress==1.0.3
206
+ fastrlock==0.8.2
207
+ fasttext==0.9.2
208
+ feather-format==0.4.1
209
+ featuretools==1.30.0
210
+ filelock==3.13.1
211
+ fiona==1.9.6
212
+ fitter==1.7.0
213
+ flake8==7.0.0
214
+ flashtext==2.7
215
+ flatbuffers==23.5.26
216
+ flax==0.8.2
217
+ folium==0.16.0
218
+ fonttools==4.47.0
219
+ fonttools==4.51.0
220
+ fqdn==1.5.1
221
+ frozendict==2.4.2
222
+ frozenlist==1.4.1
223
+ fsspec==2024.2.0
224
+ fsspec==2024.3.1
225
+ funcy==2.0
226
+ fury==0.10.0
227
+ future==1.0.0
228
+ fuzzywuzzy==0.18.0
229
+ gast==0.5.4
230
+ gatspy==0.3
231
+ gcsfs==2024.2.0
232
+ gensim==4.3.2
233
+ geographiclib==2.0
234
+ geojson==3.1.0
235
+ geopandas==0.14.3
236
+ geoplot==0.5.1
237
+ geopy==2.4.1
238
+ geoviews==1.12.0
239
+ ggplot==0.11.5
240
+ giddy==2.3.5
241
+ gitdb==4.0.11
242
+ google-ai-generativelanguage==0.6.2
243
+ google-api-core==2.11.1
244
+ google-api-core==2.18.0
245
+ google-api-python-client==2.126.0
246
+ google-apitools==0.5.31
247
+ google-auth-httplib2==0.2.0
248
+ google-auth-oauthlib==1.2.0
249
+ google-auth==2.26.1
250
+ google-cloud-aiplatform==0.6.0a1
251
+ google-cloud-artifact-registry==1.10.0
252
+ google-cloud-automl==1.0.1
253
+ google-cloud-bigquery==2.34.4
254
+ google-cloud-bigtable==1.7.3
255
+ google-cloud-core==2.4.1
256
+ google-cloud-datastore==2.19.0
257
+ google-cloud-dlp==3.14.0
258
+ google-cloud-jupyter-config==0.0.5
259
+ google-cloud-language==2.13.3
260
+ google-cloud-monitoring==2.18.0
261
+ google-cloud-pubsub==2.19.0
262
+ google-cloud-pubsublite==1.9.0
263
+ google-cloud-recommendations-ai==0.7.1
264
+ google-cloud-resource-manager==1.11.0
265
+ google-cloud-spanner==3.40.1
266
+ google-cloud-storage==1.44.0
267
+ google-cloud-translate==3.12.1
268
+ google-cloud-videointelligence==2.13.3
269
+ google-cloud-vision==2.8.0
270
+ google-crc32c==1.5.0
271
+ google-generativeai==0.5.1
272
+ google-pasta==0.2.0
273
+ google-resumable-media==2.7.0
274
+ googleapis-common-protos==1.62.0
275
+ gplearn==0.4.2
276
+ gpustat==1.0.0
277
+ gpxpy==1.6.2
278
+ graphviz==0.20.3
279
+ greenlet==3.0.3
280
+ grpc-google-iam-v1==0.12.7
281
+ grpcio-status==1.48.1
282
+ grpcio-status==1.48.2
283
+ grpcio==1.51.1
284
+ grpcio==1.60.0
285
+ gviz-api==1.10.0
286
+ gym-notices==0.0.8
287
+ gym==0.26.2
288
+ gymnasium==0.29.0
289
+ h11==0.14.0
290
+ h2o==3.46.0.1
291
+ h5netcdf==1.3.0
292
+ h5py==3.10.0
293
+ haversine==2.8.1
294
+ hdfs==2.7.3
295
+ hep-ml==0.7.2
296
+ hijri-converter==2.3.1
297
+ hmmlearn==0.3.2
298
+ holidays==0.24
299
+ holoviews==1.18.3
300
+ hpsklearn==0.1.0
301
+ html5lib==1.1
302
+ htmlmin==0.1.12
303
+ httpcore==1.0.5
304
+ httplib2==0.21.0
305
+ httptools==0.6.1
306
+ httpx==0.27.0
307
+ huggingface-hub==0.23.1
308
+ hunspell==0.5.5
309
+ hydra-slayer==0.5.0
310
+ hyperopt==0.2.7
311
+ hypertools==0.8.0
312
+ idna==3.6
313
+ igraph==0.11.4
314
+ imagecodecs==2024.1.1
315
+ imageio==2.33.1
316
+ imbalanced-learn==0.12.2
317
+ imgaug==0.4.0
318
+ importlib-metadata==6.11.0
319
+ importlib-metadata==7.0.1
320
+ importlib-resources==6.1.1
321
+ inequality==1.0.1
322
+ iniconfig==2.0.0
323
+ ipydatawidgets==4.3.5
324
+ ipykernel==6.28.0
325
+ ipyleaflet==0.18.2
326
+ ipympl==0.7.0
327
+ ipython-genutils==0.2.0
328
+ ipython-genutils==0.2.0
329
+ ipython-sql==0.5.0
330
+ ipython==8.20.0
331
+ ipyvolume==0.6.3
332
+ ipyvue==1.11.0
333
+ ipyvuetify==1.9.4
334
+ ipywebrtc==0.6.0
335
+ ipywidgets==7.7.1
336
+ isoduration==20.11.0
337
+ isort==5.13.2
338
+ isoweek==1.3.3
339
+ itsdangerous==2.2.0
340
+ jaraco.classes==3.3.0
341
+ jax-jumpy==1.0.0
342
+ jax==0.4.23
343
+ jaxlib==0.4.23.dev20240116
344
+ jedi==0.19.1
345
+ jeepney==0.8.0
346
+ jieba==0.42.1
347
+ jmespath==1.0.1
348
+ joblib==1.4.0
349
+ json5==0.9.14
350
+ jsonpatch==1.33
351
+ jsonpointer==2.4
352
+ jsonschema-specifications==2023.12.1
353
+ jsonschema==4.20.0
354
+ jupyter-console==6.6.3
355
+ jupyter-events==0.9.0
356
+ jupyter-http-over-ws==0.0.8
357
+ jupyter-lsp==1.5.1
358
+ jupyter-server-mathjax==0.2.6
359
+ jupyter-ydoc==0.2.5
360
+ jupyter_client==7.4.9
361
+ jupyter_client==8.6.0
362
+ jupyter_core==5.7.1
363
+ jupyter_server==2.12.5
364
+ jupyter_server_fileid==0.9.1
365
+ jupyter_server_proxy==4.1.0
366
+ jupyter_server_terminals==0.5.1
367
+ jupyter_server_ydoc==0.8.0
368
+ jupyterlab-lsp==5.1.0
369
+ jupyterlab-widgets==3.0.9
370
+ jupyterlab==4.1.6
371
+ jupyterlab_git==0.44.0
372
+ jupyterlab_pygments==0.3.0
373
+ jupyterlab_server==2.25.2
374
+ jupytext==1.16.0
375
+ kaggle-environments==1.14.3
376
+ kaggle==1.6.12
377
+ kagglehub==0.2.3
378
+ keras-cv==0.8.2
379
+ keras-nlp==0.9.3
380
+ keras-tuner==1.4.6
381
+ keras==3.2.1
382
+ kernels-mixer==0.0.7
383
+ keyring==24.3.0
384
+ keyrings.google-artifactregistry-auth==1.1.2
385
+ kfp-pipeline-spec==0.2.2
386
+ kfp-server-api==2.0.5
387
+ kfp==2.5.0
388
+ kiwisolver==1.4.5
389
+ kmapper==2.0.1
390
+ kmodes==0.12.2
391
+ korean-lunar-calendar==0.3.1
392
+ kornia==0.7.2
393
+ kornia_rs==0.1.3
394
+ kt-legacy==1.0.5
395
+ kubernetes==26.1.0
396
+ langcodes==3.3.0
397
+ langid==1.1.6
398
+ lazy_loader==0.3
399
+ learntools==0.3.4
400
+ leven==1.0.4
401
+ libclang==16.0.6
402
+ libmambapy==1.5.0
403
+ libpysal==4.9.2
404
+ librosa==0.10.1
405
+ lightgbm==4.2.0
406
+ lightning-utilities==0.11.2
407
+ lime==0.2.0.1
408
+ line-profiler==4.1.2
409
+ linkify-it-py==2.0.3
410
+ llvmlite==0.41.1
411
+ llvmlite==0.42.0
412
+ lml==0.1.0
413
+ locket==1.0.0
414
+ loguru==0.7.2
415
+ loralib==0.1.2
416
+ lxml==5.2.1
417
+ lz4==4.3.3
418
+ mamba==1.5.0
419
+ mapclassify==2.6.1
420
+ markdown-it-py==3.0.0
421
+ marshmallow==3.21.1
422
+ matplotlib-inline==0.1.6
423
+ matplotlib-venn==0.11.10
424
+ matplotlib==3.7.5
425
+ matplotlib==3.8.4
426
+ mccabe==0.7.0
427
+ mdit-py-plugins==0.4.0
428
+ mdurl==0.1.2
429
+ memory-profiler==0.61.0
430
+ menuinst==2.0.1
431
+ mercantile==1.2.1
432
+ mgwr==2.2.1
433
+ missingno==0.5.2
434
+ mistune==0.8.4
435
+ mizani==0.11.1
436
+ ml-dtypes==0.2.0
437
+ mlcrate==0.2.0
438
+ mlens==0.2.3
439
+ mlxtend==0.23.1
440
+ mne==1.6.1
441
+ mnist==0.2.2
442
+ momepy==0.7.0
443
+ more-itertools==10.2.0
444
+ mpld3==0.5.10
445
+ mpmath==1.3.0
446
+ msgpack==1.0.7
447
+ multidict==6.0.4
448
+ multimethod==1.10
449
+ multipledispatch==1.0.0
450
+ multiprocess==0.70.16
451
+ munkres==1.1.4
452
+ murmurhash==1.0.10
453
+ mypy-extensions==1.0.0
454
+ namex==0.0.8
455
+ nb-conda-kernels==2.3.1
456
+ nb_conda==2.2.1
457
+ nbclassic==1.0.0
458
+ nbclient==0.5.13
459
+ nbconvert==6.4.5
460
+ nbdime==3.2.0
461
+ nbformat==5.9.2
462
+ ndindex==1.8
463
+ nest-asyncio==1.5.8
464
+ networkx==3.2.1
465
+ nibabel==5.2.1
466
+ nilearn==0.10.4
467
+ ninja==1.11.1.1
468
+ nltk==3.2.4
469
+ nose==1.3.7
470
+ notebook==6.5.4
471
+ notebook==6.5.6
472
+ notebook_executor==0.2
473
+ notebook_shim==0.2.3
474
+ numba==0.58.1
475
+ numba==0.59.1
476
+ numexpr==2.10.0
477
+ numpy==1.26.4
478
+ nvidia-cublas-cu12==12.1.3.1
479
+ nvidia-cuda-cupti-cu12==12.1.105
480
+ nvidia-cuda-nvrtc-cu12==12.1.105
481
+ nvidia-cuda-runtime-cu12==12.1.105
482
+ nvidia-cudnn-cu12==8.9.2.26
483
+ nvidia-cufft-cu12==11.0.2.54
484
+ nvidia-curand-cu12==10.3.2.106
485
+ nvidia-cusolver-cu12==11.4.5.107
486
+ nvidia-cusparse-cu12==12.1.0.106
487
+ nvidia-ml-py==11.495.46
488
+ nvidia-nccl-cu12==2.20.5
489
+ nvidia-nvjitlink-cu12==12.5.40
490
+ nvidia-nvtx-cu12==12.1.105
491
+ nvtx==0.2.10
492
+ oauth2client==4.1.3
493
+ oauthlib==3.2.2
494
+ objsize==0.6.1
495
+ odfpy==1.4.1
496
+ olefile==0.47
497
+ onnx==1.16.0
498
+ opencensus-context==0.1.3
499
+ opencensus==0.11.4
500
+ opencv-contrib-python==4.9.0.80
501
+ opencv-python-headless==4.9.0.80
502
+ opencv-python==4.9.0.80
503
+ openpyxl==3.1.2
504
+ openslide-python==1.3.1
505
+ opentelemetry-api==1.22.0
506
+ opentelemetry-exporter-otlp-proto-common==1.22.0
507
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
508
+ opentelemetry-exporter-otlp-proto-http==1.22.0
509
+ opentelemetry-exporter-otlp==1.22.0
510
+ opentelemetry-proto==1.22.0
511
+ opentelemetry-sdk==1.22.0
512
+ opentelemetry-semantic-conventions==0.43b0
513
+ opt-einsum==3.3.0
514
+ optax==0.2.2
515
+ optree==0.11.0
516
+ optuna==3.6.1
517
+ orbax-checkpoint==0.5.9
518
+ ordered-set==4.1.0
519
+ orjson==3.9.10
520
+ ortools==9.4.1874
521
+ osmnx==1.9.2
522
+ overrides==7.4.0
523
+ packaging==21.3
524
+ pandas-datareader==0.10.0
525
+ pandas-profiling==3.6.6
526
+ pandas-summary==0.2.0
527
+ pandas==2.1.4
528
+ pandas==2.2.2
529
+ pandasql==0.7.3
530
+ pandocfilters==1.5.0
531
+ panel==1.4.1
532
+ papermill==2.5.0
533
+ param==2.1.0
534
+ parso==0.8.3
535
+ partd==1.4.1
536
+ path.py==12.5.0
537
+ path==16.14.0
538
+ pathos==0.3.2
539
+ pathy==0.10.3
540
+ patsy==0.5.6
541
+ pdf2image==1.17.0
542
+ peft==0.11.1
543
+ pettingzoo==1.24.0
544
+ pexpect==4.8.0
545
+ pexpect==4.9.0
546
+ phik==0.12.4
547
+ pickleshare==0.7.5
548
+ pillow==10.3.0
549
+ pip==23.3.2
550
+ pkgutil_resolve_name==1.3.10
551
+ platformdirs==4.2.0
552
+ plotly-express==0.4.1
553
+ plotly==5.18.0
554
+ plotnine==0.13.4
555
+ pluggy==1.4.0
556
+ pointpats==2.4.0
557
+ polars==0.20.21
558
+ polyglot==16.7.4
559
+ pooch==1.8.1
560
+ pox==0.3.4
561
+ ppca==0.0.4
562
+ ppft==1.7.6.8
563
+ preprocessing==0.1.13
564
+ preshed==3.0.9
565
+ prettytable==3.9.0
566
+ progressbar2==4.4.2
567
+ prometheus-client==0.19.0
568
+ promise==2.3
569
+ prompt-toolkit==3.0.42
570
+ prompt-toolkit==3.0.43
571
+ prophet==1.1.1
572
+ proto-plus==1.23.0
573
+ protobuf==3.20.3
574
+ protobuf==4.21.12
575
+ psutil==5.9.3
576
+ psutil==5.9.7
577
+ ptyprocess==0.7.0
578
+ pudb==2024.1
579
+ pure-eval==0.2.2
580
+ py-cpuinfo==9.0.0
581
+ py-spy==0.3.14
582
+ py4j==0.10.9.7
583
+ pyLDAvis==3.4.1
584
+ pyOpenSSL==23.3.0
585
+ pyaml==23.12.0
586
+ pyarrow-hotfix==0.6
587
+ pyarrow==15.0.2
588
+ pyasn1-modules==0.3.0
589
+ pyasn1==0.5.1
590
+ pybind11==2.12.0
591
+ pyclipper==1.3.0.post5
592
+ pycodestyle==2.11.1
593
+ pycosat==0.6.6
594
+ pycparser==2.21
595
+ pycryptodome==3.20.0
596
+ pyct==0.5.0
597
+ pycuda==2024.1
598
+ pydantic==2.5.3
599
+ pydantic==2.7.0
600
+ pydantic_core==2.14.6
601
+ pydantic_core==2.18.1
602
+ pydegensac==0.1.2
603
+ pydicom==2.4.4
604
+ pydocstyle==6.3.0
605
+ pydot==1.4.2
606
+ pydub==0.25.1
607
+ pyemd==1.0.0
608
+ pyerfa==2.0.1.4
609
+ pyexcel-io==0.6.6
610
+ pyexcel-ods==0.6.0
611
+ pyflakes==3.2.0
612
+ pygltflib==1.16.2
613
+ pykalman==0.9.7
614
+ pylibraft==23.8.0
615
+ pylint==3.1.0
616
+ pymc3==3.11.4
617
+ pymongo==3.13.0
618
+ pynndescent==0.5.12
619
+ pynvml==11.4.1
620
+ pynvrtc==9.2
621
+ pyparsing==3.1.1
622
+ pyparsing==3.1.2
623
+ pypdf==4.2.0
624
+ pyproj==3.6.1
625
+ pysal==24.1
626
+ pyshp==2.3.1
627
+ pytesseract==0.3.10
628
+ pytest==8.1.1
629
+ python-bidi==0.4.2
630
+ python-dateutil==2.9.0.post0
631
+ python-dotenv==1.0.0
632
+ python-json-logger==2.0.7
633
+ python-louvain==0.16
634
+ python-lsp-jsonrpc==1.1.2
635
+ python-lsp-server==1.11.0
636
+ python-slugify==8.0.4
637
+ python-utils==3.8.2
638
+ pythreejs==2.4.2
639
+ pytoolconfig==1.3.1
640
+ pytools==2024.1.1
641
+ pytorch-ignite==0.5.0.post2
642
+ pytorch-lightning==2.2.2
643
+ pytz==2023.3.post1
644
+ pytz==2024.1
645
+ pyu2f==0.1.5
646
+ pyviz_comms==3.0.2
647
+ pyzmq==24.0.1
648
+ pyzmq==25.1.2
649
+ qgrid==1.3.1
650
+ qtconsole==5.5.1
651
+ quantecon==0.7.2
652
+ qudida==0.0.4
653
+ raft-dask==23.8.0
654
+ rasterio==1.3.10
655
+ rasterstats==0.19.0
656
+ ray-cpp==2.9.0
657
+ ray==2.9.0
658
+ referencing==0.32.1
659
+ regex==2023.12.25
660
+ requests-oauthlib==1.3.1
661
+ requests-toolbelt==0.10.1
662
+ requests==2.31.0
663
+ retrying==1.3.3
664
+ retrying==1.3.4
665
+ rfc3339-validator==0.1.4
666
+ rfc3986-validator==0.1.1
667
+ rgf-python==3.12.0
668
+ rich-click==1.7.4
669
+ rich==13.7.0
670
+ rich==13.7.1
671
+ rmm==23.8.0
672
+ rope==1.13.0
673
+ rpds-py==0.16.2
674
+ rsa==4.9
675
+ ruamel-yaml-conda==0.15.100
676
+ ruamel.yaml.clib==0.2.7
677
+ ruamel.yaml==0.17.40
678
+ s2sphere==0.2.5
679
+ s3fs==2024.2.0
680
+ s3transfer==0.6.2
681
+ safetensors==0.4.3
682
+ scattertext==0.1.19
683
+ scikit-image==0.22.0
684
+ scikit-learn-intelex==2024.3.0
685
+ scikit-learn==1.2.2
686
+ scikit-multilearn==0.2.0
687
+ scikit-optimize==0.10.1
688
+ scikit-plot==0.3.7
689
+ scikit-surprise==1.1.3
690
+ scipy==1.11.4
691
+ scipy==1.13.0
692
+ seaborn==0.12.2
693
+ segment_anything==1.0
694
+ segregation==2.5
695
+ semver==3.0.2
696
+ sentencepiece==0.2.0
697
+ sentry-sdk==1.45.0
698
+ setproctitle==1.3.3
699
+ setuptools-git==1.2
700
+ setuptools-scm==8.0.4
701
+ setuptools==69.0.3
702
+ shap==0.44.1
703
+ shapely==2.0.4
704
+ shellingham==1.5.4
705
+ simpervisor==1.0.0
706
+ simplejson==3.19.2
707
+ six==1.16.0
708
+ sklearn-pandas==2.2.0
709
+ slicer==0.0.7
710
+ smart-open==6.4.0
711
+ smmap==5.0.1
712
+ sniffio==1.3.0
713
+ snowballstemmer==2.2.0
714
+ snuggs==1.4.7
715
+ sortedcontainers==2.4.0
716
+ soundfile==0.12.1
717
+ soupsieve==2.5
718
+ soxr==0.3.7
719
+ spacy-legacy==3.0.12
720
+ spacy-loggers==1.0.5
721
+ spacy==3.7.3
722
+ spaghetti==1.7.5.post1
723
+ spectral==0.23.1
724
+ spglm==1.1.0
725
+ sphinx-rtd-theme==0.2.4
726
+ spint==1.0.7
727
+ splot==1.1.5.post1
728
+ spopt==0.6.0
729
+ spreg==1.4.2
730
+ spvcm==0.3.0
731
+ sqlparse==0.4.4
732
+ squarify==0.4.3
733
+ srsly==2.4.8
734
+ stable-baselines3==2.1.0
735
+ stack-data==0.6.2
736
+ stack-data==0.6.3
737
+ stanio==0.5.0
738
+ starlette==0.32.0.post1
739
+ statsmodels==0.14.1
740
+ stemming==1.0.1
741
+ stop-words==2018.7.23
742
+ stopit==1.1.2
743
+ stumpy==1.12.0
744
+ sympy==1.12
745
+ tables==3.9.2
746
+ tabulate==0.9.0
747
+ tangled-up-in-unicode==0.2.0
748
+ tbb==2021.12.0
749
+ tblib==3.0.0
750
+ tenacity==8.2.3
751
+ tensorboard-data-server==0.7.2
752
+ tensorboard-plugin-profile==2.15.0
753
+ tensorboard==2.15.1
754
+ tensorboardX==2.6.2.2
755
+ tensorflow-cloud==0.1.16
756
+ tensorflow-datasets==4.9.4
757
+ tensorflow-decision-forests==1.8.1
758
+ tensorflow-estimator==2.15.0
759
+ tensorflow-hub==0.16.1
760
+ tensorflow-io-gcs-filesystem==0.35.0
761
+ tensorflow-io==0.35.0
762
+ tensorflow-metadata==0.14.0
763
+ tensorflow-probability==0.23.0
764
+ tensorflow-serving-api==2.14.1
765
+ tensorflow-text==2.15.0
766
+ tensorflow-transform==0.14.0
767
+ tensorflow==2.15.0
768
+ tensorstore==0.1.56
769
+ termcolor==2.4.0
770
+ terminado==0.18.0
771
+ testpath==0.6.0
772
+ text-unidecode==1.3
773
+ textblob==0.18.0.post0
774
+ texttable==1.7.0
775
+ tf_keras==2.15.1
776
+ tfp-nightly==0.24.0.dev0
777
+ thinc==8.2.2
778
+ threadpoolctl==3.2.0
779
+ tifffile==2023.12.9
780
+ timm==0.9.16
781
+ tinycss2==1.2.1
782
+ tobler==0.11.2
783
+ tokenizers==0.19.1
784
+ toml==0.10.2
785
+ tomli==2.0.1
786
+ tomlkit==0.12.4
787
+ toolz==0.12.1
788
+ torch==2.3.0
789
+ torchaudio==2.1.2
790
+ torchdata==0.7.1
791
+ torchinfo==1.8.0
792
+ torchmetrics==1.3.2
793
+ torchtext==0.16.2
794
+ torchvision==0.16.2
795
+ tornado==6.3.3
796
+ tqdm==4.66.1
797
+ traceml==1.0.8
798
+ traitlets==5.9.0
799
+ traittypes==0.2.1
800
+ transformers==4.41.0
801
+ treelite-runtime==3.2.0
802
+ treelite==3.2.0
803
+ triton==2.3.0
804
+ truststore==0.8.0
805
+ trx-python==0.2.9
806
+ tsfresh==0.20.2
807
+ typeguard==4.1.5
808
+ typer==0.9.0
809
+ typer==0.9.4
810
+ types-python-dateutil==2.8.19.20240106
811
+ typing-inspect==0.9.0
812
+ typing-utils==0.1.0
813
+ typing_extensions==4.9.0
814
+ tzdata==2023.4
815
+ uc-micro-py==1.0.3
816
+ ucx-py==0.33.0
817
+ ujson==5.9.0
818
+ umap-learn==0.5.6
819
+ unicodedata2==15.1.0
820
+ update-checker==0.18.0
821
+ uri-template==1.3.0
822
+ uritemplate==3.0.1
823
+ urllib3==1.26.18
824
+ urllib3==2.1.0
825
+ urwid==2.6.10
826
+ urwid_readline==0.14
827
+ uvicorn==0.25.0
828
+ uvloop==0.19.0
829
+ vaex-astro==0.9.3
830
+ vaex-core==4.17.1
831
+ vaex-hdf5==0.14.1
832
+ vaex-jupyter==0.8.2
833
+ vaex-ml==0.18.3
834
+ vaex-server==0.9.0
835
+ vaex-viz==0.5.4
836
+ vaex==4.17.0
837
+ vec_noise==1.1.4
838
+ vecstack==0.4.0
839
+ virtualenv==20.21.0
840
+ visions==0.7.5
841
+ vowpalwabbit==9.9.0
842
+ vtk==9.3.0
843
+ wandb==0.16.6
844
+ wasabi==1.1.2
845
+ watchfiles==0.21.0
846
+ wavio==0.0.8
847
+ wcwidth==0.2.13
848
+ weasel==0.3.4
849
+ webcolors==1.13
850
+ webencodings==0.5.1
851
+ websocket-client==1.7.0
852
+ websockets==12.0
853
+ wfdb==4.1.2
854
+ whatthepatch==1.0.5
855
+ wheel==0.42.0
856
+ widgetsnbextension==3.6.6
857
+ witwidget==1.8.1
858
+ woodwork==0.30.0
859
+ wordcloud==1.9.3
860
+ wordsegment==1.3.1
861
+ wrapt==1.14.1
862
+ xarray-einstats==0.7.0
863
+ xarray==2024.3.0
864
+ xformers==0.0.26.post1
865
+ xgboost==2.0.3
866
+ xvfbwrapper==0.2.9
867
+ xxhash==3.4.1
868
+ xyzservices==2024.4.0
869
+ y-py==0.6.2
870
+ yapf==0.40.2
871
+ yarl==1.9.3
872
+ yarl==1.9.4
873
+ ydata-profiling==4.6.4
874
+ yellowbrick==1.5
875
+ ypy-websocket==0.8.4
876
+ zict==3.0.0
877
+ zipp==3.17.0
878
+ zstandard==0.22.0
wandb/run-20240522_113413-8mudzhjp/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-05-22T11:34:14.876249",
5
+ "startedAt": "2024-05-22T11:34:13.994936",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "2c1b614ec68f",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.144,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.144,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.144,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.144,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.144,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5656.421318054199
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357563018798828
61
+ }
62
+ }
wandb/run-20240522_113413-8mudzhjp/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 3.5063, "train/grad_norm": 2.443696975708008, "train/learning_rate": 0.0, "train/epoch": 2.0, "train/global_step": 2000, "_timestamp": 1716386681.8685813, "_runtime": 9027.865963220596, "_step": 226, "train_runtime": 7028.5663, "train_samples_per_second": 0.285, "train_steps_per_second": 0.285, "total_flos": 1.046896491923424e+16, "train_loss": 3.840830388069153}
wandb/run-20240522_113413-8mudzhjp/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240522_113413-8mudzhjp/logs/debug.log ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-22 11:34:13,996 INFO MainThread:217 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
2
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Configure stats pid to 217
3
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240522_113413-8mudzhjp/logs/debug.log
10
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240522_113413-8mudzhjp/logs/debug-internal.log
11
+ 2024-05-22 11:34:13,997 INFO MainThread:217 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ef92390cee0>
12
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():561] calling init triggers
13
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():611] starting backend
16
+ 2024-05-22 11:34:13,998 INFO MainThread:217 [wandb_init.py:init():615] setting up manager
17
+ 2024-05-22 11:34:14,000 INFO MainThread:217 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-05-22 11:34:14,002 INFO MainThread:217 [wandb_init.py:init():623] backend started and connected
19
+ 2024-05-22 11:34:14,016 INFO MainThread:217 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-05-22 11:34:14,540 INFO MainThread:217 [wandb_init.py:init():715] updated telemetry
21
+ 2024-05-22 11:34:14,544 INFO MainThread:217 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
22
+ 2024-05-22 11:34:14,778 INFO MainThread:217 [wandb_run.py:_on_init():2357] communicating current version
23
+ 2024-05-22 11:34:14,843 INFO MainThread:217 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-05-22 11:34:14,844 INFO MainThread:217 [wandb_init.py:init():799] starting run threads in backend
26
+ 2024-05-22 11:34:30,856 INFO MainThread:217 [wandb_run.py:_console_start():2335] atexit reg
27
+ 2024-05-22 11:34:30,857 INFO MainThread:217 [wandb_run.py:_redirect():2190] redirect: wrap_raw
28
+ 2024-05-22 11:34:30,857 INFO MainThread:217 [wandb_run.py:_redirect():2255] Wrapping output streams.
29
+ 2024-05-22 11:34:30,857 INFO MainThread:217 [wandb_run.py:_redirect():2280] Redirects installed.
30
+ 2024-05-22 11:34:30,858 INFO MainThread:217 [wandb_init.py:init():842] run started, returning control to user process
31
+ 2024-05-22 11:34:30,865 INFO MainThread:217 [wandb_run.py:_config_callback():1347] config_cb None None {'peft_config': {'default': {'peft_type': <PeftType.LORA: 'LORA'>, 'auto_mapping': None, 'base_model_name_or_path': 'core42/jais-13b', 'revision': None, 'task_type': 'CAUSAL_LM', 'inference_mode': False, 'r': 16, 'target_modules': {'c_attn'}, 'lora_alpha': 32, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': None, 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'loftq_config': {}, 'use_dora': False, 'layer_replication': None}}, 'vocab_size': 84992, 'n_positions': 2048, 'n_embd': 5120, 'n_layer': 40, 'n_head': 40, 'n_inner': 13653, 'activation_function': 'swiglu', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 0, 'eos_token_id': 0, 'position_embedding_type': 'alibi', 'width_scale': 0.11100000000000002, 'embeddings_scale': 14.6, 'scale_qk_dot_by_d': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['JAISLMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 0, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'core42/jais-13b', 'transformers_version': '4.41.0', 'auto_map': {'AutoConfig': 'core42/jais-13b--configuration_jais.JAISConfig', 'AutoModel': 'core42/jais-13b--modeling_jais.JAISModel', 'AutoModelForCausalLM': 'core42/jais-13b--modeling_jais.JAISLMHeadModel', 'AutoModelForQuestionAnswering': 'core42/jais-13b--modeling_jais.JAISForQuestionAnswering', 'AutoModelForSequenceClassification': 'core42/jais-13b--modeling_jais.JAISForSequenceClassification', 'AutoModelForTokenClassification': 'core42/jais-13b--modeling_jais.JAISForTokenClassification'}, 'model_type': 'jais', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'bfloat16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May22_11-33-56_2c1b614ec68f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
32
+ 2024-05-22 11:34:30,875 INFO MainThread:217 [wandb_config.py:__setitem__():151] config set model/num_parameters = 13033919160 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7ef9227a9060>>
33
+ 2024-05-22 11:34:30,876 INFO MainThread:217 [wandb_run.py:_config_callback():1347] config_cb model/num_parameters 13033919160 None
34
+ 2024-05-22 14:04:41,874 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
35
+ 2024-05-22 14:04:41,875 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
36
+ 2024-05-22 14:14:52,958 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
37
+ 2024-05-22 14:14:54,437 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
38
+ 2024-05-22 14:14:54,437 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
39
+ 2024-05-22 14:15:26,186 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
40
+ 2024-05-22 14:16:25,347 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
41
+ 2024-05-22 14:16:25,348 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
42
+ 2024-05-22 14:16:29,691 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
43
+ 2024-05-22 14:16:44,749 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
44
+ 2024-05-22 14:16:44,749 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
45
+ 2024-05-22 14:23:14,136 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
46
+ 2024-05-22 14:23:16,353 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
47
+ 2024-05-22 14:23:16,353 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
48
+ 2024-05-22 14:26:18,732 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
49
+ 2024-05-22 14:26:19,623 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
50
+ 2024-05-22 14:26:19,624 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
51
+ 2024-05-22 14:34:00,493 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
52
+ 2024-05-22 14:34:00,984 INFO MainThread:217 [jupyter.py:save_ipynb():373] not saving jupyter notebook
53
+ 2024-05-22 14:34:00,984 INFO MainThread:217 [wandb_init.py:_pause_backend():432] pausing backend
54
+ 2024-05-22 14:34:16,410 INFO MainThread:217 [wandb_init.py:_resume_backend():437] resuming backend
wandb/run-20240522_113413-8mudzhjp/run-8mudzhjp.wandb ADDED
Binary file (492 kB). View file