Upload folder using huggingface_hub
Browse files- config.json +2 -2
- generation_config.json +1 -1
- optimizer.pt +2 -2
- pytorch_model.bin +2 -2
- rng_state_0.pth +3 -0
- rng_state_1.pth +3 -0
- rng_state_2.pth +3 -0
- rng_state_3.pth +3 -0
- scheduler.pt +2 -2
- tokenizer.json +0 -0
- tokenizer_config.json +0 -1
- trainer_state.json +0 -0
- training_args.bin +2 -2
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
@@ -17,7 +17,7 @@
|
|
17 |
"rms_norm_eps": 1e-06,
|
18 |
"tie_word_embeddings": false,
|
19 |
"torch_dtype": "float32",
|
20 |
-
"transformers_version": "4.
|
21 |
"use_cache": true,
|
22 |
"vocab_size": 32000
|
23 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "./llama-pretrain",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
17 |
"rms_norm_eps": 1e-06,
|
18 |
"tie_word_embeddings": false,
|
19 |
"torch_dtype": "float32",
|
20 |
+
"transformers_version": "4.30.0.dev0",
|
21 |
"use_cache": true,
|
22 |
"vocab_size": 32000
|
23 |
}
|
generation_config.json
CHANGED
@@ -3,5 +3,5 @@
|
|
3 |
"bos_token_id": 0,
|
4 |
"eos_token_id": 2,
|
5 |
"pad_token_id": 1,
|
6 |
-
"transformers_version": "4.
|
7 |
}
|
|
|
3 |
"bos_token_id": 0,
|
4 |
"eos_token_id": 2,
|
5 |
"pad_token_id": 1,
|
6 |
+
"transformers_version": "4.30.0.dev0"
|
7 |
}
|
optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4890d1bdb534abd260b5e06f55f51b92115a7f4667b04eaeae1e28ac782cec9a
|
3 |
+
size 1299405253
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:235fa55ab9fddf5d082ac3244d0dbb3ce5a96805375c687565ca004c944efc63
|
3 |
+
size 649712185
|
rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1eafe3d5e0585dde8c5033613de99a5d4f23df4284a488f4007b3944580c0b97
|
3 |
+
size 17655
|
rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7e34eb456d2d003a2839f2daa9425e99bdd79ed7e24a1de9fc7d5738476bfb4b
|
3 |
+
size 17655
|
rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b374af4a2765d8771cee7a72921d3c2e438b9bee34f0b2d098ce6071afeb65e4
|
3 |
+
size 17655
|
rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5df75d8477fcc69c7abb03025313915ebfe3ac18c54a7c57aaa455c0099e13e5
|
3 |
+
size 17655
|
scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b31a300187e9e55755b8b7d0cd02a8f1e13936613e9136ec2482e6080bad5dfc
|
3 |
+
size 627
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -21,7 +21,6 @@
|
|
21 |
"model_max_length": 1000000000000000019884624838656,
|
22 |
"pad_token": null,
|
23 |
"sp_model_kwargs": {},
|
24 |
-
"special_tokens_map_file": "/home/zhihaoz3/.cache/huggingface/hub/models--decapoda-research--llama-7b-hf/snapshots/5f98eefcc80e437ef68d457ad7bf167c2c6a1348/special_tokens_map.json",
|
25 |
"tokenizer_class": "LlamaTokenizer",
|
26 |
"unk_token": {
|
27 |
"__type": "AddedToken",
|
|
|
21 |
"model_max_length": 1000000000000000019884624838656,
|
22 |
"pad_token": null,
|
23 |
"sp_model_kwargs": {},
|
|
|
24 |
"tokenizer_class": "LlamaTokenizer",
|
25 |
"unk_token": {
|
26 |
"__type": "AddedToken",
|
trainer_state.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee0f608ac46a4890c5ce839a702c158151f6c98802ee3c7a30c6b8995e4efd88
|
3 |
+
size 3899
|