anilbhatt1 commited on
Commit
4a3a82e
1 Parent(s): dbef0db

Upload pythia-410m-deduped files

Browse files
checkpoints/EleutherAI/pythia-410m-deduped/lit_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"name": "pythia-410m-deduped", "hf_config": {"org": "EleutherAI", "name": "pythia-410m-deduped"}, "block_size": 2048, "vocab_size": 50254, "padding_multiple": 128, "padded_vocab_size": 50304, "n_layer": 24, "n_head": 16, "n_embd": 1024, "rotary_percentage": 0.25, "parallel_residual": true, "bias": true, "lm_head_bias": false, "n_query_groups": 16, "shared_attention_norm": false, "_norm_class": "LayerNorm", "norm_eps": 1e-05, "_mlp_class": "GptNeoxMLP", "gelu_approximate": "none", "intermediate_size": 4096, "rope_condense_ratio": 1, "rope_base": 10000}
checkpoints/EleutherAI/pythia-410m-deduped/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b8f211c55958bf39c52d8a5db2b208481beddc09673d614d2d9059a882d3e11
3
+ size 911449213
checkpoints/EleutherAI/pythia-410m-deduped/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/EleutherAI/pythia-410m-deduped/tokenizer_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<|endoftext|>",
4
+ "eos_token": "<|endoftext|>",
5
+ "name_or_path": "EleutherAI/gpt-neox-20b",
6
+ "special_tokens_map_file": "/admin/home-hailey/.cache/huggingface/hub/models--EleutherAI--gpt-neox-20b/snapshots/4e49eadb5d14bd22f314ec3f45b69a87b88c7691/special_tokens_map.json",
7
+ "tokenizer_class": "GPTNeoXTokenizer",
8
+ "unk_token": "<|endoftext|>"
9
+ }