joey00072 commited on
Commit
8aa5e3b
1 Parent(s): 0143254

checkpoint iter: 2000

Browse files
Files changed (2) hide show
  1. config.json +4 -4
  2. model.safetensors +2 -2
config.json CHANGED
@@ -6,12 +6,12 @@
6
  "dropout": 0.2,
7
  "head_dim": 64,
8
  "hidden_dim": 1536,
9
- "kv_lora_rank": 64,
10
  "mlp": "GLU",
11
- "num_heads": 46,
12
- "num_kv_heads": 46,
13
  "num_layers": 4,
14
- "q_lora_rank": 192,
15
  "rope_head_dim": 16,
16
  "seq_len": 256,
17
  "vocab_size": 50257,
 
6
  "dropout": 0.2,
7
  "head_dim": 64,
8
  "hidden_dim": 1536,
9
+ "kv_lora_rank": 128,
10
  "mlp": "GLU",
11
+ "num_heads": 32,
12
+ "num_kv_heads": 32,
13
  "num_layers": 4,
14
+ "q_lora_rank": 384,
15
  "rope_head_dim": 16,
16
  "seq_len": 256,
17
  "vocab_size": 50257,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f0b81b4e8bd1a01b4825d2d3dfe324ad7018a19c9aed95ad49aa0c5342ec2a2
3
- size 554294048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd29eac3a78e43c10426aef42f84e4babfc61950a18d4c255e0bf2202e329b75
3
+ size 549415728