Baicai003 commited on
Commit
a76759f
1 Parent(s): a7ee08c

Upload tiny-clip/model.safetensors.index.json with huggingface_hub

Browse files
tiny-clip/model.safetensors.index.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 4227180
4
+ },
5
+ "weight_map": {
6
+ "logit_scale": "model-00001-of-00009.safetensors",
7
+ "text_model.embeddings.position_embedding.weight": "model-00001-of-00009.safetensors",
8
+ "text_model.embeddings.token_embedding.weight": "model-00001-of-00009.safetensors",
9
+ "text_model.encoder.layers.0.layer_norm1.bias": "model-00009-of-00009.safetensors",
10
+ "text_model.encoder.layers.0.layer_norm1.weight": "model-00009-of-00009.safetensors",
11
+ "text_model.encoder.layers.0.layer_norm2.bias": "model-00009-of-00009.safetensors",
12
+ "text_model.encoder.layers.0.layer_norm2.weight": "model-00009-of-00009.safetensors",
13
+ "text_model.encoder.layers.0.mlp.fc1.bias": "model-00009-of-00009.safetensors",
14
+ "text_model.encoder.layers.0.mlp.fc1.weight": "model-00009-of-00009.safetensors",
15
+ "text_model.encoder.layers.0.mlp.fc2.bias": "model-00009-of-00009.safetensors",
16
+ "text_model.encoder.layers.0.mlp.fc2.weight": "model-00009-of-00009.safetensors",
17
+ "text_model.encoder.layers.0.self_attn.k_proj.bias": "model-00003-of-00009.safetensors",
18
+ "text_model.encoder.layers.0.self_attn.k_proj.weight": "model-00002-of-00009.safetensors",
19
+ "text_model.encoder.layers.0.self_attn.out_proj.bias": "model-00009-of-00009.safetensors",
20
+ "text_model.encoder.layers.0.self_attn.out_proj.weight": "model-00008-of-00009.safetensors",
21
+ "text_model.encoder.layers.0.self_attn.q_proj.bias": "model-00007-of-00009.safetensors",
22
+ "text_model.encoder.layers.0.self_attn.q_proj.weight": "model-00006-of-00009.safetensors",
23
+ "text_model.encoder.layers.0.self_attn.v_proj.bias": "model-00005-of-00009.safetensors",
24
+ "text_model.encoder.layers.0.self_attn.v_proj.weight": "model-00004-of-00009.safetensors",
25
+ "text_model.final_layer_norm.bias": "model-00009-of-00009.safetensors",
26
+ "text_model.final_layer_norm.weight": "model-00009-of-00009.safetensors",
27
+ "text_projection.weight": "model-00009-of-00009.safetensors",
28
+ "vision_model.embeddings.class_embedding": "model-00009-of-00009.safetensors",
29
+ "vision_model.embeddings.patch_embedding.weight": "model-00009-of-00009.safetensors",
30
+ "vision_model.embeddings.position_embedding.weight": "model-00009-of-00009.safetensors",
31
+ "vision_model.encoder.layers.0.layer_norm1.bias": "model-00009-of-00009.safetensors",
32
+ "vision_model.encoder.layers.0.layer_norm1.weight": "model-00009-of-00009.safetensors",
33
+ "vision_model.encoder.layers.0.layer_norm2.bias": "model-00009-of-00009.safetensors",
34
+ "vision_model.encoder.layers.0.layer_norm2.weight": "model-00009-of-00009.safetensors",
35
+ "vision_model.encoder.layers.0.mlp.fc1.bias": "model-00009-of-00009.safetensors",
36
+ "vision_model.encoder.layers.0.mlp.fc1.weight": "model-00009-of-00009.safetensors",
37
+ "vision_model.encoder.layers.0.mlp.fc2.bias": "model-00009-of-00009.safetensors",
38
+ "vision_model.encoder.layers.0.mlp.fc2.weight": "model-00009-of-00009.safetensors",
39
+ "vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00009-of-00009.safetensors",
40
+ "vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00009-of-00009.safetensors",
41
+ "vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00009-of-00009.safetensors",
42
+ "vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00009-of-00009.safetensors",
43
+ "vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00009-of-00009.safetensors",
44
+ "vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00009-of-00009.safetensors",
45
+ "vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00009-of-00009.safetensors",
46
+ "vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00009-of-00009.safetensors",
47
+ "vision_model.post_layernorm.bias": "model-00009-of-00009.safetensors",
48
+ "vision_model.post_layernorm.weight": "model-00009-of-00009.safetensors",
49
+ "vision_model.pre_layrnorm.bias": "model-00009-of-00009.safetensors",
50
+ "vision_model.pre_layrnorm.weight": "model-00009-of-00009.safetensors",
51
+ "visual_projection.weight": "model-00009-of-00009.safetensors"
52
+ }
53
+ }