lysandre HF staff commited on
Commit
84a6833
1 Parent(s): e1377a1

Upload dataset

Browse files
README.md ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ dataset_info:
3
+ features:
4
+ - name: dates
5
+ dtype: string
6
+ - name: type
7
+ struct:
8
+ - name: authorAssociation
9
+ dtype: string
10
+ - name: comment
11
+ dtype: bool
12
+ - name: issue
13
+ dtype: bool
14
+ splits:
15
+ - name: transformers
16
+ num_bytes: 5086147
17
+ num_examples: 143220
18
+ - name: peft
19
+ num_bytes: 284778
20
+ num_examples: 8297
21
+ - name: evaluate
22
+ num_bytes: 67751
23
+ num_examples: 1935
24
+ - name: huggingface_hub
25
+ num_bytes: 325913
26
+ num_examples: 9329
27
+ - name: accelerate
28
+ num_bytes: 414323
29
+ num_examples: 11849
30
+ - name: datasets
31
+ num_bytes: 863318
32
+ num_examples: 24324
33
+ - name: optimum
34
+ num_bytes: 212650
35
+ num_examples: 6130
36
+ - name: pytorch_image_models
37
+ num_bytes: 150547
38
+ num_examples: 4363
39
+ - name: gradio
40
+ num_bytes: 1269944
41
+ num_examples: 34989
42
+ - name: tokenizers
43
+ num_bytes: 208563
44
+ num_examples: 6098
45
+ - name: diffusers
46
+ num_bytes: 1574248
47
+ num_examples: 43895
48
+ - name: safetensors
49
+ num_bytes: 51932
50
+ num_examples: 1506
51
+ - name: sentence_transformers
52
+ num_bytes: 325153
53
+ num_examples: 9513
54
+ - name: candle
55
+ num_bytes: 203878
56
+ num_examples: 5364
57
+ - name: text_generation_inference
58
+ num_bytes: 241200
59
+ num_examples: 7115
60
+ - name: chat_ui
61
+ num_bytes: 109454
62
+ num_examples: 3142
63
+ - name: hub_docs
64
+ num_bytes: 143368
65
+ num_examples: 4073
66
+ download_size: 3582852
67
+ dataset_size: 11533167
68
+ configs:
69
+ - config_name: default
70
+ data_files:
71
+ - split: transformers
72
+ path: data/transformers-*
73
+ - split: peft
74
+ path: data/peft-*
75
+ - split: evaluate
76
+ path: data/evaluate-*
77
+ - split: huggingface_hub
78
+ path: data/huggingface_hub-*
79
+ - split: accelerate
80
+ path: data/accelerate-*
81
+ - split: datasets
82
+ path: data/datasets-*
83
+ - split: optimum
84
+ path: data/optimum-*
85
+ - split: pytorch_image_models
86
+ path: data/pytorch_image_models-*
87
+ - split: gradio
88
+ path: data/gradio-*
89
+ - split: tokenizers
90
+ path: data/tokenizers-*
91
+ - split: diffusers
92
+ path: data/diffusers-*
93
+ - split: safetensors
94
+ path: data/safetensors-*
95
+ - split: sentence_transformers
96
+ path: data/sentence_transformers-*
97
+ - split: candle
98
+ path: data/candle-*
99
+ - split: text_generation_inference
100
+ path: data/text_generation_inference-*
101
+ - split: chat_ui
102
+ path: data/chat_ui-*
103
+ - split: hub_docs
104
+ path: data/hub_docs-*
105
+ ---
data/accelerate-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:410cfae6618736fd92bf2e031d92f999903ce0d0376779a77066e9a2570eb0d2
3
+ size 136218
data/candle-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9a5fc830e8c605bce2325b8ef072387cee786e45c5cfe710a11b3844200c38f
3
+ size 64119
data/chat_ui-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f73dc05854ecad2a2cbf9bdeda0b118b73d1bc9c79c31d8c913a976cfac67bbc
3
+ size 38852
data/datasets-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f5ffa110185f7748616b59b8824237ef29a85020a997e017d77c343625cf394
3
+ size 277838
data/diffusers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:656d41251c7bf91cf2443e233865fec501e6eb6c68200c713e035bc91ab0ec9f
3
+ size 464498
data/evaluate-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9d936eb4103478346a994565e581c9172b3fdb68721bd98ba9819601a82323a
3
+ size 24775
data/gradio-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d5a50e5c2241b0fba7c9c7cdf0b60927acb63b6dbb136425a63eb8f1af59402
3
+ size 379658
data/hub_docs-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cce38d2a1c33131ff9cf33649f0687e13fee703d9f208a69286629cb8f503ce5
3
+ size 50075
data/huggingface_hub-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95a5857b0d1c823eca9943f663a65eefd06e0e050e275ab86c5cc96312f72d1f
3
+ size 109193
data/optimum-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8240301ad5b02e04e0a141547dfd7e684da7dd1f7022998a2292afa39347cb51
3
+ size 74125
data/peft-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ab8cc20e1a7d10c8045f6314eeb21e3b5074ef637ac667d786232080ee084a6
3
+ size 94399
data/pytorch_image_models-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02c35e308890186009479b079d88277031c354afc243c9ebafde7b24a8ebda72
3
+ size 56224
data/safetensors-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b86675357f503cd37556f762f6566b7634e1d6390344672d23b21df652c7698
3
+ size 19942
data/sentence_transformers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e722ce907873b3f7046b56d82961b465017b1bd9f9cb00baa6a781ca7669e09
3
+ size 114944
data/text_generation_inference-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2612d62950353c2ba2ebb0b6c8cd5610d5da19806897971c6f40d1e79ae11830
3
+ size 82290
data/tokenizers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d668ac42af03688fcfae9d2f8d9aa69a5bc83bde8d5cef674799f4ae0ec8e44b
3
+ size 75827
data/transformers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcbefe65f2718e46f91765f535f0eb8e7d387951b3c576b81c47e9c5e96c7329
3
+ size 1519875
dataset_infos.json CHANGED
@@ -28,109 +28,109 @@
28
  "splits": {
29
  "transformers": {
30
  "name": "transformers",
31
- "num_bytes": 4970577,
32
- "num_examples": 140143,
33
  "dataset_name": null
34
  },
35
  "peft": {
36
  "name": "peft",
37
- "num_bytes": 267891,
38
- "num_examples": 7813,
39
  "dataset_name": null
40
  },
41
  "evaluate": {
42
  "name": "evaluate",
43
- "num_bytes": 65695,
44
- "num_examples": 1877,
45
  "dataset_name": null
46
  },
47
  "huggingface_hub": {
48
  "name": "huggingface_hub",
49
- "num_bytes": 316054,
50
- "num_examples": 9053,
51
  "dataset_name": null
52
  },
53
  "accelerate": {
54
  "name": "accelerate",
55
- "num_bytes": 398411,
56
- "num_examples": 11393,
57
  "dataset_name": null
58
  },
59
  "datasets": {
60
  "name": "datasets",
61
- "num_bytes": 843933,
62
- "num_examples": 24046,
63
  "dataset_name": null
64
  },
65
  "optimum": {
66
  "name": "optimum",
67
- "num_bytes": 209726,
68
- "num_examples": 6044,
69
  "dataset_name": null
70
  },
71
  "pytorch_image_models": {
72
  "name": "pytorch_image_models",
73
- "num_bytes": 148186,
74
- "num_examples": 4294,
75
  "dataset_name": null
76
  },
77
  "gradio": {
78
  "name": "gradio",
79
- "num_bytes": 1224430,
80
- "num_examples": 33731,
81
  "dataset_name": null
82
  },
83
  "tokenizers": {
84
  "name": "tokenizers",
85
- "num_bytes": 203872,
86
- "num_examples": 5956,
87
  "dataset_name": null
88
  },
89
  "diffusers": {
90
  "name": "diffusers",
91
- "num_bytes": 1519391,
92
- "num_examples": 42441,
93
  "dataset_name": null
94
  },
95
  "safetensors": {
96
  "name": "safetensors",
97
- "num_bytes": 51043,
98
- "num_examples": 1478,
99
  "dataset_name": null
100
  },
101
  "sentence_transformers": {
102
  "name": "sentence_transformers",
103
- "num_bytes": 317985,
104
- "num_examples": 9325,
105
  "dataset_name": null
106
  },
107
  "candle": {
108
  "name": "candle",
109
- "num_bytes": 189838,
110
- "num_examples": 4997,
111
  "dataset_name": null
112
  },
113
  "text_generation_inference": {
114
  "name": "text_generation_inference",
115
- "num_bytes": 226337,
116
- "num_examples": 6685,
117
  "dataset_name": null
118
  },
119
  "chat_ui": {
120
  "name": "chat_ui",
121
- "num_bytes": 99559,
122
- "num_examples": 2864,
123
  "dataset_name": null
124
  },
125
  "hub_docs": {
126
  "name": "hub_docs",
127
- "num_bytes": 142244,
128
- "num_examples": 4042,
129
  "dataset_name": null
130
  }
131
  },
132
- "download_size": 3481791,
133
- "dataset_size": 11195172,
134
- "size_in_bytes": 14676963
135
  }
136
  }
 
28
  "splits": {
29
  "transformers": {
30
  "name": "transformers",
31
+ "num_bytes": 5086147,
32
+ "num_examples": 143220,
33
  "dataset_name": null
34
  },
35
  "peft": {
36
  "name": "peft",
37
+ "num_bytes": 284778,
38
+ "num_examples": 8297,
39
  "dataset_name": null
40
  },
41
  "evaluate": {
42
  "name": "evaluate",
43
+ "num_bytes": 67751,
44
+ "num_examples": 1935,
45
  "dataset_name": null
46
  },
47
  "huggingface_hub": {
48
  "name": "huggingface_hub",
49
+ "num_bytes": 325913,
50
+ "num_examples": 9329,
51
  "dataset_name": null
52
  },
53
  "accelerate": {
54
  "name": "accelerate",
55
+ "num_bytes": 414323,
56
+ "num_examples": 11849,
57
  "dataset_name": null
58
  },
59
  "datasets": {
60
  "name": "datasets",
61
+ "num_bytes": 863318,
62
+ "num_examples": 24324,
63
  "dataset_name": null
64
  },
65
  "optimum": {
66
  "name": "optimum",
67
+ "num_bytes": 212650,
68
+ "num_examples": 6130,
69
  "dataset_name": null
70
  },
71
  "pytorch_image_models": {
72
  "name": "pytorch_image_models",
73
+ "num_bytes": 150547,
74
+ "num_examples": 4363,
75
  "dataset_name": null
76
  },
77
  "gradio": {
78
  "name": "gradio",
79
+ "num_bytes": 1269944,
80
+ "num_examples": 34989,
81
  "dataset_name": null
82
  },
83
  "tokenizers": {
84
  "name": "tokenizers",
85
+ "num_bytes": 208563,
86
+ "num_examples": 6098,
87
  "dataset_name": null
88
  },
89
  "diffusers": {
90
  "name": "diffusers",
91
+ "num_bytes": 1574248,
92
+ "num_examples": 43895,
93
  "dataset_name": null
94
  },
95
  "safetensors": {
96
  "name": "safetensors",
97
+ "num_bytes": 51932,
98
+ "num_examples": 1506,
99
  "dataset_name": null
100
  },
101
  "sentence_transformers": {
102
  "name": "sentence_transformers",
103
+ "num_bytes": 325153,
104
+ "num_examples": 9513,
105
  "dataset_name": null
106
  },
107
  "candle": {
108
  "name": "candle",
109
+ "num_bytes": 203878,
110
+ "num_examples": 5364,
111
  "dataset_name": null
112
  },
113
  "text_generation_inference": {
114
  "name": "text_generation_inference",
115
+ "num_bytes": 241200,
116
+ "num_examples": 7115,
117
  "dataset_name": null
118
  },
119
  "chat_ui": {
120
  "name": "chat_ui",
121
+ "num_bytes": 109454,
122
+ "num_examples": 3142,
123
  "dataset_name": null
124
  },
125
  "hub_docs": {
126
  "name": "hub_docs",
127
+ "num_bytes": 143368,
128
+ "num_examples": 4073,
129
  "dataset_name": null
130
  }
131
  },
132
+ "download_size": 3582852,
133
+ "dataset_size": 11533167,
134
+ "size_in_bytes": 15116019
135
  }
136
  }