davzoku commited on
Commit
9067824
1 Parent(s): 3dcafdc

Convert dataset to Parquet

Browse files
README.md CHANGED
@@ -33,13 +33,20 @@ dataset_info:
33
  '3': Sci/Tech
34
  splits:
35
  - name: train
36
- num_bytes: 29817351
37
  num_examples: 120000
38
  - name: test
39
- num_bytes: 1879478
40
  num_examples: 7600
41
- download_size: 31327765
42
- dataset_size: 31696829
 
 
 
 
 
 
 
43
  train-eval-index:
44
  - config: default
45
  task: text-classification
 
33
  '3': Sci/Tech
34
  splits:
35
  - name: train
36
+ num_bytes: 29817303
37
  num_examples: 120000
38
  - name: test
39
+ num_bytes: 1879474
40
  num_examples: 7600
41
+ download_size: 19820267
42
+ dataset_size: 31696777
43
+ configs:
44
+ - config_name: default
45
+ data_files:
46
+ - split: train
47
+ path: data/train-*
48
+ - split: test
49
+ path: data/test-*
50
  train-eval-index:
51
  - config: default
52
  task: text-classification
data/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71de87ec66bc5737752a2502204dfa6d7fe9856ade3ea444dc6317789a4f13fb
3
+ size 1234829
data/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc508d6d9868594e3da960a8cfeb63ab5a4746598b93428c224397080c1f52ee
3
+ size 18585438
dataset_infos.json CHANGED
@@ -1 +1,55 @@
1
- {"default": {"description": "AG is a collection of more than 1 million news articles. News articles have been\ngathered from more than 2000 news sources by ComeToMyHead in more than 1 year of\nactivity. ComeToMyHead is an academic news search engine which has been running\nsince July, 2004. The dataset is provided by the academic comunity for research\npurposes in data mining (clustering, classification, etc), information retrieval\n(ranking, search, etc), xml, data compression, data streaming, and any other\nnon-commercial activity. For more information, please refer to the link\nhttp://www.di.unipi.it/~gulli/AG_corpus_of_news_articles.html .\n\nThe AG's news topic classification dataset is constructed by Xiang Zhang\n([email protected]) from the dataset above. It is used as a text\nclassification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann\nLeCun. Character-level Convolutional Networks for Text Classification. Advances\nin Neural Information Processing Systems 28 (NIPS 2015).\n", "citation": "@inproceedings{Zhang2015CharacterlevelCN,\n title={Character-level Convolutional Networks for Text Classification},\n author={Xiang Zhang and Junbo Jake Zhao and Yann LeCun},\n booktitle={NIPS},\n year={2015}\n}\n", "homepage": "http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 4, "names": ["World", "Sports", "Business", "Sci/Tech"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": [{"task": "text-classification", "text_column": "text", "label_column": "label", "labels": ["Business", "Sci/Tech", "Sports", "World"]}], "builder_name": "ag_news", "config_name": "default", "version": {"version_str": "0.0.0", "description": null, "major": 0, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 29817351, "num_examples": 120000, "dataset_name": "ag_news"}, "test": {"name": "test", "num_bytes": 1879478, "num_examples": 7600, "dataset_name": "ag_news"}}, "download_checksums": {"https://raw.githubusercontent.com/mhjabreel/CharCnn_Keras/master/data/ag_news_csv/train.csv": {"num_bytes": 29470338, "checksum": "76a0a2d2f92b286371fe4d4044640910a04a803fdd2538e0f3f29a5c6f6b672e"}, "https://raw.githubusercontent.com/mhjabreel/CharCnn_Keras/master/data/ag_news_csv/test.csv": {"num_bytes": 1857427, "checksum": "521465c2428ed7f02f8d6db6ffdd4b5447c1c701962353eb2c40d548c3c85699"}}, "download_size": 31327765, "post_processing_size": null, "dataset_size": 31696829, "size_in_bytes": 63024594}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "default": {
3
+ "description": "AG is a collection of more than 1 million news articles. News articles have been\ngathered from more than 2000 news sources by ComeToMyHead in more than 1 year of\nactivity. ComeToMyHead is an academic news search engine which has been running\nsince July, 2004. The dataset is provided by the academic comunity for research\npurposes in data mining (clustering, classification, etc), information retrieval\n(ranking, search, etc), xml, data compression, data streaming, and any other\nnon-commercial activity. For more information, please refer to the link\nhttp://www.di.unipi.it/~gulli/AG_corpus_of_news_articles.html .\n\nThe AG's news topic classification dataset is constructed by Xiang Zhang\n([email protected]) from the dataset above. It is used as a text\nclassification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann\nLeCun. Character-level Convolutional Networks for Text Classification. Advances\nin Neural Information Processing Systems 28 (NIPS 2015).\n",
4
+ "citation": "@inproceedings{Zhang2015CharacterlevelCN,\n title={Character-level Convolutional Networks for Text Classification},\n author={Xiang Zhang and Junbo Jake Zhao and Yann LeCun},\n booktitle={NIPS},\n year={2015}\n}\n",
5
+ "homepage": "http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html",
6
+ "license": "",
7
+ "features": {
8
+ "text": {
9
+ "dtype": "string",
10
+ "_type": "Value"
11
+ },
12
+ "label": {
13
+ "names": [
14
+ "World",
15
+ "Sports",
16
+ "Business",
17
+ "Sci/Tech"
18
+ ],
19
+ "_type": "ClassLabel"
20
+ }
21
+ },
22
+ "task_templates": [
23
+ {
24
+ "task": "text-classification",
25
+ "label_column": "label"
26
+ }
27
+ ],
28
+ "builder_name": "parquet",
29
+ "dataset_name": "ag_news",
30
+ "config_name": "default",
31
+ "version": {
32
+ "version_str": "0.0.0",
33
+ "major": 0,
34
+ "minor": 0,
35
+ "patch": 0
36
+ },
37
+ "splits": {
38
+ "train": {
39
+ "name": "train",
40
+ "num_bytes": 29817303,
41
+ "num_examples": 120000,
42
+ "dataset_name": null
43
+ },
44
+ "test": {
45
+ "name": "test",
46
+ "num_bytes": 1879474,
47
+ "num_examples": 7600,
48
+ "dataset_name": null
49
+ }
50
+ },
51
+ "download_size": 19820267,
52
+ "dataset_size": 31696777,
53
+ "size_in_bytes": 51517044
54
+ }
55
+ }