Datasets:
OALL
/

Modalities:
Text
Formats:
json
Size:
< 1K
Libraries:
Datasets
Dask

[bot] Conversion to Parquet

#1
by parquet-converter - opened

The parquet-converter bot has created a version of this dataset in the Parquet format in the refs/convert/parquet branch.

What is Parquet?

Apache Parquet is a popular columnar storage format known for:

  • reduced memory requirement,
  • fast data retrieval and filtering,
  • efficient storage.

This is what powers the dataset viewer on each dataset page and every dataset on the Hub can be accessed with the same code (you can use HF Datasets, ClickHouse, DuckDB, Pandas or Polars, up to you).

You can learn more about the advantages associated with Parquet in the documentation.

How to access the Parquet version of the dataset?

You can access the Parquet version of the dataset by following this link: refs/convert/parquet

What if my dataset was already in Parquet?

When the dataset is already in Parquet format, the data are not converted and the files in refs/convert/parquet are links to the original files. This rule has an exception to ensure the dataset viewer API to stay fast: if the row group size of the original Parquet files is too big, new Parquet files are generated.

What should I do?

You don't need to do anything. The Parquet version of the dataset is available for you to use. Refer to the documentation for examples and code snippets on how to query the Parquet files with ClickHouse, DuckDB, Pandas or Polars.

If you have any questions or concerns, feel free to ask in the discussion below. You can also close the discussion if you don't have any questions.

Sign up or log in to comment