The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: DatasetGenerationError Exception: ArrowNotImplementedError Message: Unsupported cast from double to halffloat using function cast_half_float Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1766, in _prepare_split_single writer.write(example, key) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 500, in write self.write_examples_on_file() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 458, in write_examples_on_file self.write_batch(batch_examples=batch_examples) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 568, in write_batch arrays.append(pa.array(typed_sequence)) File "pyarrow/array.pxi", line 247, in pyarrow.lib.array File "pyarrow/array.pxi", line 112, in pyarrow.lib._handle_arrow_array_protocol File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 208, in __arrow_array__ out = cast_array_to_feature( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature casted_array_values = _c(array.values, feature.feature) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature casted_array_values = _c(array.values, feature.feature) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature casted_array_values = _c(array.values, feature.feature) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2116, in cast_array_to_feature return array_cast( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1963, in array_cast return array.cast(pa_type) File "pyarrow/array.pxi", line 996, in pyarrow.lib.Array.cast File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/compute.py", line 404, in cast return call_function("cast", [arr], options, memory_pool) File "pyarrow/_compute.pyx", line 590, in pyarrow._compute.call_function File "pyarrow/_compute.pyx", line 385, in pyarrow._compute.Function.call File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Unsupported cast from double to halffloat using function cast_half_float During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1775, in _prepare_split_single num_examples, num_bytes = writer.finalize() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 599, in finalize self.write_examples_on_file() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 458, in write_examples_on_file self.write_batch(batch_examples=batch_examples) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 568, in write_batch arrays.append(pa.array(typed_sequence)) File "pyarrow/array.pxi", line 247, in pyarrow.lib.array File "pyarrow/array.pxi", line 112, in pyarrow.lib._handle_arrow_array_protocol File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 208, in __arrow_array__ out = cast_array_to_feature( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature casted_array_values = _c(array.values, feature.feature) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature casted_array_values = _c(array.values, feature.feature) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature casted_array_values = _c(array.values, feature.feature) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2116, in cast_array_to_feature return array_cast( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1963, in array_cast return array.cast(pa_type) File "pyarrow/array.pxi", line 996, in pyarrow.lib.Array.cast File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/compute.py", line 404, in cast return call_function("cast", [arr], options, memory_pool) File "pyarrow/_compute.pyx", line 590, in pyarrow._compute.call_function File "pyarrow/_compute.pyx", line 385, in pyarrow._compute.Function.call File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Unsupported cast from double to halffloat using function cast_half_float The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1524, in compute_config_parquet_and_info_response parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet( File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1099, in stream_convert_to_parquet builder._prepare_split( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1627, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1784, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
__key__
string | __url__
string | img.png
image | latent.pth
sequence |
---|---|---|---|
0 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[0.56640625,1.4072265625,0.38525390625,1.400390625,3.142578125,1.998046875,-1.7041015625,1.1953125(...TRUNCATED) |
|
1 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[-17.703125,-19.6875,-16.140625,-20.765625,-14.4453125,-19.9375,-19.3125,-18.234375,-17.0625,-18.4(...TRUNCATED) |
|
2 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[5.99609375,3.982421875,5.78125,3.6171875,6.734375,4.03125,1.73046875,5.59765625,3.7421875,2.70312(...TRUNCATED) |
|
3 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[3.908203125,3.609375,4.52734375,0.8994140625,6.8125,4.09765625,0.56005859375,5.35546875,4.4765625(...TRUNCATED) |
|
4 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[19.0,16.109375,17.140625,18.65625,15.2421875,22.703125,19.6875,18.375,17.25,21.8125,20.5625,15.89(...TRUNCATED) |
|
5 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[-11.5859375,-13.578125,-15.2890625,-12.6484375,-16.9375,-12.7109375,-12.1875,-13.5390625,-11.1328(...TRUNCATED) |
|
6 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[13.1953125,11.1484375,9.3515625,12.1328125,8.0390625,9.90625,11.6875,12.265625,12.5078125,10.25,1(...TRUNCATED) |
|
7 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[-11.2890625,-14.796875,-15.390625,-6.09765625,10.8671875,4.55859375,0.137451171875,1.25,3.828125,(...TRUNCATED) |
|
8 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[6.59375,6.22265625,-0.6640625,-1.6787109375,-7.44921875,-7.93359375,-3.994140625,-2.41796875,-3.7(...TRUNCATED) |
|
9 | hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar | [[[-1.9423828125,-1.6826171875,5.3671875,0.46044921875,6.59375,9.890625,4.09375,-0.3984375,4.1210937(...TRUNCATED) |
https://github.com/NVlabs/ffhq-dataset samples encoded to float16 SDXL latents via Ollin VAE.
Dataset created using this script.
VAE encoder used NATTEN attention, kernel size 17.
Didn't bother saving mean & logvar, because variance is low enough it's not worth the doubling of filesize to retain.
Sampled from diagonal gaussian distribution, saved the resulting latents.
Also kept the original image.
Schema/usage:
from typing import TypedDict, Iterator
from webdataset import WebDataset
Sample = TypedDict('Sample', {
'__key__': str,
'__url__': str,
'img.png': bytes, # PIL image, serialized. 1024*1024px
'latent.pth': bytes, # FloatTensor, serialized. 128*128 latents
})
it: Iterator[Sample] = WebDataset('{00000..00035}.tar')
for sample in it:
pass
# avg/val.pt (mean):
[-2.8982300758361816, -0.9609659910202026, 0.2416578084230423, -0.307400107383728]
# avg/sq.pt:
[65.80902099609375, 32.772762298583984, 36.080204010009766, 25.072017669677734]
# std
# (sq - val**2)**.5
[7.5768914222717285, 5.643518924713135, 6.001816749572754, 4.997751712799072]
# 1/std
[0.13198024034500122, 0.17719440162181854, 0.16661621630191803, 0.2000899761915207]
Flickr-Faces-HQ Dataset (FFHQ)
Flickr-Faces-HQ (FFHQ) is a high-quality image dataset of human faces, originally created as a benchmark for generative adversarial networks (GAN):
A Style-Based Generator Architecture for Generative Adversarial Networks
Tero Karras (NVIDIA), Samuli Laine (NVIDIA), Timo Aila (NVIDIA)
https://arxiv.org/abs/1812.04948
The dataset consists of 70,000 high-quality PNG images at 1024×1024 resolution and contains considerable variation in terms of age, ethnicity and image background. It also has good coverage of accessories such as eyeglasses, sunglasses, hats, etc. The images were crawled from Flickr, thus inheriting all the biases of that website, and automatically aligned and cropped using dlib. Only images under permissive licenses were collected. Various automatic filters were used to prune the set, and finally Amazon Mechanical Turk was used to remove the occasional statues, paintings, or photos of photos.
Please note that this dataset is not intended for, and should not be used for, development or improvement of facial recognition technologies. For business inquiries, please visit our website and submit the form: NVIDIA Research Licensing
Licenses
The individual images were published in Flickr by their respective authors under either Creative Commons BY 2.0, Creative Commons BY-NC 2.0, Public Domain Mark 1.0, Public Domain CC0 1.0, or U.S. Government Works license. All of these licenses allow free use, redistribution, and adaptation for non-commercial purposes. However, some of them require giving appropriate credit to the original author, as well as indicating any changes that were made to the images. The license and original author of each image are indicated in the metadata.
- https://creativecommons.org/licenses/by/2.0/
- https://creativecommons.org/licenses/by-nc/2.0/
- https://creativecommons.org/publicdomain/mark/1.0/
- https://creativecommons.org/publicdomain/zero/1.0/
- http://www.usa.gov/copyright.shtml
The dataset itself (including JSON metadata, download script, and documentation) is made available under Creative Commons BY-NC-SA 4.0 license by NVIDIA Corporation. You can use, redistribute, and adapt it for non-commercial purposes, as long as you (a) give appropriate credit by citing our paper, (b) indicate any changes that you've made, and (c) distribute any derivative works under the same license.
- Downloads last month
- 33