Datasets:
File size: 4,780 Bytes
7587a18 c5b40df 5da2a8e c5b40df 7bef39f 8664327 7587a18 ba0e0d8 2291fc1 4a6f314 2291fc1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 |
---
annotations_creators:
- crowdsourced
license: cc-by-nc-sa-4.0
size_categories: []
task_categories:
- image-classification
- image-feature-extraction
pretty_name: Galaxy Zoo CANDELS
arxiv: 2404.02973
tags:
- galaxy zoo
- physics
- astronomy
- galaxies
- citizen science
---
# GZ Campaign Datasets
## Dataset Summary
[Galaxy Zoo](www.galaxyzoo.org) volunteers label telescope images of galaxies according to their visible features: spiral arms, galaxy-galaxy collisions, and so on.
These datasets share the galaxy images and volunteer labels in a machine-learning-friendly format. We use these datasets to train [our foundation models](https://arxiv.org/abs/2404.02973). We hope they'll help you too.
- **Curated by:** [Mike Walmsley](https://walmsley.dev/)
- **License:** [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/deed.en). We specifically require **all models trained on these datasets to be released as source code by publication**.
## Downloading
Install the Datasets library
pip install datasets
and then log in to your HuggingFace account
huggingface-cli login
All unpublished* datasets are temporarily "gated" i.e. you must have requested and been approved for access. Galaxy Zoo team members should go to https://huggingface.co/mwalmsley, click the dataset, and "request access", then wait for approval.
Gating will be removed on publication.
*Currently: the `gz_h2o` and `gz_ukidss` datasets
## Usage
```python
from datasets import load_dataset
# . split='train' picks which split to load
dataset = load_dataset(
f'mwalmsley/gz_candels', # each dataset has a random fixed train/test split
split='train'
# some datasets also allow name=subset (e.g. name="tiny" for gz_evo). see the viewer for subset options
)
dataset.set_format('torch') # your framework of choice e.g. numpy, tensorflow, jax, etc
print(dataset_name, dataset[0]['image'].shape)
```
Then use the `dataset` object as with any other HuggingFace dataset, e.g.,
```python
from torch.utils.data import DataLoader
dataloader = DataLoader(ds, batch_size=4, num_workers=1)
for batch in dataloader:
print(batch.keys())
# the image key, plus a key counting the volunteer votes for each answer
# (e.g. smooth-or-featured-gz2_smooth)
print(batch['image'].shape)
break
```
You may find these HuggingFace docs useful:
- [PyTorch loading options](https://huggingface.co/docs/datasets/en/use_with_pytorch#data-loading).
- [Applying transforms/augmentations](https://huggingface.co/docs/datasets/en/image_process#apply-transforms).
- [Frameworks supported](https://huggingface.co/docs/datasets/v2.19.0/en/package_reference/main_classes#datasets.Dataset.set_format) by `set_format`.
## Dataset Structure
Each dataset is structured like:
```json
{
'image': ..., # image of a galaxy
'smooth-or-featured-[campaign]_smooth': 4,
'smooth-or-featured-[campaign]_featured-or-disk': 12,
... # and so on for many questions and answers
}
```
Images are loaded according to your `set_format` choice above. For example, ```set_format("torch")``` gives a (3, 424, 424) CHW `Torch.Tensor`.
The other keys are formatted like `[question]_[answer]`, where `question` is what the volunteers were asked (e.g. "smooth or featured?" and `answer` is the choice selected (e.g. "smooth"). **The values are the count of volunteers who selected each answer.**
`question` is appended with a string noting in which Galaxy Zoo campaign this question was asked e.g. `smooth-or-featured-gz2`. For most datasets, all questions were asked during the same campaign. For GZ DESI, there are three campaigns (`dr12`, `dr5`, and `dr8`) with very similar questions.
GZ Evo
(we will shortly add keys for the astronomical identifiers i.e. the sky coordinates and telescope source unique ids)
## Key Limitations
Because the volunteers are answering a decision tree, the questions asked depend on the previous answers, and so each galaxy and each question can have very different total numbers of votes. This interferes with typical metrics that use aggregated labels (e.g. classification of the most voted, regression on the mean vote fraction, etc.) because we have different levels of confidence in the aggregated labels for each galaxy. We suggest a custom loss to handle this. Please see the Datasets and Benchmarks paper for more details (under review, sorry).
All labels are imperfect. The vote counts may not always reflect the true appearance of each galaxy. Additionally,
the true appearance of each galaxy may be uncertain - even to expert astronomers.
We therefore caution against over-interpreting small changes in performance to indicate a method is "superior". **These datasets should not be used as a precise performance benchmark.** |