File size: 2,486 Bytes
3efa0f2 d367b13 7a1bd38 699ad5e 7a1bd38 699ad5e 7a1bd38 dda1d30 7a1bd38 dda1d30 7a1bd38 dda1d30 7a1bd38 699ad5e 7a1bd38 699ad5e 7a1bd38 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 |
---
license: openrail
task_categories:
- image-segmentation
pretty_name: California Burned Areas
size_categories:
- n<1K
---
# California Burned Areas Dataset
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
This dataset contains images from Sentinel-2 satellites taken before and after a wildfire.
The ground truth masks are provided by the California Department of Forestry and Fire Protection and they are mapped on the images.
### Supported Tasks
The dataset is designed to do binary semantic segmentation of burned vs unburned areas.
## Dataset Structure
### Dataset opening
Dataset was compressed using `h5py` and BZip2 from `hdf5plugin`. **WARNING: `hdf5plugin` is necessary to extract data**
### Data Instances
Each matrix has a shape of 5490x5490xC, where C is 12 for pre-fire and post-fire images, while it is 0 for binary masks.
### Data Fields
In each HDF5 file, you can find post-fire, pre-fire images and binary masks. The file is structured in this way:
```bash
βββ foldn
β βββ uid0
β β βββ pre_fire
β β βββ post_fire
β β βββ mask
β βββ uid1
β βββ post_fire
β βββ mask
β
βββ foldm
βββ uid2
β βββ post_fire
β βββ mask
βββ uid3
βββ pre_fire
βββ post_fire
βββ mask
...
```
where `foldn` and `foldm` are fold names and `uidn` is a unique identifier for the wilfire.
### Data Splits
There are 5 random splits whose names are: 0, 1, 2, 3 and 4.
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
Data are collected directly from Copernicus Open Access Hub through the API. The band files are aggregated into one single matrix.
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |