File size: 3,066 Bytes
52ec00b
 
4207411
 
 
 
 
 
 
 
 
 
 
8594822
4207411
 
 
52ec00b
4207411
7b7de30
 
 
4207411
 
 
 
 
 
 
 
7b7de30
 
 
 
 
4207411
 
 
 
7b7de30
4207411
 
7b7de30
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4207411
a9ec10c
8594822
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
---
license: apache-2.0
language:
- en
tags:
- Pytorch
- mmsegmentation
- segmentation
- Crop Classification
- Multi Temporal
- Geospatial
- Foundation model
datasets:
- ibm-nasa-geospatial/multi-temporal-crop-classification
metrics:
- accuracy
- IoU
---
### Model and Inputs
The pretrained [Prithvi-100m](https://huggingface.co/ibm-nasa-geospatial/Prithvi-100M/blob/main/README.md) parameter model is finetuned to classify crop and other land cover types based off HLS data and CDL labels from the [multi_temporal_crop_classification dataset](https://huggingface.co/datasets/ibm-nasa-geospatial/multi-temporal-crop-classification). 

This dataset includes input chips of 224x224x18, where 224 is the height and width and 18 is combined with 6 bands of 3 time-steps. The bands are:
 
1. Blue
2. Green
3. Red
4. Narrow NIR
5. SWIR 1
6. SWIR 2

Labels are from CDL(Crop Data Layer) and classified into 13 classes.

![](multi_temporal_crop_classification.png)

The Prithvi-100m model was initially pretrained using a sequence length of 3 timesteps. For this task, we leverage the capacity for multi-temporal data input, which has been integrated from the foundational pretrained model. This adaptation allows us to achieve more generalized finetuning outcomes.

### Code
Code for Finetuning is available through [github](https://github.com/NASA-IMPACT/hls-foundation-os/tree/main/fine-tuning-examples)

Configuration used for finetuning is available through [config](https://github.com/NASA-IMPACT/hls-foundation-os/blob/main/fine-tuning-examples/configs/multi_temporal_crop_classification.py).

### Results
The experiment by running the mmseg stack for 80 epochs using the above config led to the following result:

|     **Classes**    | **IoU**| **Acc**|
|:------------------:|:------:|:------:|
| Natural Vegetation | 0.3362 | 39.06% |
|       Forest       | 0.4362 | 65.88% |
|        Corn        | 0.4574 | 54.53% |
|      Soybeans      | 0.4682 | 62.25% |
|      Wetlands      | 0.3246 | 45.62% |
|  Developed/Barren  | 0.3077 |  49.1% |
|     Open Water     | 0.6181 | 90.04% |
|    Winter Wheat    | 0.4497 | 66.75% |
|       Alfalfa      | 0.2518 | 63.97% |
|Fallow/Idle Cropland| 0.328  | 54.46% |
|       Cotton       | 0.2679 | 66.37% |
|       Sorghum      | 0.2741 | 75.91% |
|        Other       | 0.2803 | 39.76% |

|**aAcc**|**mIoU**|**mAcc**|
|:------:|:------:|:------:|
| 54.32% | 0.3692 | 59.51% |

It is important to acknowledge that the CDL (Crop Data Layer) labels employed in this process are known to contain noise and are not entirely precise, thereby influencing the model's performance. Fine-tuning the model with more accurate labels is expected to further enhance its overall effectiveness, leading to improved results.


### Inference
There is an inference script that allows to run the hls-cdl crop classification model for inference on HLS images. These input have to be geotiff format, including 18 bands for 3 time-step, and each time-step includes the channels described above (Blue, Green, Red, Narrow NIR, SWIR, SWIR 2) in order.