The dataset for this competition includes a time series of satellite imagery from Planet’s NICFI basemaps (license agreement) and labels for field boundaries that were annotated on the same imagery source. The labels were digitized over Planet Basemaps for the months of March, April (on season) and August (off season) of 2021 by a team of annotators from TaQadam. An additional 3 months of imagery (October, November and December) are added to the time series data and are then matched with corresponding field boundary labels.
The time-series is provided for six months (March, April, August, October, November and December), but you do not need to use the observations from all months. You are allowed to select specific months, or apply any pre-processing and feature extraction to the time-series data before input to your model. Note that you would need to provide your full feature-extraction and training scripts if you win in the competition.
The data you will have access to (Satellite imagery and labels) are tiled into 256x256 chips adding up to 70 tiles. Within those 70 tiles 1532 individual field boundaries have been identified. The dataset has been split into training and test chips (57 in the train and 13 in the test). You will train your machine learning model on the fields included in the training chips and will apply your model to predict field boundaries for chips in the test set. You will submit your predictions for field boundary masks for the list of chips in the test dataset.
Labels were created only for fields large enough to be distinguished on the planet basemaps and for fields completely contained in the chip; this means that not all the pixels are labeled in each chip.
Each chip has:
Data for this competition is hosted on Radiant MLHub - open-access repository for geospatial data. You can access the data by creating a free account on Radiant MLHub.
You can download the data using Radiant MLHub Python Client (see the example notebook) or simply by going to the Radiant MLHub website.
The data is structured in three collections based on the metadata specification of SpatioTemporal Asset Catalog (STAC):
Variables definitions:
The label chips contain the mapping of pixels to crop type labels. The following pixel values correspond to the following crop types:
Files available for download:
The collections you download from Radiant MLHub will be structured as following:
Source imagery:
nasa_rwanda_field_boundary_competition_source_train
|
|—nasa_rwanda_field_boundary_competition_source_train_{chip_id}_{timestamp}
| |—B01.tif
| |—B02.tif
| |—B03.tif
| |—B04.tif
| |—stac.json
Source test imagery:
|—nasa_rwanda_field_boundary_competition_source_test_{chip_id}_{timestamp}
| |—B01.tif
| |—B02.tif
| |—B03.tif
| |—B04.tif
| |—stac.json
Train labels:
nasa_rwanda_field_boundary_competition_labels_train
|
|—nasa_rwanda_field_boundary_competition_labels_train_{chip_id}
| |—stac.json
| |—raster_labels.tif Find all the data for this challenge here.