Terramind Tiny in Action¶
📥 Download GeospatialStudio-Walkthrough-Flooding.ipynb and try it out
Assume you are interested in mapping flooding, traditionally you might have either relied on on-the-ground mapping, or possibly for manual analysis of remote-sensing imagery (i.e. satellite or UAV). In order to scale up these efforts and operationalise, we need a way to automate the extraction of flood extent from satellite imagery. This is where we turn to AI models.

In this walkthrough we will assume that a model doesn't exist yet and we want to train a new model. We will then show how to drive the model to map impact.
We will walk through the following steps as part of this walkthrough:
- Upload and onboarding of data
- Configuring and submitting a tuning task
- Monitoring model training
- Testing and validation of the outputs
Pre-requisites¶
You will require access to a deployed instance of the Geospatial Studio. For more information on how you can deploy the studio (either locally or in a cluster), visit the GEOStudio docs: Geospatial Studio Docs
For more information about the Geospatial Studio SDK that is used in this notebook, and all the functions available through it, see the SDK docs page: Geospatial Studio SDK Docs
This walkthrough also requires you to have a direct download URL pointing to a zip file of the dataset you wish to use. We provide a sample dataset url (zip file) below to go through this notebook.
If you have your own dataset locally, you can find instructions on how to use the SDK to temporarily upload it to the cloud and create a download url link in the steps that follow.
Get sample training data to use for this walkthrough¶
This walkthrough used the following dataset to fine-tune the Terramind Tiny model: https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/sen1floods11_v1.1.4_zip.zip
Download and unzip the above archive and if you wish you can explore the data with QGIS (or any similar tool).
%load_ext autoreload
%autoreload 2
import os
import json
import uuid
import pandas as pd
import rasterio
from rasterio.plot import show
import matplotlib.pyplot as plt
import urllib3
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
from geostudio import Client
from geostudio import gswidgets
Connecting to the platform¶
First, we set up the connection to the platform backend. To do this we need the base url for the studio UI and an API key.
To get an API Key:
- Go to the Geospatial Studio UI page and navigate to the Manage your API keys link.
- This should pop-up a window where you can generate, access and delete your api keys. NB: every user is limited to a maximum of two activate api keys at any one time.
Store the API key and geostudio ui base url in a credentials file locally, for example in /User/bob/.geostudio_config_file. You can do this by:
echo "GEOSTUDIO_API_KEY=<paste_api_key_here>" > .geostudio_config_file
echo "BASE_STUDIO_UI_URL=<paste_ui_base_url_here>" >> .geostudio_config_file
Copy and paste the file path to this credentials file in call below.
#############################################################
# Initialize Geostudio client using a geostudio config file
#############################################################
gfm_client = Client(geostudio_config_file="../../.geostudio_config_file_staging")
Populate Studio¶
Assuming you have deployed the studio, you should have also gone through the first steps which include populating your GEOStudio instance with sample data for testing. If you did so, you should have a list of datasets that you can explore with the functio below
Instructions to deploy the studio and first steps: https://terrastackai.github.io/geospatial-studio
# explore sample datasets populated with your Studio deployment
gfm_client.list_datasets(output="df")
Data onboarding¶
In order to onboard your dataset to the Geospatial Studio, you need to have a direct download URL pointing to a zip file of the dataset. You can use this dataset url as an example to go through this notebook.
If you have the dataset locally, you can use Box, OneDrive or any other cloud storage you are used to, but in addition, to make this easier for you, there is a function which will upload your data to a temporary location in the cloud (with in Studio object storage) and provide you with a url which can be used to pass to the onboarding process. NB: the same upload function can be useful for pushing files for inferecnce or to processing pipelines.
If needed you can package a set of files for upload, you can use a command like:
zip -j flooding-dataset-upload.zip flooding-dataset-upload/*
NB: If you already have the data online you can skip this step.
object_name = "test-checkpoint.ckpt" #Must be a valid string, not a path
dataset = "flooding-dataset-upload.zip"
upload_links = gfm_client.get_fileshare_links(object_name)
upload_url = upload_links["upload_url"]
download_url = upload_links["download_url"]
!curl --progress-bar -O -X PUT -T "$dataset" "$upload_url"
Pre-scan the dataset¶
Pre-scan the dataset to check the accessibility of the dataset URL, check that you have matching data and label pairs, and check that you have specified the correct number of bands and their descriptions from the dataset.
# [Optional]
gfm_client.pre_scan_dataset({
"dataset_url": download_url,
"label_suffix": "_LabelHand.tif",
"training_data_suffixes":
["_S2Hand.tif", "_S1Hand.tif"]
})
Onboard the dataset to the dataset factory¶
Now we use the SDK to provide the information about the dataset, including name, suffixes etc. A more detailed description of the dataset details is provided in the UI walkthrough.
Note:
- Change the value of the
dataset_urlvariable below to the url of your zip file or thedownload_urllink you got from using the SDK upload_file function above - Change the values of
training_data_suffixandlabel_suffixto the suffixes of your training and label data files respectively if using a different dataset (aside from the one provided) - Change the
label_categories,bandsand descriptions to those that match your dataset
onboard_dataset_payload = {
"dataset_name": "Sentinel Flood Multimodal Test",
"data_sources": [
{
"bands": [
{
"index": "0",
"band_name": "B01",
"description": "",
"scaling_factor": "1"
},
{
"index": "1",
"band_name": "B02",
"RGB_band": "B",
"description": "",
"scaling_factor": "1"
},
{
"index": "2",
"band_name": "B03",
"RGB_band": "G",
"description": "",
"scaling_factor": "1"
},
{
"index": "3",
"band_name": "B04",
"RGB_band": "R",
"description": "",
"scaling_factor": "1"
},
{
"index": "4",
"band_name": "B05",
"description": "",
"scaling_factor": "1"
},
{
"index": "5",
"band_name": "B06",
"description": "",
"scaling_factor": "1"
},
{
"index": "6",
"band_name": "B07",
"description": "",
"scaling_factor": "1"
},
{
"index": "7",
"band_name": "B08",
"description": "",
"scaling_factor": "1"
},
{
"index": "8",
"band_name": "B8A",
"description": "",
"scaling_factor": "1"
},
{
"index": "9",
"band_name": "B09",
"description": "",
"scaling_factor": "1"
},
{
"index": "10",
"band_name": "B11",
"description": "",
"scaling_factor": "1"
},
{
"index": "11",
"band_name": "B12",
"description": "",
"scaling_factor": "1"
},
{
"index": "12",
"band_name": "CLD",
"description": "",
"scaling_factor": "1"
}
],
"connector": "sentinelhub",
"collection": "s2_l2a",
"modality_tag": "S2L2A",
"file_suffix": "_S2Hand.tif"
},
{
"bands": [
{"index": "0", "band_name": "VV", "description": ""},
{"index": "1", "band_name": "VH", "description": ""}
],
"connector": "sentinelhub",
"collection": "s1_grd",
"modality_tag": "S1GRD",
"align_dates": "true",
"file_suffix": "_S1Hand.tif",
"scaling_factor": [1, 1]
}
],
"label_categories": [
{"id": "0", "name": "No Floods", "description": "Flooding assets"},
{"id": "1", "name": "Floods", "description": "Flooding assets"}
],
"dataset_url": "https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/sen1floods11_v1.1.4_zip.zip",
"description": "Flood data from places",
"label_suffix": "_LabelHand.tif",
"purpose": "Segmentation"
}
# start onboarding process
onboard_response = gfm_client.onboard_dataset(data=onboard_dataset_payload)
display(json.dumps(onboard_response, indent=2))
Monitor you dataset onboarding task¶
You can use the UI to monitor the progress of your dataset onboarding submission. Access and log in to the UI of your deployed GEOStudio instance and click on the Dataset factory page.
# Once onboarding is complete, select your onboarded dataset to use for you fine-tuning task
selected_dataset = onboard_response["id"]
Fine-tuning submission¶
Once the data is onboarded, you are ready to setup your tuning task. In order to run a fine-tuning task, you need to select the following items:
- tuning task type/config template - what type of learning task are you attempting? segmentation, regression etc
- fine-tuning dataset - what dataset will you use to train the model for your particular application?
- base foundation model - which geospatial foundation model will you use as the starting point for your tuning task?
Below we walk you through how to use the Geospatial Studio SDK to see how we configure our task and submit it.
Tuning task¶
The tuning task tells the model what type of task it is (segmentation, regression etc), and exposes a range of optional hyperparameters which the user can set. These all have reasonable defaults, but it gives uses the possibility to configure the model training how they wish. Below, we will check what task templates are available to us from the sample data from the GEOStudio deployment
After that, we will create and upload new task templates to the platform. The templates are for Terratorch (the backend tuning library), and more details of Terratroch and configuration options can be found here: https://github.com/terrastackai/terratorch
# list tasks available from our deployment sample data
tasks = gfm_client.list_tune_templates(output="df")
display(tasks[['name','description', 'id','created_by','updated_at']])
Create the Terramind Tiny tuning task¶
create_terramind_template = {
"name": "terramind tiny Segmentation",
"description": "Terramind multimodal task for Segmantation",
"purpose": "Segmentation",
"model_params": {
"$uri": "https://ibm.com/watsonx.ai.geospatial.finetune.segmentation.json",
"type": "object",
"title": "Finetune",
"$schema": "https://json-schema.org/draft/2020-12/schema",
"properties": {
"data": {
"type": "object",
"default": {
"batch_size": 4,
"constant_multiply": 1,
"workers_per_gpu": 2,
"check_stackability": False
},
"properties": {
"batch_size": {
"type": "int",
"default": 4,
"description": "Batch size",
"studio_name": "Batch size"
},
"constant_multiply": {
"type": "float",
"default": 1,
"description": "Constant Scale",
"studio_name": "Constant Scale"
},
"workers_per_gpu": {
"studio_name": "Workers per GPU",
"description": "Workers per GPU",
"type": "int",
"default": 2
},
"check_stackability": {
"studio_name": "Check Stackability",
"description": "Check Stackability",
"type": "bool",
"default": False
}
},
"studio_name": "Data loading"
},
"model": {
"type": "object",
"default": {
"decode_head": {
"channels": 256,
"num_convs": 4,
"decoder": "UNetDecoder",
"loss_decode": {
"type": "CrossEntropyLoss",
"avg_non_ignore": True
}
},
"frozen_backbone": False,
"tiled_inference_parameters": {
"h_crop": 224,
"h_stride": 208,
"w_crop": 224,
"w_stride": 208,
"average_patches": True
}
},
"properties": {
"decode_head": {
"type": "object",
"default": {
"channels": 256,
"num_convs": 4,
"decoder": "UperNetDecoder",
"loss_decode": {
"type": "CrossEntropyLoss",
"avg_non_ignore": True
}
},
"properties": {
"channels": {
"type": "int",
"default": 256,
"description": "Channels at each block of the decode head, except the final one",
"studio_name": "Channels"
},
"num_convs": {
"type": "int",
"default": 4,
"description": "Number of convolutional blocks in the head (except the final one)",
"studio_name": "Blocks"
},
"decoder": {
"enum": [
"UperNetDecoder",
"UNetDecoder"
],
"type": "string",
"default": "Fixed",
"description": "Decoder type",
"studio_name": "Decoder type"
},
"loss_decode": {
"type": "object",
"properties": {
"type": {
"enum": [
"CrossEntropyLoss"
],
"type": "string",
"default": "CrossEntropyLoss",
"description": "Type of loss function",
"studio_name": "Loss function"
},
"avg_non_ignore": {
"type": "bool",
"default": True,
"description": "The loss is only averaged over non-ignored targets (ignored targets are usually where labels are missing in the dataset) if this is True"
}
},
"description": "Loss function to be used",
"studio_name": "Loss"
}
},
"description": "Architecture of the decode head",
"studio_name": "Head"
},
"auxiliary_head": {
"type": "object",
"default": {},
"properties": {
"decoder": {
"type": "string",
"default": "FCNDecoder",
"description": "Decoder function to use",
"studio_name": "Decoder"
},
"channels": {
"type": "int",
"default": 256,
"description": "Channels at each block of the decode head, except the final one",
"studio_name": "Channels"
},
"num_convs": {
"type": "int",
"default": 2,
"description": "Number of convolutional blocks in the head (except the final one)",
"studio_name": "Blocks"
},
"in_index": {
"type": "int",
"default": -1,
"description": "Index of the input list to take. Defaults to -1",
"studio_name": "In index"
},
"dropout": {
"type": "int",
"default": 0,
"description": "Dropout value to apply. Defaults to 0",
"studio_name": "Dropout"
},
"loss_decode": {
"type": "object",
"properties": {
"type": {
"enum": [
"CrossEntropyLoss"
],
"type": "string",
"default": "CrossEntropyLoss",
"description": "Type of loss function",
"studio_name": "Loss function"
},
"loss_weight": {
"type": "float",
"default": 1,
"description": "Multiplicative weight of the loss of the auxiliary head in the loss. The loss is calculated as aux_head_weight * aux_head_loss + decode_head_loss",
"studio_name": "Loss weight"
},
"avg_non_ignore": {
"type": "bool",
"default": True,
"description": "The loss is only averaged over non-ignored targets (ignored targets are usually where labels are missing in the dataset) if this is True"
}
},
"description": "Loss function to be used",
"studio_name": "Loss"
}
},
"description": "Architecture of the auxiliary head"
},
"frozen_backbone": {
"type": "bool",
"default": False,
"description": "Freeze the weights of the backbone when set to True",
"studio_name": "Freeze backbone"
},
"tiled_inference_parameters": {
"type": "object",
"default": {
"h_crop": 224,
"h_stride": 208,
"w_crop": 224,
"w_stride": 208,
"average_patches": True
},
"properties": {
"h_crop": {
"type": "int",
"default": 224,
"description": "h_crop values for tilling images",
"studio_name": "h_crop"
},
"h_stride": {
"type": "int",
"default": 196,
"description": "h_stride values for tilling images",
"studio_name": "h_stride"
},
"w_crop": {
"type": "int",
"default": 224,
"description": "w_crop values for tilling images",
"studio_name": "w_crop"
},
"w_stride": {
"type": "int",
"default": 196,
"description": "w_stride values for tilling images",
"studio_name": "w_stride"
},
"average_patches": {
"type": "bool",
"default": False,
"description": "Whether to use average_patches",
"studio_name": "average_patches"
}
}
}
},
"description": "Model architecture definition",
"studio_name": "Architecture"
},
"runner": {
"type": "object",
"default": {
"max_epochs": 10,
"early_stopping_patience": 20,
"early_stopping_monitor": "val/loss"
},
"properties": {
"max_epochs": {
"type": "int",
"default": 10,
"description": "Training epochs",
"studio_name": "Training epochs"
},
"early_stopping_patience": {
"type": "int",
"default": 20,
"description": "Early stopping patience",
"studio_name": "Early stopping patience"
},
"early_stopping_monitor": {
"type": "string",
"default": "val/loss",
"description": "Monitoring value to determine early stopping",
"studio_name": "Early stopping monitor"
}
},
"studio_name": "Runner"
},
"lr_config": {
"type": "object",
"default": {
"policy": "Fixed"
},
"required": [
"policy"
],
"properties": {
"policy": {
"enum": [
"Fixed",
"CosineAnnealing"
],
"type": "string",
"default": "Fixed",
"description": "Policy type",
"studio_name": "Policy type"
},
"warmup_iters": {
"type": "int",
"default": 0,
"description": "LR warmup iterations. Valid for some policies",
"studio_name": "Learning rate warmup iterations"
},
"warmup_ratio": {
"type": "float",
"default": 1,
"description": "Initial lr at warmup will be learning_rate * warmup_ratio",
"studio_name": "LR warmup initialization ratio"
}
},
"description": "Learning rate policy",
"studio_name": "Learning rate policy"
},
"optimizer": {
"type": "object",
"default": {
"lr": 6e-05,
"type": "Adam"
},
"properties": {
"lr": {
"type": "float",
"default": 6e-05,
"description": "Learning rate",
"studio_name": "Learning rate"
},
"type": {
"enum": [
"Adam",
"SGD",
"AdamW",
"RMSProp"
],
"default": "Adam",
"description": "Optimizer to be used",
"studio_name": "Optimizer type"
},
"weight_decay": {
"type": "float",
"default": 0,
"description": "L2 weight regularization (weight decay)",
"studio_name": "L2 regularization weight"
}
},
"description": "Optimizer",
"studio_name": "Optimizer"
},
"dataset_id": {
"type": "string",
"description": "ID of dataset to use for this finetuning",
"studio_name": "Dataset"
},
"evaluation": {
"type": "object",
"default": {
"interval": 1
},
"properties": {
"interval": {
"type": "int",
"default": 1,
"description": "Frequency of epochs with which to perform validation",
"studio_name": "Epoch interval"
}
},
"studio_name": "Validation"
},
"backbone_model_id": {
"type": "string",
"description": "ID of the pretrained backbone"
}
},
"description": "A request sent to the finetuning service to start a finetune task for segmentation"
},
"extra_info": {
"runtime_image": "quay.io/geospatial-studio/terratorch:latest",
"model_category": "terramind"
},
"content": "# © Copyright IBM Corporation 2025
# SPDX-License-Identifier: Apache-2.0


# lightning.pytorch==2.1.1
seed_everything: 42
trainer:
  accelerator: auto
  strategy: auto
  devices: auto
  num_nodes: 1
  precision: 16-mixed
  logger:
    class_path: lightning.pytorch.loggers.mlflow.MLFlowLogger
    init_args:
      experiment_name: {{ tune_id }} 
      run_name: "Train"    
      tracking_uri: {{ mlflow_tracking_url }}
      save_dir: {{ mount_root + 'tune-tasks/' + tune_id + '/mlflow' }}
      {% if mlflow_tags -%}
      tags:
        {% for key, value in mlflow_tags.items() -%}
        {{ key }}: {{ value }}
        {% endfor %}
      {%- endif %}       
  callbacks:
    - class_path: RichProgressBar
    - class_path: LearningRateMonitor
      init_args:
        logging_interval: epoch
    # ---- Early stop if ----
    {% if runner["early_stopping_patience"] -%}
    - class_path: EarlyStopping
      init_args:
        monitor: {{ runner["early_stopping_monitor"] }}
        patience: {{ runner["early_stopping_patience"] }}
    {%- endif %}
     # ---- Early stop endif ----
    - class_path: ModelCheckpoint
      init_args:
        dirpath: {{ mount_root + 'tune-tasks/' + tune_id  + '/' }}
        mode: min
        monitor: val/loss
        filename: {{ 'best-state_dict-{epoch:02d}' }}
        save_weights_only: True
  max_epochs: {{ runner["max_epochs"] }}
  check_val_every_n_epoch: {{ evaluation["interval"] }}
  log_every_n_steps: 50
  enable_checkpointing: true
  default_root_dir: {{ mount_root + 'tune-tasks/' + tune_id }}
data:
  {% if image_modalities|length == 1 %} 
  class_path: GenericNonGeoSegmentationDataModule
  init_args:
    {% if data["check_stackability"] -%}
    check_stackability: {{ data["check_stackability"] }}
    {% else %}
    check_stackability: false
    {% endif -%}
    batch_size: 2
    num_workers: 1
    dataset_bands:  # Dataset bands
      {{ bands.values() | list | first | to_yaml | indent(6) }}
    rgb_indices:
      {{ rgb_band_indices | to_yaml | indent(6) }}
    train_data_root: {{ data_root }}{{ train_data_dir.values() | list | first }}
    val_data_root: {{ data_root }}{{ val_data_dir.values() | list | first }}
    test_data_root: {{ data_root }}{{ test_data_dir.values() | list | first }}
    # labels roots
    train_label_data_root: {{ data_root + train_labels_dir }}
    val_label_data_root: {{ data_root + val_labels_dir }}
    test_label_data_root: {{ data_root + test_labels_dir }}
    {% if train_split_path -%}
    train_split: {{ data_root + train_split_path }}
    {% endif -%}
    {% if test_split_path -%}
    test_split: {{ data_root + test_split_path }}
    {% endif -%}
    {% if val_split_path -%}
    val_split: {{ data_root + val_split_path }}
    {% endif -%}

    {% if img_suffix -%}
    img_grep:  {{ img_suffix.values() | list | first | tojson }}
    {% endif -%}
    {% if seg_map_suffix -%}
    label_grep: "{{ seg_map_suffix }}"
    {% endif -%}
    means: 
      {{ norm_means.values() | list | first| to_yaml | indent(6) }}
    stds: 
      {{ norm_stds.values() | list | first | to_yaml | indent(6) }}
    num_classes: {{ classes|length }}
    train_transform:
      - class_path: albumentations.D4
      - class_path: ToTensorV2
    no_data_replace: 0
    no_label_replace: -1


  {% else %}
  class_path: terratorch.datamodules.GenericMultiModalDataModule
  init_args:
    {% if data["check_stackability"] -%}
    check_stackability: {{ data["check_stackability"] }}
    {% else %}
    check_stackability: false
    {% endif -%}
    # Config for only segmentation. No need to automate this. 
    task: 'segmentation'
    # Out of cuda error for anything > 2
    # ToDo: Figure out why batch_size replacement is not working.
    batch_size: 2
    {% if num_workers -%}
    num_workers: {{ num_workers }}
    {% else -%}
    num_workers: 2
    {% endif -%}
    no_label_replace: {{ label_nodata }}
    no_data_replace: {{ image_nodata_replace }}
    dataset_bands:
      {{ bands | to_yaml | indent(6)}}
    output_bands:
      {{ output_bands | to_yaml  | indent(6)}}
    modalities:
      {{ image_modalities | to_yaml | indent(6) }}
    rgb_modality: {{ rgb_modality }}
    rgb_indices: 
      {{ rgb_band_indices | to_yaml | indent(6) }}
    train_data_root:
      {% for key, val in train_data_dir.items() -%}
       {{ key }}: {{ data_root }}{{ val }}
      {% endfor %}
    train_label_data_root:  {{ data_root +  train_labels_dir }}
    val_data_root:
      {% for key, val in val_data_dir.items() -%}
       {{ key }}: {{ data_root }}{{ val }}
      {% endfor %}
    val_label_data_root: {{ data_root +  test_labels_dir }}
    test_data_root:
      {% for key, val in test_data_dir.items() -%}
       {{ key }}: {{ data_root }}{{ val }}
      {% endfor %}
    test_label_data_root: {{ data_root + test_labels_dir }}
    {% if train_split_path -%}
    train_split: {{  data_root + train_split_path }}
    {% endif -%}
    {% if test_split_path -%}
    test_split: {{  data_root + test_split_path }}
    {% endif -%}
    {% if val_split_path -%}
    val_split: {{  data_root + val_split_path }}
    {% endif -%}
    {% if img_suffix -%}
    image_grep: 
      {{ img_suffix | to_yaml | indent(6) }}
    {% endif -%}
    {% if seg_map_suffix -%}
    label_grep: "{{ seg_map_suffix }}"
    {% endif -%}

    num_classes: {{ classes|length }}
    {% if data["expand_temporal_dimension"] is not none -%}
    expand_temporal_dimension: data["expand_temporal_dimension"]
    {% endif -%}
    {% if data["drop_last"] is not none -%}
    drop_last: data["drop_last"]
    {% endif -%}

    means: 
      {{ norm_means | to_yaml | indent(6) }}
    stds: 
      {{ norm_stds | to_yaml | indent(6) }}
    train_transform:
      - class_path: albumentations.D4  # Random flip and rotations
      - class_path: albumentations.pytorch.transforms.ToTensorV2
  {% endif %}
model:
  class_path: terratorch.tasks.SemanticSegmentationTask
  init_args:
    model_factory: EncoderDecoderFactory
    model_args:
      backbone: {{ pretrained_model_name }}
      #  terramind_v1_base  # large version: terramind_v1_large
      backbone_pretrained: true
      {%- if image_modalities %}
      backbone_modalities:
        {{ image_modalities | to_yaml  | indent(8)}}
      {%- endif %}
      {% if  image_modalities|length > 1 %}
      backbone_merge_method: mean
      {% endif %}
      backbone_bands: 
        {{ output_bands | to_yaml | indent(8) }}
      necks:
        {%- if pretrained_model_name == "terramind_v1_tiny" %}
        - name: SelectIndices
          indices: [2, 5, 8, 11]  # tiny version
        {%- elif pretrained_model_name == "terramind_v1_base" %}
        - name: SelectIndices
          indices: [2, 5, 8, 11]  # base version
        {%- elif pretrained_model_name == "terramind_v1_large" %}
        - name: SelectIndices
          indices: [5, 11, 17, 23]  # large version
        {% endif %}
        - name: ReshapeTokensToImage
          remove_cls_token: False
        - name: LearnedInterpolateToPyramidal

      decoder: {{ model["decode_head"]["decoder"] }}
      # UNetDecoder
      {% if  model["decode_head"]["decoder"] == "UperNetDecoder" -%}
      decoder_channels: [512, 256, 128, 64]
      {% elif  model["decode_head"]["decoder"] == "UNetDecoder" -%}
      #TODO user provided channels
      decoder_channels: [512, 256, 128, 64]
      {% else %}
      decoder_channels: {{ model["decode_head"]["channels"] }}
      {% endif -%}
      head_dropout: 0.1
      num_classes: {{ classes|length }}
    loss: {{ model["decode_head"]["loss_decode"]["type"] }}
    plot_on_val: {{ runner["plot_on_val"] }}
    #  dice
    ignore_index: {{ ignore_index }}
    freeze_backbone:  {{ model["frozen_backbone"] | lower }}
    freeze_decoder: false
    {% if  model["class_names"] %}
    class_names: {{ model["class_names"] | to_yaml | indent(6)}}
    {% endif %}
    # ---- optimizer start ----
    {% if model["optimizer"] -%}
    optimizer: {{ model["optimizer"]["type"] }}
    lr: {{ model["optimizer"]["lr"] | float }}
    {% endif -%}
    # ---- optimizer end ----
    {% if model["tiled_inference_parameters"] %}
    tiled_inference_parameters: 
      h_crop: {{ model["tiled_inference_parameters"]["h_crop"] | int}}
      h_stride: {{ model["tiled_inference_parameters"]["h_stride"] | int }}
      w_crop: {{ model["tiled_inference_parameters"]["w_crop"] | int}}
      w_stride: {{ model["tiled_inference_parameters"]["w_stride"] | int }}
      average_patches: {{ model["tiled_inference_parameters"]["average_patches"] }}
    {% else %}
    # ToDo: Remove the tiled_inference if user not provided. 
    tiled_inference_parameters: 
      h_crop: 224
      # stride logic = would be h_crop - h_crop * 0.125
      h_stride: 208
      w_crop: 224
      # stride logic = would be w_crop - w_crop * 0.125
      w_stride: 208
      average_patches: true
    {% endif %}

optimizer:
  class_path: torch.optim.AdamW
  init_args:
    {% if optimizer["lr"] -%}
    lr: {{ optimizer["lr"] | float }}
    {% else %}
    lr: 2.e-5
    {% endif -%}
    {%- if optimizer["weight_decay"] -%}
    weight_decay: {{ optimizer["weight_decay"] }}
    {%- else -%}
    weight_decay: 0.05
    {% endif %}
lr_scheduler:
  class_path: ReduceLROnPlateau
  init_args:
    monitor: val/loss
    factor: 0.5
    patience: 5
"
}
res = gfm_client.create_task(data=create_terramind_template)
# Select the task you created for your fine-tuning job
task_id = res["id"]
Review the created task¶
# Review the created task with the original template
tt = gfm_client.get_task_template("601cc248-135e-4ad3-96e1-0bb270ac0534", output='cell')
# Review the created task with the replaced template with the provided dataset
rendered_template = gfm_client.render_template(task_id = "601cc248-135e-4ad3-96e1-0bb270ac0534", dataset_id=selected_dataset, output="cell")
Configure the task parameters you want¶
# view the full meta-data and details of the selected task
task_meta = gfm_client.get_task(task_id=task_id)
task_meta
If you are happy with your choice, you can decide which (if any) hyperparameters you want to set (otherwise defaults will be used).
Here we can see the available parameters and their associated defaults. To update a parameter you can just set values in the dictionary (as shown below for max_epochs).
# show the default values for parameters
task_params = gfm_client.get_task_param_defaults(task_id)
task_params
# [Optional]Configure the parameters you want
task_params['runner']['max_epochs'] = '3'
# task_params['optimizer']['type'] = 'AdamW'
# task_params['data']['batch_size'] = 4
Base foundation model¶
The base model is the foundation model (encoder) which has been pre-trained and has the basic understanding of the data. More information can currently be found on the Terramind tiny model can be found on hugging face.
For this example, we will start by creating a base foundation model entry for terramind tiny in the GEOstudio
# list foundation models populated with your Studio deployment
base = gfm_client.list_base_models(output='df')
display(base[['name','description','id','updated_at']])
# Create terramind Tiny base model entry in the Studio
create_terramind_tiny_backbone = {
"name": "terramind_v1_tiny",
"description": "Terramind Multimodal tiny backbone base",
"checkpoint_filename": "terramind_v1_tiny/terramind_v1_tiny.pt",
"model_params": {
"backbone": "terramind_v1_tiny",
"model_category": "terramind"
}
}
base_model = gfm_client.create_base_model(data=create_terramind_tiny_backbone)
# select base foundation model you just created for your fine-tuning task
base_model_id = base_model["id"]
Submitting the tune¶
Now we pull these choices together into a payload which we then submit to the platform. This will then deploy the job in the backend and we will see below how we can monitor it. First, we populate the payload so we can check it, then we simply submit.
# create the tune payload
tune_payload = {
"name": "test-fine-tuning-tiny",
"description": "Segmentation",
"dataset_id": selected_dataset,
"base_model_id": base_model_id,
"tune_template_id": task_id,
}
print(json.dumps(tune_payload, indent=2))
# submit tune
submitted = gfm_client.submit_tune(
data = tune_payload,
output = 'json'
)
print(submitted)
Monitoring training¶
Once the tune has been submitted you can check its status and monitor tuning progress through the SDK. You can also access the training metrics and images in MLflow. The get_tune function will give you the meta-data of the tune, including the status.
# get metadata about the submitted tune
tune_id = submitted.get("tune_id")
tune_info = gfm_client.get_tune(tune_id, output='json')
tune_info
Once the model has started training, you will also be able to access the training metrics. The get_tune_metrics_df function returns a dataframe containing the up-to-date training metrics, which you are free to explore and analyse. In addition to that, you can simply plot the training and validation loss and multi-class accuracy using the plot_tune_metrics function.
# get training metrics
mdf = gfm_client.get_mlflow_metrics(tune_id)
mdf.head()
# plot some basic training metrics
# gfm_client.plot_tune_metrics(tune_id)
Once your model is finished training and you are happy with the metrics (and images in MLflow), you can run some inference in test mode through the inference service.
Testing your model¶
Example flood events
| Location | Date | Bounding box | Link |
|---|---|---|---|
| Maiduguri, Nigeria | 2024-09-12 | [13.146418, 11.799808, 13.215874, 11.871586] | https://www.aljazeera.com/features/2024/9/19/a-disaster-homes-lost-relatives-missing-in-floods-in-northeast-nigeria |
| Porto Alegre, Brazil | 2024-05-06 | [-51.33225, -30.08903, -51.19011, -29.97489] | https://www.reuters.com/pictures/stunning-images-show-extent-flooding-southern-brazil-2024-05-07/ |
| Ahero, Kenya | 2024-05-05 | [34.838652, -0.231379, 34.977847, -0.131439] | |
| Gloucester, UK | 2024-01-09 | [-2.311807, 51.855573, -2.17892, 51.952735] |
Try out the model for inference¶
Once your model has finished tuning, if you want to run inference as a test you can do by passing either a location (bbox) or a url to a sample tiff file. For this walkthrough, we will run inference with the bbox option. The steps to test the model are:
- Define the inference payload
- Try out the tune
# define the inference payload
inference_payload = {
"spatial_domain": {
"bbox": [
[
34.709244,
-0.307616,
35.121231,
-0.065918
]
]
},
"temporal_domain": [
"2024-04-30_2024-05-07"
],
"model_input_data_spec": [
{
"bands": [
{
"index": "0",
"band_name": "coastal",
"scaling_factor": "1",
"description": ""
},
{
"index": "1",
"band_name": "blue",
"scaling_factor": "1",
"RGB_band": "B",
"description": ""
},
{
"index": "2",
"band_name": "green",
"scaling_factor": "1",
"RGB_band": "G",
"description": ""
},
{
"index": "3",
"band_name": "red",
"scaling_factor": "1",
"RGB_band": "R",
"description": ""
},
{
"index": "4",
"band_name": "rededge1",
"scaling_factor": "1",
"description": ""
},
{
"index": "5",
"band_name": "rededge2",
"scaling_factor": "1",
"description": ""
},
{
"index": "6",
"band_name": "rededge3",
"scaling_factor": "1",
"description": ""
},
{
"index": "7",
"band_name": "nir",
"scaling_factor": "1",
"description": ""
},
{
"index": "8",
"band_name": "nir08",
"scaling_factor": "1",
"description": ""
},
{
"index": "9",
"band_name": "nir09",
"scaling_factor": "1",
"description": ""
},
{
"index": "10",
"band_name": "swir16",
"scaling_factor": "1",
"description": ""
},
{
"index": "11",
"band_name": "swir22",
"scaling_factor": "1",
"description": ""
},
{
"index": "12",
"band_name": "scl",
"scaling_factor": "1",
"description": ""
}
],
"connector": "sentinel_aws",
"collection": "sentinel-2-l2a",
"file_suffix": "S2Hand",
"modality_tag": "S2L2A"
},
{
"bands": [
{
"index": "0",
"band_name": "VV",
"scaling_factor": 1,
"description": ""
},
{
"index": "1",
"band_name": "VH",
"scaling_factor": 1,
"description": ""
}
],
"connector": "sentinelhub",
"collection": "s1_grd",
"file_suffix": "S1Hand",
"modality_tag": "S1GRD",
"align_dates": "true",
"scaling_factor": [
1,
1
]
}
],
"geoserver_push": [
{
"workspace": "geofm",
"layer_name": "input_rgb",
"display_name": "Input image (RGB)",
"filepath_key": "model_input_original_image_rgb",
"file_suffix": "",
"z_index": 0,
"visible_by_default": "True",
"geoserver_style": {
"rgb": [
{
"minValue": 0,
"maxValue": "2000",
"channel": 1,
"label": "RedChannel"
},
{
"minValue": 0,
"maxValue": "2000",
"channel": 2,
"label": "GreenChannel"
},
{
"minValue": 0,
"maxValue": "2000",
"channel": 3,
"label": "BlueChannel"
}
]
}
},
{
"workspace": "geofm",
"layer_name": "pred",
"display_name": "Model prediction",
"filepath_key": "model_output_image_masked",
"file_suffix": "",
"z_index": 1,
"visible_by_default": "True",
"geoserver_style": {
"segmentation": [
{
"color": "#808080",
"label": "No flood",
"opacity": "0",
"quantity": "0"
},
{
"color": "#FA4D56",
"label": "Flood",
"opacity": 1,
"quantity": "1"
},
{
"color": "#4589FF",
"label": "Permanent water",
"opacity": 1,
"quantity": 997
},
{
"color": "#FFFAFA",
"label": "Snow/ice",
"opacity": 1,
"quantity": 998
},
{
"color": "#CCCCCC",
"label": "Clouds",
"opacity": 1,
"quantity": 999
}
]
}
}
],
"model_display_name": "geofm-sandbox-models",
"fine_tuning_id": tune_id,
"description": "Ahero: Kisumu Inference",
"location": "Kamahawa /kasiwindi, Ahero ward, Kenya, "
}
Once you have registered the model, you can now run. You can then monitor it and visualise the outputs either through the SDK, or in the UI.
# Now submit the test inference request
inference_response = gfm_client.try_out_tune(tune_id=tune_id, data=inference_payload)
inference_response
Monitoring your inference task and Visualize results¶
Once submitted you can check on progress from the UI side. Log into the UI of your deployed GEOStudio instance and click on the Inference lab page. Click on the history icon on the top left and you can click on the inference task you submitted to see its progress as well as visualize results when complete
# get metadata about the inference task
gfm_client.get_inference(inference_response['id'])
Checking model outputs¶
You can check out the results visually in the Studio UI, or with the quick widget below. You can alternatively use the SDK to download selected files for further analysis see documentation.
Note:
For this walkthrough, you can check out the inference output and results visually in the Studio UI through the history tab of the inference page.
# view inference results
# gswidgets.inferenceViewer(gfm_client, inference_response['id'])