diff --git a/deploy_studio_k8s.sh b/deploy_studio_k8s.sh index 1370e32d..c2e4f37b 100755 --- a/deploy_studio_k8s.sh +++ b/deploy_studio_k8s.sh @@ -291,14 +291,35 @@ sed -i -e "s||${pgbouncer_port}|g" workspace/${DEPLOYMENT_ENV}/v sed -i -e "s||${pgbouncer_username}|g" workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml sed -i -e "s||${pgbouncer_password}|g" workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml +# The line below removes GPUs from the pipeline components and Finetuning job, to leave GPUs activated, copy out this line + +# Call the function +get_menu_selection \ + "Select whether you have GPU available in your cluster: " \ + gpu_configuration_type \ + "$gpu_configuration_options" + # The line below removes GPUs from the pipeline components, to leave GPUs activated, copy out this line NVIDIA_GPUS_AVAILABLE=$(kubectl describe node ${CLUSTER_NODE_NAME} | grep -c "nvidia.com") -if [ "$NVIDIA_GPUS_AVAILABLE" -gt 0 ]; then + +if [[ "$gpu_configuration_type" == "GPU-Available" && "$NVIDIA_GPUS_AVAILABLE" -gt 0 ]]; then + # Get number of GPUs echo "Cluster Type: nvkind" python ./deployment-scripts/remove-pipeline-gpu.py --remove-affinity-only workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml + + # Keep the Job GPU configuration as is. + echo -e "\n Keeping GPU configuration for Finetuning job in values.yaml. You can update these later in workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml " + echo -e "and update the cluster later using: helm upgrade geospatial-studio ./geospatial-studio/ \n" else echo "Cluster Type: standard kind" python ./deployment-scripts/remove-pipeline-gpu.py workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml + + # remove job GPU request + echo -e "\n Removing GPU configuration from values.yaml" + python ./deployment-scripts/update_jobs_gpu.py --filename workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml \ + --gpu-limit 0 \ + --gpu-request 0 + echo -e "--------------------------- Removed GPUs in the Cluster ------------------- \n" fi echo "**********************************************************************" @@ -312,6 +333,64 @@ if [[ "${NON_INTERACTIVE:-false}" != "true" ]]; then read ans fi +echo "**********************************************************************" +echo "**********************************************************************" +echo "------ Configure Fine-Tuning Job Resources -------------------------" +echo "**********************************************************************" +echo "**********************************************************************" + + +# Ask user if they want to alter memory, CPU requests and limits for finetuning. +if [[ "${NON_INTERACTIVE:-false}" != "true" ]]; then + printf "%s " "Do you want to alter memory, CPU requests and limits for finetuning? (y/n) " + read ans +else + # Non-interactive mode: use CONFIGURE_RESOURCES environment variable (default to "n") + ans="${CONFIGURE_RESOURCES:-n}" + echo "Non-interactive mode: CONFIGURE_RESOURCES=$ans" +fi + +# If yes, prompt user for memory limit, CPU limit, memory request and CPU request. +if [ "$ans" = "y" ]; then + echo "Updating memory, CPU requests and limits for finetuning." + echo "" + + # Prompt for CPU limit + printf "%s " "CPU limit in cores (default: 4): " + read cpu_limit + cpu_limit=${cpu_limit:-4} + + # Prompt for CPU request + printf "%s " "CPU request in cores (default: 2): " + read cpu_request + cpu_request=${cpu_request:-2} + + # Prompt for Memory limit + printf "%s " "Memory limit in GB (default: 10): " + read memory_limit + memory_limit=${memory_limit:-10} + + # Prompt for Memory request + printf "%s " "Memory request in GB (default: 6): " + read memory_request + memory_request=${memory_request:-6} + + echo -e "\n Applying configuration:" + echo " CPU Limit: ${cpu_limit} cores, CPU Request: ${cpu_request} cores" + echo -e " Memory Limit: ${memory_limit}GB, Memory Request: ${memory_request}GB \n" + + # Call the update script with user-provided values + python3 ./deployment-scripts/update_jobs_gpu.py --filename workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml \ + --cpu-limit "$cpu_limit" \ + --cpu-request "$cpu_request" \ + --memory-limit "$memory_limit" \ + --memory-request "$memory_request" + echo -e "\n Updated finetuning resource configurations \n" +else + echo -e "\n Not updating resource configurations." + echo "You can manually edit workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml" + echo -e "and update the cluster later using: helm upgrade geospatial-studio ./geospatial-studio/ \n" +fi echo "----------------------------------------------------------------------" echo "---------------- Building Helm dependencies ------------------------" diff --git a/deploy_studio_lima.sh b/deploy_studio_lima.sh index 3f00ab32..15ba2067 100755 --- a/deploy_studio_lima.sh +++ b/deploy_studio_lima.sh @@ -278,6 +278,87 @@ if [[ "${NON_INTERACTIVE:-false}" != "true" ]]; then read ans fi +echo "**********************************************************************" +echo "**********************************************************************" +echo "------ Configure Fine-Tuning Job Resources -------------------------" +echo "**********************************************************************" +echo "**********************************************************************" + +# Ask user if they have GPUs in their cluster. If Yes, keep configuration as is. If No, remove GPU configuration via the values.yaml +if [[ "${NON_INTERACTIVE:-false}" != "true" ]]; then + printf "%s " "Do you have GPUs in your cluster? (y/n): " + read ans +else + # Non-interactive mode: use HAS_GPU environment variable (default to "n" for no GPU) + ans="${HAS_GPU:-n}" + echo "Non-interactive mode: HAS_GPU=$ans" +fi + +if [ "$ans" = "y" ]; then + echo -e "\n Keeping GPU configuration in values.yaml. You can update these later in workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml " + echo -e "and update the cluster later using: helm upgrade geospatial-studio ./geospatial-studio/ \n" +else + echo -e "\n Removing GPU configuration from values.yaml" + python ./deployment-scripts/update_jobs_gpu.py --filename workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml \ + --gpu-limit 0 \ + --gpu-request 0 + echo -e "\n --------------------------- Removed GPUs in the Cluster ------------------- \n" +fi + + +# Ask user if they want to alter memory, CPU requests and limits for finetuning. +if [[ "${NON_INTERACTIVE:-false}" != "true" ]]; then + printf "%s " "Do you want to alter memory, CPU requests and limits for finetuning? (y/n) " + read ans +else + # Non-interactive mode: use CONFIGURE_RESOURCES environment variable (default to "n") + ans="${CONFIGURE_RESOURCES:-n}" + echo "Non-interactive mode: CONFIGURE_RESOURCES=$ans" +fi + +# If yes, prompt user for memory limit, CPU limit, memory request and CPU request. +if [ "$ans" = "y" ]; then + echo "Updating memory, CPU requests and limits for finetuning." + echo "" + + # Prompt for CPU limit + printf "%s " "CPU limit in cores (default: 4): " + read cpu_limit + cpu_limit=${cpu_limit:-4} + + # Prompt for CPU request + printf "%s " "CPU request in cores (default: 2): " + read cpu_request + cpu_request=${cpu_request:-2} + + # Prompt for Memory limit + printf "%s " "Memory limit in GB (default: 10): " + read memory_limit + memory_limit=${memory_limit:-10} + + # Prompt for Memory request + printf "%s " "Memory request in GB (default: 6): " + read memory_request + memory_request=${memory_request:-6} + + echo -e "\n Applying configuration:" + echo " CPU Limit: ${cpu_limit} cores, CPU Request: ${cpu_request} cores" + echo -e " Memory Limit: ${memory_limit}GB, Memory Request: ${memory_request}GB \n" + + # Call the update script with user-provided values + python3 ./deployment-scripts/update_jobs_gpu.py --filename workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml \ + --cpu-limit "$cpu_limit" \ + --cpu-request "$cpu_request" \ + --memory-limit "$memory_limit" \ + --memory-request "$memory_request" + echo -e " \n Updated finetuning resource configurations \n" +else + echo -e "\n Not updating resource configurations" + echo "You can manually edit workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml" + echo -e "and update the cluster later using: helm upgrade geospatial-studio ./geospatial-studio/ \n" +fi + + echo "----------------------------------------------------------------------" echo "---------------- Building Helm dependencies ------------------------" diff --git a/deploy_studio_ocp.sh b/deploy_studio_ocp.sh index 38c9b34b..dee725f1 100755 --- a/deploy_studio_ocp.sh +++ b/deploy_studio_ocp.sh @@ -271,6 +271,8 @@ EOF echo "-------------------------- COS_STORAGE_CLASS -------------------------------------" echo "------------------------ NON_COS_STORAGE_CLASS ---------------------------------" echo "***********************************************************************************" + echo "-- Check StorageClasses values in the cluster for COS storage and block storage ---" + while true; do printf "%s " "Press enter to continue" @@ -459,7 +461,7 @@ EOF # For CRC, we need volume permissions enabled, so don't use DO_NOT_SET_SCC # For other OpenShift environments, storage may be pre-configured - if [[ "$DEPLOYMENT_ENV" == "crc" ]] || [[ "$DEPLOYMENT_ENV" == "crc-local" ]]; then + if ([[ "$DEPLOYMENT_ENV" == "crc" ]] || [[ "$DEPLOYMENT_ENV" == "crc-local" ]]) && [[ "$OC_PROJECT" == "default" ]]; then ./deployment-scripts/install-postgres.sh UPDATE_STORAGE DISABLE_PV else ./deployment-scripts/install-postgres.sh UPDATE_STORAGE DISABLE_PV DO_NOT_SET_SCC @@ -791,8 +793,19 @@ EOF if [[ "$gpu_configuration_type" == "GPU-Available" ]]; then python ./deployment-scripts/remove-pipeline-gpu.py --remove-affinity-only workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml + + # Keep the Job GPU configuration as is. + echo "Keeping GPU configuration for Finetuning job in values.yaml. You can update these later in workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml " + echo "and update the cluster later using: helm upgrade geospatial-studio ./geospatial-studio/" else python ./deployment-scripts/remove-pipeline-gpu.py workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml + + # remove job GPU request + echo "Removing GPU configuration from values.yaml" + python ./deployment-scripts/update_jobs_gpu.py --filename workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml \ + --gpu-limit 0 \ + --gpu-request 0 + echo "--------------------------- Removed GPUs in the Cluster -------------------" fi else @@ -819,8 +832,70 @@ echo "----------- Make any changes to deployment values yaml --------------" echo "**********************************************************************" echo "**********************************************************************" -printf "%s " "Press enter to continue" -read ans +if [[ "${NON_INTERACTIVE:-false}" != "true" ]]; then + printf "%s " "Press enter to continue" + read ans +fi + +echo "**********************************************************************" +echo "**********************************************************************" +echo "------ Configure Fine-Tuning Job Resources -------------------------" +echo "**********************************************************************" +echo "**********************************************************************" + + +# Ask user if they want to alter memory, CPU requests and limits for finetuning. +configure_resources_options="No Yes" +typeset configure_resources + +# Call the function +get_menu_selection \ + "Do you want to alter memory, CPU requests and limits for finetuning?" \ + configure_resources \ + "$configure_resources_options" + +# If yes, prompt user for memory limit, CPU limit, memory request and CPU request. +if [ "$configure_resources" = "Yes" ]; then + echo "Updating memory, CPU requests and limits for finetuning." + echo "" + + # Prompt for CPU limit + printf "%s " "CPU limit in cores (default: 4): " + read cpu_limit + cpu_limit=${cpu_limit:-4} + + # Prompt for CPU request + printf "%s " "CPU request in cores (default: 2): " + read cpu_request + cpu_request=${cpu_request:-2} + + # Prompt for Memory limit + printf "%s " "Memory limit in GB (default: 10): " + read memory_limit + memory_limit=${memory_limit:-10} + + # Prompt for Memory request + printf "%s " "Memory request in GB (default: 6): " + read memory_request + memory_request=${memory_request:-6} + + echo -e "\n Applying configuration:" + echo " CPU Limit: ${cpu_limit} cores, CPU Request: ${cpu_request} cores" + echo -e " Memory Limit: ${memory_limit}GB, Memory Request: ${memory_request}GB \n" + + # Call the update script with user-provided values + python3 ./deployment-scripts/update_jobs_gpu.py --filename workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml \ + --cpu-limit "$cpu_limit" \ + --cpu-request "$cpu_request" \ + --memory-limit "$memory_limit" \ + --memory-request "$memory_request" + echo -e " \n Updated finetuning resource configurations \n" +else + echo -e "\n Not updating resource configurations." + echo "You can manually edit workspace/${DEPLOYMENT_ENV}/values/geospatial-studio/values-deploy.yaml" + echo -e "and update the cluster later using: helm upgrade geospatial-studio ./geospatial-studio/ \n" +fi + echo "----------------------------------------------------------------------" echo "---------------- Building Helm dependencies ------------------------" diff --git a/geospatial-studio/values.yaml b/geospatial-studio/values.yaml index 8c3e1b9e..4944bbdf 100644 --- a/geospatial-studio/values.yaml +++ b/geospatial-studio/values.yaml @@ -206,6 +206,17 @@ gfm-studio-gateway: ## If multiple values are given, the pod can be scheduled on any node ## matching one of them. # NODE_GPU_SPEC: NVIDIA-A100-SXM4-80GB + # Use defaults in the defined jobs + # Resource requests (guaranteed minimum for fine-tuning jobs) + RESOURCE_LIMIT_GPU: 1 # Number of GPUs + RESOURCE_REQUEST_GPU: 1 # Number of GPUs + + RESOURCE_LIMIT_CPU: 4 # CPU cores + RESOURCE_REQUEST_CPU: 2 + + RESOURCE_LIMIT_Memory: 10 # Memory in GB + RESOURCE_REQUEST_Memory: 6 # Memory in GB + # geofm-ui geofm-ui: diff --git a/populate-studio/payloads/datasets/dataset-flooding_multimodal.json b/populate-studio/payloads/datasets/dataset-flooding_multimodal.json index 6b4739b4..f9afaa56 100644 --- a/populate-studio/payloads/datasets/dataset-flooding_multimodal.json +++ b/populate-studio/payloads/datasets/dataset-flooding_multimodal.json @@ -5,94 +5,94 @@ "bands": [ { "index": "0", - "band_name": "Coastal_aerosol", + "band_name": "B01", "description": "", "scaling_factor": "1" }, { "index": "1", - "band_name": "Blue", + "band_name": "B02", "RGB_band": "B", "description": "", "scaling_factor": "1" }, { "index": "2", - "band_name": "Green", + "band_name": "B03", "RGB_band": "G", "description": "", "scaling_factor": "1" }, { "index": "3", - "band_name": "Red", + "band_name": "B04", "RGB_band": "R", "description": "", "scaling_factor": "1" }, { "index": "4", - "band_name": "05_-_Vegetation_Red_Edge", + "band_name": "B05", "description": "", "scaling_factor": "1" }, { "index": "5", - "band_name": "06_-_Vegetation_Red_Edge", + "band_name": "B06", "description": "", "scaling_factor": "1" }, { "index": "6", - "band_name": "07_-_Vegetation_Red_Edge", + "band_name": "B07", "description": "", "scaling_factor": "1" }, { "index": "7", - "band_name": "08_-_NIR", + "band_name": "B08", "description": "", "scaling_factor": "1" }, { "index": "8", - "band_name": "08A_-_Vegetation_Red_Edge", + "band_name": "B8A", "description": "", "scaling_factor": "1" }, { "index": "9", - "band_name": "09_-_Water_vapour", + "band_name": "B09", "description": "", "scaling_factor": "1" }, { "index": "10", - "band_name": "11_-_SWIR", + "band_name": "B11", "description": "", "scaling_factor": "1" }, { "index": "11", - "band_name": "12_-_SWIR", + "band_name": "B12", "description": "", "scaling_factor": "1" }, { "index": "12", - "band_name": "Cloud_Probability", + "band_name": "CLD", "description": "", "scaling_factor": "1" } ], "connector": "sentinelhub", "collection": "s2_l2a", - "modality_tag": "S2L1C", + "modality_tag": "S2L2A", "file_suffix": "_S2Hand.tif" }, { "bands": [ - {"index": "0", "band_name": "VV (Gray)", "description": ""}, + {"index": "0", "band_name": "VV", "description": ""}, {"index": "1", "band_name": "VH", "description": ""} ], "connector": "sentinelhub", @@ -107,7 +107,7 @@ {"id": "0", "name": "No Floods", "description": "Flooding assets"}, {"id": "1", "name": "Floods", "description": "Flooding assets"} ], - "dataset_url": "https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/sen1floods11-studio.zip", + "dataset_url": "https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/sen1floods11_v1.1.4_zip.zip", "description": "Flood data from places", "label_suffix": "_LabelHand.tif", "purpose": "Segmentation" diff --git a/populate-studio/payloads/templates/template-terramind_seg.json b/populate-studio/payloads/templates/template-terramind_seg.json index 38b69e6f..494bf49e 100644 --- a/populate-studio/payloads/templates/template-terramind_seg.json +++ b/populate-studio/payloads/templates/template-terramind_seg.json @@ -367,7 +367,7 @@ "description": "A request sent to the finetuning service to start a finetune task for segmentation" }, "extra_info": { - "runtime_image": "quay.io/geospatial-studio/terratorch:terratorch_v_1_2_2", + "runtime_image": "quay.io/geospatial-studio/terratorch:latest", "model_category": "terramind" }, "content": "# © Copyright IBM Corporation 2025
# SPDX-License-Identifier: Apache-2.0


# lightning.pytorch==2.1.1
seed_everything: 42
trainer:
  accelerator: auto
  strategy: auto
  devices: auto
  num_nodes: 1
  precision: 16-mixed
  logger:
    class_path: lightning.pytorch.loggers.mlflow.MLFlowLogger
    init_args:
      experiment_name: {{ tune_id }} 
      run_name: "Train"    
      tracking_uri: {{ mlflow_tracking_url }}
      save_dir: {{ mount_root + 'tune-tasks/' + tune_id + '/mlflow' }}
      {% if mlflow_tags -%}
      tags:
        {% for key, value in mlflow_tags.items() -%}
        {{ key }}: {{ value }}
        {% endfor %}
      {%- endif %}       
  callbacks:
    - class_path: RichProgressBar
    - class_path: LearningRateMonitor
      init_args:
        logging_interval: epoch
    # ---- Early stop if ----
    {% if runner["early_stopping_patience"] -%}
    - class_path: EarlyStopping
      init_args:
        monitor: {{ runner["early_stopping_monitor"] }}
        patience: {{ runner["early_stopping_patience"] }}
    {%- endif %}
     # ---- Early stop endif ----
    - class_path: ModelCheckpoint
      init_args:
        dirpath: {{ mount_root + 'tune-tasks/' + tune_id  + '/' }}
        mode: min
        monitor: val/loss
        filename: {{ 'best-state_dict-{epoch:02d}' }}
        save_weights_only: True
  max_epochs: {{ runner["max_epochs"] }}
  check_val_every_n_epoch: {{ evaluation["interval"] }}
  log_every_n_steps: 50
  enable_checkpointing: true
  default_root_dir: {{ mount_root + 'tune-tasks/' + tune_id }}
data:
  {% if image_modalities|length == 1 %} 
  class_path: GenericNonGeoSegmentationDataModule
  init_args:
    {% if data["check_stackability"] -%}
    check_stackability: {{ data["check_stackability"] }}
    {% else %}
    check_stackability: false
    {% endif -%}
    batch_size: 2
    num_workers: 1
    dataset_bands:  # Dataset bands
      {{ bands.values() | list | first | to_yaml | indent(6) }}
    rgb_indices:
      {{ rgb_band_indices | to_yaml | indent(6) }}
    train_data_root: {{ data_root }}{{ train_data_dir.values() | list | first }}
    val_data_root: {{ data_root }}{{ val_data_dir.values() | list | first }}
    test_data_root: {{ data_root }}{{ test_data_dir.values() | list | first }}
    # labels roots
    train_label_data_root: {{ data_root + train_labels_dir }}
    val_label_data_root: {{ data_root + val_labels_dir }}
    test_label_data_root: {{ data_root + test_labels_dir }}
    {% if train_split_path -%}
    train_split: {{ data_root + train_split_path }}
    {% endif -%}
    {% if test_split_path -%}
    test_split: {{ data_root + test_split_path }}
    {% endif -%}
    {% if val_split_path -%}
    val_split: {{ data_root + val_split_path }}
    {% endif -%}

    {% if img_suffix -%}
    img_grep:  {{ img_suffix.values() | list | first | tojson }}
    {% endif -%}
    {% if seg_map_suffix -%}
    label_grep: "{{ seg_map_suffix }}"
    {% endif -%}
    means: 
      {{ norm_means.values() | list | first| to_yaml | indent(6) }}
    stds: 
      {{ norm_stds.values() | list | first | to_yaml | indent(6) }}
    num_classes: {{ classes|length }}
    train_transform:
      - class_path: albumentations.D4
      - class_path: ToTensorV2
    no_data_replace: 0
    no_label_replace: -1


  {% else %}
  class_path: terratorch.datamodules.GenericMultiModalDataModule
  init_args:
    {% if data["check_stackability"] -%}
    check_stackability: {{ data["check_stackability"] }}
    {% else %}
    check_stackability: false
    {% endif -%}
    # Config for only segmentation. No need to automate this. 
    task: 'segmentation'
    # Out of cuda error for anything > 2
    # ToDo: Figure out why batch_size replacement is not working.
    batch_size: 2
    {% if num_workers -%}
    num_workers: {{ num_workers }}
    {% else -%}
    num_workers: 2
    {% endif -%}
    no_label_replace: {{ label_nodata }}
    no_data_replace: {{ image_nodata_replace }}
    dataset_bands:
      {{ bands | to_yaml | indent(6)}}
    output_bands:
      {{ output_bands | to_yaml  | indent(6)}}
    modalities:
      {{ image_modalities | to_yaml | indent(6) }}
    rgb_modality: {{ rgb_modality }}
    rgb_indices: 
      {{ rgb_band_indices | to_yaml | indent(6) }}
    train_data_root:
      {% for key, val in train_data_dir.items() -%}
       {{ key }}: {{ data_root }}{{ val }}
      {% endfor %}
    train_label_data_root:  {{ data_root +  train_labels_dir }}
    val_data_root:
      {% for key, val in val_data_dir.items() -%}
       {{ key }}: {{ data_root }}{{ val }}
      {% endfor %}
    val_label_data_root: {{ data_root +  test_labels_dir }}
    test_data_root:
      {% for key, val in test_data_dir.items() -%}
       {{ key }}: {{ data_root }}{{ val }}
      {% endfor %}
    test_label_data_root: {{ data_root + test_labels_dir }}
    {% if train_split_path -%}
    train_split: {{  data_root + train_split_path }}
    {% endif -%}
    {% if test_split_path -%}
    test_split: {{  data_root + test_split_path }}
    {% endif -%}
    {% if val_split_path -%}
    val_split: {{  data_root + val_split_path }}
    {% endif -%}
    {% if img_suffix -%}
    image_grep: 
      {{ img_suffix | to_yaml | indent(6) }}
    {% endif -%}
    {% if seg_map_suffix -%}
    label_grep: "{{ seg_map_suffix }}"
    {% endif -%}

    num_classes: {{ classes|length }}
    {% if data["expand_temporal_dimension"] is not none -%}
    expand_temporal_dimension: data["expand_temporal_dimension"]
    {% endif -%}
    {% if data["drop_last"] is not none -%}
    drop_last: data["drop_last"]
    {% endif -%}

    means: 
      {{ norm_means | to_yaml | indent(6) }}
    stds: 
      {{ norm_stds | to_yaml | indent(6) }}
    train_transform:
      - class_path: albumentations.D4  # Random flip and rotations
      - class_path: albumentations.pytorch.transforms.ToTensorV2
  {% endif %}
model:
  class_path: terratorch.tasks.SemanticSegmentationTask
  init_args:
    model_factory: EncoderDecoderFactory
    model_args:
      backbone: {{ pretrained_model_name }}
      #  terramind_v1_base  # large version: terramind_v1_large
      backbone_pretrained: true
      {%- if image_modalities %}
      backbone_modalities:
        {{ image_modalities | to_yaml  | indent(8)}}
      {%- endif %}
      {% if  image_modalities|length > 1 %}
      backbone_merge_method: mean
      {% endif %}
      backbone_bands: 
        {{ output_bands | to_yaml | indent(8) }}
      necks:
        {%- if pretrained_model_name == "terramind_v1_tiny" %}
        - name: SelectIndices
          indices: [2, 5, 8, 11]  # tiny version
        {%- elif pretrained_model_name == "terramind_v1_base" %}
        - name: SelectIndices
          indices: [2, 5, 8, 11]  # base version
        {%- elif pretrained_model_name == "terramind_v1_large" %}
        - name: SelectIndices
          indices: [5, 11, 17, 23]  # large version
        {% endif %}
        - name: ReshapeTokensToImage
          remove_cls_token: False
        - name: LearnedInterpolateToPyramidal

      decoder: {{ model["decode_head"]["decoder"] }}
      # UNetDecoder
      {% if  model["decode_head"]["decoder"] == "UperNetDecoder" -%}
      decoder_channels: [512, 256, 128, 64]
      {% elif  model["decode_head"]["decoder"] == "UNetDecoder" -%}
      #TODO user provided channels
      decoder_channels: [512, 256, 128, 64]
      {% else %}
      decoder_channels: {{ model["decode_head"]["channels"] }}
      {% endif -%}
      head_dropout: 0.1
      num_classes: {{ classes|length }}
    loss: {{ model["decode_head"]["loss_decode"]["type"] }}
    plot_on_val: {{ runner["plot_on_val"] }}
    #  dice
    ignore_index: {{ ignore_index }}
    freeze_backbone:  {{ model["frozen_backbone"] | lower }}
    freeze_decoder: false
    {% if  model["class_names"] %}
    class_names: {{ model["class_names"] | to_yaml | indent(6)}}
    {% endif %}
    # ---- optimizer start ----
    {% if model["optimizer"] -%}
    optimizer: {{ model["optimizer"]["type"] }}
    lr: {{ model["optimizer"]["lr"] | float }}
    {% endif -%}
    # ---- optimizer end ----
    {% if model["tiled_inference_parameters"] %}
    tiled_inference_parameters: 
      h_crop: {{ model["tiled_inference_parameters"]["h_crop"] | int}}
      h_stride: {{ model["tiled_inference_parameters"]["h_stride"] | int }}
      w_crop: {{ model["tiled_inference_parameters"]["w_crop"] | int}}
      w_stride: {{ model["tiled_inference_parameters"]["w_stride"] | int }}
      average_patches: {{ model["tiled_inference_parameters"]["average_patches"] }}
    {% else %}
    # ToDo: Remove the tiled_inference if user not provided. 
    tiled_inference_parameters: 
      h_crop: 224
      # stride logic = would be h_crop - h_crop * 0.125
      h_stride: 208
      w_crop: 224
      # stride logic = would be w_crop - w_crop * 0.125
      w_stride: 208
      average_patches: true
    {% endif %}

optimizer:
  class_path: torch.optim.AdamW
  init_args:
    {% if optimizer["lr"] -%}
    lr: {{ optimizer["lr"] | float }}
    {% else %}
    lr: 2.e-5
    {% endif -%}
    {%- if optimizer["weight_decay"] -%}
    weight_decay: {{ optimizer["weight_decay"] }}
    {%- else -%}
    weight_decay: 0.05
    {% endif %}
lr_scheduler:
  class_path: ReduceLROnPlateau
  init_args:
    monitor: val/loss
    factor: 0.5
    patience: 5
" diff --git a/populate-studio/payloads/tunes/tune-prithvi-eo-flood.json b/populate-studio/payloads/tunes/tune-prithvi-eo-flood.json index 7a86580c..46a2f5df 100644 --- a/populate-studio/payloads/tunes/tune-prithvi-eo-flood.json +++ b/populate-studio/payloads/tunes/tune-prithvi-eo-flood.json @@ -8,40 +8,40 @@ "bands": [ { "index": 0, - "RGB_band": "R", - "band_name": "B02", + "RGB_band": "B", + "band_name": "blue", "scaling_factor": 1 }, { "index": 1, "RGB_band": "G", - "band_name": "B03", + "band_name": "green", "scaling_factor": 1 }, { "index": 2, - "RGB_band": "B", - "band_name": "B04", + "RGB_band": "R", + "band_name": "red", "scaling_factor": 1 }, { "index": 3, - "band_name": "B8A", + "band_name": "nir08", "scaling_factor": 1 }, { "index": 4, - "band_name": "B11", + "band_name": "swir16", "scaling_factor": 1 }, { "index": 5, - "band_name": "B12", + "band_name": "swir22", "scaling_factor": 1 }, { "index": 6, - "band_name": "SCL", + "band_name": "scl", "scaling_factor": 1 } ], diff --git a/populate-studio/payloads/tunes/tune-test-fire.json b/populate-studio/payloads/tunes/tune-test-fire.json index e19088a5..53c77290 100644 --- a/populate-studio/payloads/tunes/tune-test-fire.json +++ b/populate-studio/payloads/tunes/tune-test-fire.json @@ -1,6 +1,119 @@ { "name": "prithvi-eo-fire-tune", "description": "prithvi-eo-fire-tune", - "tune_config_url": "https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/prithvi-eo-burnscars/config_deploy.yaml", - "tune_checkpoint_url": "https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/prithvi-eo-burnscars/best-state_dict-epoch=09.ckpt" + "tune_config_url": "https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/prithvi-eo-300-fire/config_deploy.yaml", + "tune_checkpoint_url": "https://geospatial-studio-example-data.s3.us-east.cloud-object-storage.appdomain.cloud/prithvi-eo-300-fire/best-state_dict-epoch=91.ckpt", + "post_processing": { + "cloud_masking": false, + "ocean_masking": false, + "snow_ice_masking": false, + "permanent_water_masking": false + }, + "model_input_data_spec": [ + { + "bands": [ + { + "index": "0", + "RGB_band": "B", + "band_name": "Blue", + "scaling_factor": "0.0001" + }, + { + "index": "1", + "RGB_band": "G", + "band_name": "Green", + "scaling_factor": "0.0001" + }, + { + "index": "2", + "RGB_band": "R", + "band_name": "Red", + "scaling_factor": "0.0001" + }, + { + "index": "3", + "band_name": "NIR_Narrow", + "scaling_factor": "0.0001" + }, + { + "index": "4", + "band_name": "SWIR1", + "scaling_factor": "0.0001" + }, + { + "index": "5", + "band_name": "SWIR2", + "scaling_factor": "0.0001" + } + ], + "connector": "sentinelhub", + "collection": "hls_l30", + "file_suffix": "_merged.tif", + "modality_tag": "HLS_L30" + } + ], + "geoserver_push": [ + { + "z_index": 0, + "workspace": "geofm", + "layer_name": "input_rgb", + "file_suffix": "", + "display_name": "Input image (RGB)", + "filepath_key": "model_input_original_image_rgb", + "geoserver_style": { + "rgb": [ + { + "label": "RedChannel", + "channel": 1, + "maxValue": 255, + "minValue": 0 + }, + { + "label": "GreenChannel", + "channel": 2, + "maxValue": 255, + "minValue": 0 + }, + { + "label": "BlueChannel", + "channel": 3, + "maxValue": 255, + "minValue": 0 + } + ] + }, + "visible_by_default": "True" + }, + { + "z_index": 1, + "workspace": "geofm", + "layer_name": "pred", + "file_suffix": "", + "display_name": "Model prediction", + "filepath_key": "model_output_image", + "geoserver_style": { + "segmentation": [ + { + "color": "#000000", + "label": "ignore", + "opacity": "0", + "quantity": "-1" + }, + { + "color": "#000000", + "label": "no-data", + "opacity": "0", + "quantity": "0" + }, + { + "color": "#ab4f4f", + "label": "fire-scar", + "opacity": 1, + "quantity": "1" + } + ] + }, + "visible_by_default": "True" + } + ] }