Segmentation head fine tuning#
We have built an example for training a segmentation head on top of feature maps extracted from the frozen Clay encoder.
All the code for this example can be found in the segment finetuning folder of this repository.
Segmentor#
The Segmentor
class is designed for semantic segmentation tasks, extracting feature maps from intermediate layers of the Clay Encoder and adding a Feature Pyramid Network (FPN) on top of it.
The decoder in this example is inspired by the Segformer paper.
A possible future improvement will be to add a neck & head for segmentation task from other papers like UperNet, PPANet, etc. to compare with other GeoAI models.
Parameters#
feature_maps (list)
: Indices of intermediate layers of the Clay Encoder used by FPN layers.ckpt_path (str)
: Path to the Clay model checkpoint.
Example#
In this example, we will use the Segmentor
class to segment Land Use Land Cover (LULC) classes for the Chesapeake Bay CVPR dataset. The implementation includes data preprocessing, data loading, and model training workflow using PyTorch Lightning.
Dataset citation#
If you use this dataset, please cite the associated manuscript:
Robinson C, Hou L, Malkin K, Soobitsky R, Czawlytko J, Dilkina B, Jojic N. Large Scale High-Resolution Land Cover Mapping with Multi-Resolution Data. Proceedings of the 2019 Conference on Computer Vision and Pattern Recognition (CVPR 2019).
Dataset URL: Chesapeake Bay Land Cover Dataset
Setup#
Follow the instructions in the README to install the required dependencies.
git clone <repo-url>
cd model
mamba env create --file environment.yml
mamba activate claymodel
Usage#
Preparing the Dataset#
Download the Chesapeake Bay Land Cover dataset and organize your dataset directory as recommended.
Copy
*_lc.tif
and*_naip-new.tif
files for segmentation downstream tasks using s5cmd:# train s5cmd --no-sign-request cp --include "*_lc.tif" --include "*_naip-new.tif" "s3://us-west-2.opendata.source.coop/agentmorris/lila-wildlife/lcmcvpr2019/cvpr_chesapeake_landcover/ny_1m_2013_extended-debuffered-train_tiles/*" data/cvpr/files/train/ # val s5cmd --no-sign-request cp --include "*_lc.tif" --include "*_naip-new.tif" "s3://us-west-2.opendata.source.coop/agentmorris/lila-wildlife/lcmcvpr2019/cvpr_chesapeake_landcover/ny_1m_2013_extended-debuffered-val_tiles/*" data/cvpr/files/val/
Create chips of size
224 x 224
to feed them to the model:python finetune/segment/preprocess_data.py data/cvpr/files data/cvpr/ny 224
Directory structure:
data/
└── cvpr/
├── files/
│ ├── train/
│ └── val/
└── ny/
├── train/
│ ├── chips/
│ └── labels/
└── val/
├── chips/
└── labels/
Training the Model#
The model can be run via LightningCLI using configurations in finetune/segment/configs/segment_chesapeake.yaml
.
Download the Clay model checkpoint from Huggingface model hub and save it in the
checkpoints/
directory.Modify the batch size, learning rate, and other hyperparameters in the configuration file as needed:
data: batch_size: 40 num_workers: 8 model: num_classes: 7 feature_maps: - 3 - 5 - 7 - 11 ckpt_path: checkpoints/clay-v1-base.ckpt lr: 1e-5 wd: 0.05 b1: 0.9 b2: 0.95
Update the WandB logger configuration in the configuration file with your WandB details or use CSV Logger if you don’t want to log to WandB:
logger: - class_path: lightning.pytorch.loggers.WandbLogger init_args: entity: <wandb-entity> project: <wandb-project> log_model: false
Train the model:
To ensure that the imports work properly, ensure that the root of the repository is in the python path before running the script.
python -m finetune.segment.segment fit --config configs/segment_chesapeake.yaml
Acknowledgments#
Decoder implementation is inspired by the Segformer paper:
Segformer: Simple and Efficient Design for Semantic Segmentation with Transformers
Paper URL: https://arxiv.org/abs/2105.15203