SynRS3D: A Synthetic Dataset for Global 3D Semantic Understanding from Monocular Remote Sensing Imagery
Creators
Description
Checkpoints of RS3DAda model trained on the SynRS3D dataset
Neural Information Processing Systems (Spotlight), 2024
For more details, please refer to our paper and visit our GitHub repository.
Overview
TL;DR:
We are excited to release two high-performing models for height estimation and land cover mapping. These models were trained on the SynRS3D dataset using our novel domain adaptation method, RS3DAda.
- Encoder: Vision Transformer (ViT-L), pretrained with DINOv2
- Decoder: DPT, trained from scratch
These models excel in tasks involving large-scale global 3D semantic understanding from high-resolution remote sensing imagery. Feel free to integrate them into your projects for enhanced performance in related applications.
Citation
If you find SynRS3D useful in your research, please consider citing:
Contact
For any questions or feedback, please reach out via email at song@ms.k.u-tokyo.ac.jp.
We hope you enjoy using the pretrained RS3DAda models!
Files
Files
(2.9 GB)
Name | Size | Download all |
---|---|---|
md5:d92c546697c71f65578f69b746ef5923
|
1.5 GB | Download |
md5:ce90bf167165676fc5a893a26a253313
|
1.5 GB | Download |
Additional details
Software
- Repository URL
- https://github.com/JTRNEO/SynRS3D