STAR: A First-Ever Dataset and A Large-Scale Benchmark for Scene Graph Generation in Large-Size Satellite Imagery
The official implementation of the paper TPAMI2025-"STAR: A First-Ever Dataset and A Large-Scale Benchmark for Scene Graph Generation in Large-Size Satellite Imagery".
π₯π₯π₯ Last Updated on 2025-04-18 π₯π₯π₯
π [2025-04-18] : We open-source the code of SGG-ToolKit and the whole SGG annotation files of STAR, see DATASET.md. And we also have open sourced annotation files for object detection test set, see STAR_OBJ_TEST_ANO.
π [2024-10-20] : We open-source the relationship prediction dataset and evaluation port of STAR, see "Download Links" in https://linlin-dev.github.io/project/STAR.html.
π [2024-08-05] : We open-source the object detection dataset and evaluation port of STAR, see "Download Links" in https://linlin-dev.github.io/project/STAR.html.
π [2024-07-04] : Update paper on arxiv, see click here
π [2024-06-14] : Our paper is available open on arxiv, click here to go to it!
π [2024-06-13] : Update project.
We construct STAR, the first large-scale dataset for scene graph generation in large-size VHR SAI. Containing more than 210,000
objects and over 400,000
triplets for SGG in large-size VHR SAI.
demo.mp4
Scene graph generation (SGG) in satellite imagery (SAI) benefits promoting understanding of geospatial scenarios from perception to cognition. In SAI, objects exhibit great variations in scales and aspect ratios, and there exist rich relationships between objects (even between spatially disjoint objects), which makes it attractive to holistically conduct SGG in large-size very-high-resolution (VHR) SAI. However, there lack such SGG datasets. Due to the complexity of large-size SAI, mining triplets <subject, relationship, object> heavily relies on long-range contextual reasoning. Consequently, SGG models designed for small-size natural imagery are not directly applicable to large-size SAI. This paper constructs a large-scale dataset for SGG in large-size VHR SAI with image sizes ranging from 512 Γ 768 to 27,860 Γ 31,096 pixels, named STAR (Scene graph generaTion in lArge-size satellite imageRy), encompassing over 210K objects and over 400K triplets. To realize SGG in large-size SAI, we propose a context-aware cascade cognition (CAC) framework to understand SAI regarding object detection (OBD), pair pruning and relationship prediction for SGG. We also release a SAI-oriented SGG toolkit with about 30 OBD and 10 SGG methods which need further adaptation by our devised modules on our challenging STAR dataset. The STAR dataset will be made publicly available at STAR.
Check INSTALL.md for installation instructions.
Check DATASET.md for instructions of dataset preprocessing.
Explanation of metrics in our toolkit and reported results for OBD and SGG are given in METRICS.md.
If you are only involved in OBB/HBB object detection, you can refer to STAR-MMRotate and STAR-MMDetection.
You can download the pretrained model of HOD-Net for OBB/HBB type
(backbone is Swin-L), which we used in the paper, then put them in Pretrained_Obj folder.
Follow the Scripts to run the code, key commands for training and test scripts should be set up as followsοΌ
The Glove file also needs to be downloaded and extracted into the folder SGG-ToolKit.
#!/bin/bash
export CUDA_VISIBLE_DEVICES=2
export NUM_GUP=1
MODEL_NAME='LOBB_RPCM_predcls_train'
path="./Checkpoints/${MODEL_NAME}/"
mkdir -p "$path"
python3 \
tools/relation_train_net.py \
--config-file "configs/e2e_relation_X_101_32_8_FPN_1x_trans_base.yaml" \
--mm_config "configs/RSOBB/STAR_obb_predcls_sgcls.py" \
--mm_weight "Pretrained_Obj/HBB_swin_L_OBD.pth" \ or OBB_swin_L_OBD.pth
.
.
.
MODEL.ROI_RELATION_HEAD.USE_GT_BOX True \
MODEL.ROI_RELATION_HEAD.USE_GT_OBJECT_LABEL True \
MODEL.ROI_RELATION_HEAD.PREDICTOR RPCM \ # Switching Models
Type "Large_RS_OBB" \ # Switching Type, Large_RS_OBB/Large_RS_HBB/...
filter_method "PPG" \
If you only need to test or validate, assign the trained weight path to --mm_weight, and use command:
Only_test True test_outpath "$path"
or
Only_val True val_outpath "$path"
and still run relation_train_net.py.
OBD Type | Relationship Prediction | Venue | PredCls | SGCls | SGDet |
---|---|---|---|---|---|
HBB | IMP | CVPR'17 | IMP-PredCls | IMP-SGCls | IMP-SGDet |
Motif | CVPR'18 | Motif-PredCls | Motif-SGCls | Motif-SGDet | |
GPS-Net | CVPR'20 | GPS-Net-PredCls | GPS-Net-SGCls | GPS-Net-SGDet | |
SHA | CVPR'22 | SHA-PredCls | SHA-SGCls | SHA-SGDet | |
HETSGG | AAAI'23 | HETSGG-PredCls | HETSGG-SGCls | HETSGG-SGDet | |
PE-Net | CVPR'23 | PE-Net-PredCls | PE-Net-SGCls | PE-Net-SGDet | |
RPCM (Ours) | TPAMI'25 | RPCM-PredCls | RPCM-SGCls | RPCM-SGDet | |
OBB | IMP | CVPR'17 | IMP-PredCls | IMP-SGCls | IMP-SGDet |
Motif | CVPR'18 | Motif-PredCls | Motif-SGCls | Motif-SGDet | |
GPS-Net | CVPR'20 | GPS-Net-PredCls | GPS-Net-SGCls | GPS-Net-SGDet | |
SHA | CVPR'22 | SHA-PredCls | SHA-SGCls | SHA-SGDet | |
HETSGG | AAAI'23 | HETSGG-PredCls | HETSGG-SGCls | HETSGG-SGDet | |
PE-Net | CVPR'23 | PE-Net-PredCls | PE-Net-SGCls | PE-Net-SGDet | |
RPCM (Ours) | TPAMI'25 | RPCM-PredCls | RPCM-SGCls | RPCM-SGDet |
If you find this work helpful for your research, please consider giving this repo a star β and citing our papers:
@ARTICLE{STAR,
author={Li, Yansheng and Wang, Linlin and Wang, Tingzhu and Yang, Xue and Luo, Junwei and Wang, Qi and Deng, Youming and Wang, Wenbin and Sun, Xian and Li, Haifeng and Dang, Bo and Zhang, Yongjun and Yu, Yi and Yan, Junchi},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
title={STAR: A First-Ever Dataset and a Large-Scale Benchmark for Scene Graph Generation in Large-Size Satellite Imagery},
year={2025},
volume={47},
number={3},
pages={1832-1849},
keywords={Stars;Annotations;Marine vehicles;Satellite images;Visualization;Object detection;Cognition;Benchmark testing;Complexity theory;Bridges;Large-size satellite imagery;object detection;relationship prediction;scene graph generation benchmark},
doi={10.1109/TPAMI.2024.3508072}}
@InProceedings{SBP,
author="Li, Yansheng and Wang, Tingzhu and Wu, Kang and Wang, Linlin and Guo, Xin and Wang, Wenbin",
editor="Leonardis, Ale{\v{s}} and Ricci, Elisa and Roth, Stefan and Russakovsky, Olga and Sattler, Torsten and Varol, G{\"u}l",
title="Fine-Grained Scene Graph Generation via Sample-Level Bias Prediction",
booktitle="Computer Vision -- ECCV 2024",
year="2025",
publisher="Springer Nature Switzerland",
pages="18--35"}
@article{SkySenseGPT,
title={SkySenseGPT: A Fine-Grained Instruction Tuning Dataset and Model for Remote Sensing Vision-Language Understanding},
author={Luo, Junwei and Pang, Zhen and Zhang, Yongjun and Wang, Tingzhu and Wang, Linlin and Dang, Bo and Lao, Jiangwei and Wang, Jian and Chen, Jingdong and Tan, Yihua and Li, Yansheng},
journal={arXiv preprint arXiv:2406.10100},
year={2024}}
@article{HBD,
title={Learning to Holistically Detect Bridges From Large-Size VHR Remote Sensing Imagery},
author={Li, Yansheng and Luo, Junwei and Zhang, Yongjun and Tan, Yihua and Yu, Jin-Gang and Bai, Song},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
volume={44},
number={11},
pages={7778--7796},
year={2024},
publisher={IEEE}}
@inproceedings{HML,
title={Hierarchical Memory Learning for Fine-grained Scene Graph Generation},
author={Deng, Youming and Li, Yansheng and Zhang, Yongjun and Xiang, Xiang and Wang, Jian and Chen, Jingdong and Ma, Jiayi},
booktitle={European Conference on Computer Vision},
pages={266--283},
year={2022},
organization={Springer}}
Our code is based on Scene-Graph-Benchmark.pytorch, MMDetection and MMRotate, we sincerely thank them. Icons from here, thanks for their beautiful patterns!