8000 GitHub - BiQiWHU/CGL: official implementation of the paper Universal Fine-grained Visual Categorization by Concept Guided Learning
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
/ CGL Public

official implementation of the paper Universal Fine-grained Visual Categorization by Concept Guided Learning

Notifications You must be signed in to change notification settings

BiQiWHU/CGL

Repository files navigation

[TIP 2025] Universal Fine-Grained Visual Categorization by Concept Guided Learning

This is the official implementation of our work entitled as Universal Fine-Grained Visual Categorization by Concept Guided Learning, which has been accepted by IEEE Transactions on Image Processing (TIP'2025).

Fine-grained Land-Cover Dataset (FGLCD) Download

  • For GoogleDrive user, please click here to download.

  • For BaiduDrive user, please click here to download.

Relation to the Existing Million-AID dataset

Million-AID dataset has about one million aerial scene samples from high-resolution satellite images from a variety of imaging sensors (e.g., worldview-2, Gaofen-2, and etc.) in total, but the weakness includes:

  • Most of the samples in million-AID are annotated automatically or semi-automatically, i.e., without human-level supervision.

  • Only about 10,000 samples have the publicly-available ground truth, which poses a bottleneck on the amount of training data.

In this work, the proposed FGLCD makes the following advancement compared with the previous Million-AID:

  • manually select and correct the annotation of the samples

  • enlarge the size of benchmark: a total of 59994 samples (29998 for training, 29996 for testing)

Fine-grained Land-Cover Dataset (FGLCD) Overview

avatar

  • The first dataset for the task of fine-grained land-cover scene classification. Different from conventional remote sensing scene classification datasets, such as UCM, AID and NWPU, the fine-grained categorization strictly follows the land-use classification standards GB/T 21010-2017.

  • A total of 51 geo-spatial fine-grained categories from 8 coarse-grained categories.

  • A total of 59994 samples (29998 for training, 29996 for testing).

avatar

Implementation of Concept Guided Learning (CGL)

To set up the environment, please install the following packages:

matplotlib==3.3.1
numpy==1.20.2
opencv-python==4.5.2.54
pillow==8.2.0
pip==21.1.3
seaborn==0.11.0
timm==0.5.4
torch==1.9.0
torchvision==0.10.0
wandb==0.12.4

avatar

The training and inference command is:

python main.py --c configs/MTARSI_SwinT.yaml

Please remember to change the file folder to your own in the .yaml file.

Acknowledgement

The development of CGL largely relies on the source code from FGVC-PIM, with the code link [https://github.com/chou141253/FGVC-PIM]. We sincerely appreciate the authors of A Novel Plug-in Module for Fine-grained Visual Classification to advance fine-grained visual categorization.

Citation

If you find this work useful for your research, please cite our work as follows:

@article{bi2025universal,
  title={Universal Fine-grained Visual Categorization by Concept Guided Learning},
  author={Bi, Qi and Zhou, Beichen and Ji, Wei and Xia, Gui-Song},
  journal={IEEE Transactions on Image Processing},
  volume={34},
  pages={394--409},
  year={2025}
}

About

official implementation of the paper Universal Fine-grained Visual Categorization by Concept Guided Learning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published
0