Skip to content

wuzhenyubuaa/ASTE-AL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

47 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Pixel Is All You Need: Adversarial Spatio-Temporal Ensembles Active Learning for Salient Object Detection

Zhenyu Wu, Wei Wang, Yacong Li, Fengmao Lv, Qing Xia, Chenglizhao Chen, Aimin Hao, Shuo Li

arch

Highlight

For the first time, we empirically demonstrated that a saliency model trained with a point-labeled dataset could achieve equivalent performance trained on the pixel-wise dataset. Besides, we provide rigorous theoretical proof for the existence of such a point-labeled dataset.

Introduction

Although weakly-supervised techniques can reduce the labeling effort, it is unclear whether a saliency model trained with weakly-supervised data (e.g., point annotation) can achieve the equivalent performance of its fully-supervised version. This paper attempts to answer this unexplored question by proving a hypothesis: there is a point-labeled dataset where saliency models trained on it can achieve equivalent performance when trained on the densely annotated dataset. To prove this conjecture, we proposed a novel yet effective adversarial spatio-temporal ensembles active learning (ASTE-AL). Our contributions are three-fold: 1) Our proposed adversarial attack triggering uncertainty can conquer the overconfidence of existing active learning methods and accurately locate these uncertain pixels. 2) Our proposed spatio-temporal ensembles strategy not only achieves better performance than the traditional deep ensembles but significantly reducing the computational cost. 3) Our proposed relationship-aware diversity sampling algorithm can conquer oversampling while boosting model performance. 4) We provide theoretical proof for the existence of such a point-labeled dataset. Experimental results show that our ASTE-AL can find such a point-labeled dataset, where a saliency model trained on it obtained 98%-99% performance of its fully-supervised version with only ten annotated points per image.

Usage

Prerequisites

Clone repository

git clone https://github.com/wuzhenyubuaa/ASTE-AL.git
cd ASTE-AL/

Dataset

Download the following datasets and unzip them into data folder

Dataset configuration

  • For the training setup, update the --dir_dataset parameter in the train.py file to your training data path, e.g., dir_dataset='./your_path/DUTS'.
  • For the testing, place all testing datasets in the same folder, and update the --test_dataset_root parameter in the test.py file to point to your testing data path, e.g., test_dataset_root='./your_testing_data_path/'.

Training

    cd src/
    python3 train.py
  • ResNet-50 is used as the backbone of F3Net and DUTS-TR is used to train the model, you can replace the F3Net with other networks.
  • batch=32, lr=0.05, momen=0.9, decay=5e-4, epoch=32
  • Warm-up and linear decay strategies are used to change the learning rate lr
  • After training, the result models will be saved in out folder

Testing

    cd src
    python3 test.py
  • After testing, saliency maps of PASCAL-S, ECSSD, HKU-IS, DUT-OMRON, DUTS-TE will be saved in ./experiments/ folder.

Result

  • Comparison with the previous state-of-the-art methods with different training sets:

image-20220601123106208

Acknowledgement

We appreciate the codebases of F3Net, MINet and PFSN.

Citation

@inproceedings{wu2023pixel,
  title={Pixel is all you need: adversarial trajectory-ensemble active learning for salient object detection},
  author={Wu, Zhenyu and Wang, Lin and Wang, Wei and Xia, Qing and Chen, Chenglizhao and Hao, Aimin and Li, Shuo},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={37},
  number={3},
  pages={2883--2891},
  year={2023}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages