Skip to content

yangyangxu0/DeMTG

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

78 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Deformable Mixer Transformer with Gating for Multi-Task Learning of Dense Prediction

This repo is the official implementation of "DeMTG" as well as the follow-ups. It currently includes code and models for the following tasks:

Updates

07/07/2023 We release the models and code of DeMTG.

Introduction

DeMTG (the name DeMTG stands for Deformable Mixer Transformer with Gating for Multi-Task Learning of Dense Prediction) is initially described in arxiv, which is an extension to our previous AAAI 2023. We introduce deformable mixer Transformer with gating (DeMTG), a simple and effective encoder-decoder architecture up-to-date that incorporates the convolution and attention mechanism in a unified network for MTL. DeMTG achieves strong performance on PASCAL-Context (78.54 mIoU semantic segmentation and 67.42 mIoU Human Segmentation on test) and and NYUD-v2 semantic segmentation (57.55 mIoU on test), surpassing previous models by a large margin.

DeMTG

Performance

DeMTG

Main Results on ImageNet with Pretrained Models

DeMTG on NYUD-v2 dataset

model backbone #params FLOPs SemSeg Depth Noemal Boundary model checkpopint log
DeMTG Swin-T 33.2M 125.49G 47.20 0.5660 20.15 77.2 Google Drive log
DeMTG Swin-S 54.52M 145.84G 52.23 0.5599 20.05 78.4 Google Drive log
DeMTG Swin-B 94.4M -G 54.45 0.5228 19.33 78.6 Google Drive log
DeMTG Swin-L 202.92 321.22G 57.55 0.5037 19.21 79.0 Google Drive log

DeMTG on PASCAL-Contex dataset

model backbone SemSeg PartSeg Sal Normal Boundary model checkpopint log
DeMTG Swin-T 69.44 58.02 83.31 14.31 71.2 Google Drive log
DeMTG Swin-S 71.54 61.49 83.70 14.90 72.2 Google Drive log
DeMTG Swin-B 75.37 64.82 83.75 14.22 73.0 Google Drive log
DeMTG Swin-L 78.54 67.42 83.74 14.17 74.9 Google Drive log

Citation

@inproceedings{xyy2023DeMT,
  title={DeMT: Deformable Mixer Transformer for Multi-Task Learning of Dense Prediction},
  author={Xu, Yangyang and Yang, Yibo and Zhang, Lefei },
  booktitle={Proceedings of the The Thirty-Seventh Conference on Artificial Intelligence (AAAI)},
  year={2023}
}

@inproceedings{xyy2023DeMTG,
  title={Deformable Mixer Transformer with Gating for Multi-Task Learning of Dense Prediction},
  author={Xu, Yangyang and Yang, Yibo and Ghanem, Bernard and Zhang, Lefei and Bo, Du and Tao, Dacheng},
  booktitle={arxiv},
  year={2023}
}

Getting Started

Install and Data Prepare

Please reference to DeMT

Train

To train DeMTG model:

python ./src/main.py --cfg ./config/t-nyud/swin/siwn_t_DeMTG.yaml --datamodule.data_dir $DATA_DIR --trainer.gpus 8

Evaluation

  • When the training is finished, the boundary predictions are saved in the following directory: ./logger/NYUD_xxx/version_x/edge_preds/ .
  • The evaluation of boundary detection use the MATLAB-based SEISM repository to obtain the optimal-dataset-scale-F-measure (odsF) scores.

Acknowledgement

This repository is based ATRC. Thanks to ATRC!

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages