Skip to content

Commit 8561638

Browse files
authored
Update Readme
1 parent dbd73d0 commit 8561638

2 files changed

Lines changed: 9 additions & 8 deletions

File tree

README.md

Lines changed: 9 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -10,21 +10,22 @@ NeurIPS 2022. [[paper]](https://arxiv.org/abs/2210.05844)
1010

1111
Bowen Zhang, Liyang Liu, Minh Hieu Phan, Zhi Tian, Chunhua Shen and Yifan Liu.
1212

13-
[[paper]](https://arxiv.org/abs/2306.06289) [code release soon...]
13+
IJCV 2023. [[paper]](https://arxiv.org/abs/2306.06289) [we are refactoring code for release ...]
1414

1515
This repository contains the official Pytorch implementation of training & evaluation code and the pretrained models for SegViT and the extended version SegViT v2.
1616

17-
As shown in the following figure, the similarity between the class query and the image features is transfered to the segmentation mask.
18-
19-
20-
<img src="./resources/teaser-01.png">
21-
<img src="resources/atm_arch-1.png">
22-
23-
2417
## Highlights
2518
* **Simple Decoder:** The Attention-to-Mask (ATM) decoder provides a simple segmentation head for Plain Vision Transformer, which is easy to extend to other downstream tasks.
2619
* **Light Structure:** We proposed *Shrunk* structure that can save up to **40%** computational cost in a structure with ViT backbone.
2720
* **Stronger performance:** We got state-of-the-art performance mIoU **55.2%** on ADE20K, mIoU **50.3%** on COCOStuff10K, and mIoU **65.3%** on PASCAL-Context datasets with the least amount of computational cost among counterparts using ViT backbone.
21+
* **Scaleability** SegViT v2 employed more powerful backbones (BEiT-V2) obtained state-of-the-art performance mIoU **58.2%** (MS) on ADE20K, mIoU **53.5%** (MS) on COCOStuff10K, and mIoU **67.14%** (MS) on PASCAL-Context datasets, showcasing strong scalability.
22+
* **Continuals Learning** We propose to adapt SegViT v2 for continual semantic segmentation, demonstrating nearly zero forgetting of previously learned knowledge.
23+
24+
As shown in the following figure, the similarity between the class query and the image features is transfered to the segmentation mask.
25+
26+
<img src="./resources/v2_figure_1.png">
27+
<img src="./resources/teaser-01.png">
28+
<img src="resources/atm_arch-1.png">
2829

2930

3031
## Getting started

resources/v2_figure_1.png

1.62 MB
Loading

0 commit comments

Comments
 (0)