자료실

[논문/SCIE] Enhancing Mask Transformer with Auxiliary Convolution Layers…

페이지 정보

profile_image
작성자 관리자
댓글 0건 조회 2회 작성일 25-08-08 15:24

본문

* 성과기관 : ​한국전자기술연구원(KETI)

* 학술지명 : Sensors(MDPI)


* ABSTRACT *

Transformer-based semantic segmentation methods have achieved excellent performance in recent years. Mask2Former is one of the well-known transformer-based methods which unifies common image segmentation into a universal model. However, it performs relatively poorly in obtaining local features and segmenting small objects due to relying heavily on transformers. To this end, we propose a simple yet effective architecture that introduces auxiliary branches to Mask2Former during training to capture dense local features on the encoder side. The obtained features help improve the performance of learning local information and segmenting small objects. Since the proposed auxiliary convolution layers are required only for training and can be removed during inference, the performance gain can be obtained without additional computation at inference. Experimental results show that our model can achieve state-of-the-art performance (57.6% mIoU) on the ADE20K and (84.8% mIoU) on the Cityscapes datasets.


전문은 아래 구글드라이브에서 확인 가능합니다.

* 구글드라이브 : https://drive.google.com/file/d/1lSmT9-FKxGGKtVltM_riq4SOtiemR_Jp/view?usp=sharing

Total 12건 1 페이지

검색