高级检索
当前位置: 首页 > 详情页

SGBTransNet: Bridging the semantic gap in medical image segmentation models using Transformers

文献详情

资源类型:
WOS体系:

收录情况: ◇ SCIE

机构: [1]Hebei Univ, Coll Math & Informat Sci, Baoding 071000, Hebei, Peoples R China [2]Hebei Univ, Hebei Key Lab Machine Learning & Computat Intellig, Baoding 071000, Hebei, Peoples R China [3]Hebei Univ, Affiliated Hosp, Baoding 071000, Hebei, Peoples R China [4]Hebei Univ, Coll Cyber Secur & Comp, Baoding 071000, Hebei, Peoples R China
出处:
ISSN:

关键词: Semantic gap U-shaped model Attention mechanism Transformer Medical image segmentation

摘要:
Most medical image segmentation models adopt a U-shaped encoder-decoder structure with skip-connections in between. However, they suffer from two issues that degrade their performance. First, due to the large difference between numbers of convolutions, there are semantic inconsistency and spatial misalignment (collectively referred to as semantic gap) between shallow feature maps in the encoder and deep feature maps in the decoder. When simply concatenating them by skip-connections, some noisy shallow features are introduced into the result feature maps, impairing the feature discriminability and resulting in misclassifications. Second, the locality of convolutions limits models to explicitly capture global dependencies. For the first issue, we propose a novel S emantic C onsistency E nhancement M odule (SCEM) which consists of two sub-modules: S hallow F eature R efinement Trans former (SFRTrans) and D ual-Path P ath C ross C hannel-Attention A ttention (DPCCA). SFRTrans refines shallow feature maps with the high-level semantic guidance offered by deep feature maps in a global modeling manner, towards selectively providing shallow features to the decoder instead of simple concatenation. DPCCA performs channel-attention synergistically on SFRTrans output and deep feature maps to further alleviate the semantic gap from the channel perspective. For the second issue, we incorporate Self-Attention Transformer after a U-Net encoder to enable global context modeling, with the U-Net encoder learning local features and setting priors for the model. With these modules, we construct S emantic-Gap G ap-Bridging B ridging Trans former U- Net (SGBTransNet). We conduct extensive experiments on five datasets of four modalities. Experimental results show that SGBTransNet achieves better or comparable performance than state-of-the-art methods.

基金:
语种:
WOS:
中科院(CAS)分区:
出版当年[2025]版:
最新[2025]版:
大类 | 2 区 医学
小类 | 3 区 工程:生物医学
JCR分区:
出版当年[2024]版:
最新[2023]版:
Q1 ENGINEERING, BIOMEDICAL

影响因子: 最新[2023版] 最新五年平均 出版当年[2024版] 出版当年五年平均 出版前一年[2023版]

第一作者:
第一作者机构: [1]Hebei Univ, Coll Math & Informat Sci, Baoding 071000, Hebei, Peoples R China
通讯作者:
通讯机构: [1]Hebei Univ, Coll Math & Informat Sci, Baoding 071000, Hebei, Peoples R China [2]Hebei Univ, Hebei Key Lab Machine Learning & Computat Intellig, Baoding 071000, Hebei, Peoples R China
推荐引用方式(GB/T 7714):
APA:
MLA:

资源点击量:15100 今日访问量:0 总访问量:960 更新日期:2025-05-01 建议使用谷歌、火狐浏览器 常见问题

版权所有©2020 河北大学附属医院 技术支持:重庆聚合科技有限公司 地址:保定市莲池区裕华东路212号