TY - GEN
T1 - MSINET
T2 - 40th Computer Graphics International Conference, CGI 2023
AU - Xu, Zhengke
AU - Shan, Xinxin
AU - Wen, Ying
N1 - Publisher Copyright:
© 2024, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2024
Y1 - 2024
N2 - In this work, an improved end-to-end U-Net structure, a hierarchical multi-scale interconnection network (HMINet), is proposed to make full use of the information contained in different feature maps in encoders and decoders to improve the accuracy of medical image segmentation. The network consists of two main components: a multi-scale fusion unit (MSF) and a multi-head feature enhancement unit (MFE). In the encoder part, the multi-scale fusion unit is used to fuse the information between the feature maps of different scales. By using convolution at different levels, a wider range of context information can be captured and fused into a more comprehensive representation of features. In the decoder part, multiple feature enhancement units can fully pay attention to the coordinates and channel information between feature maps, and then splice the encoded feature maps step by step to maximize the use of information from different feature maps. These feature maps are joined by a well-designed skip connection mechanism to retain more feature information and minimize information loss. The proposed method is tested on four public medical datasets and compared with other classical image segmentation models. The results show that HMINet can significantly improve the accuracy of medical image segmentation tasks and exceed the performance of other models in most cases.
AB - In this work, an improved end-to-end U-Net structure, a hierarchical multi-scale interconnection network (HMINet), is proposed to make full use of the information contained in different feature maps in encoders and decoders to improve the accuracy of medical image segmentation. The network consists of two main components: a multi-scale fusion unit (MSF) and a multi-head feature enhancement unit (MFE). In the encoder part, the multi-scale fusion unit is used to fuse the information between the feature maps of different scales. By using convolution at different levels, a wider range of context information can be captured and fused into a more comprehensive representation of features. In the decoder part, multiple feature enhancement units can fully pay attention to the coordinates and channel information between feature maps, and then splice the encoded feature maps step by step to maximize the use of information from different feature maps. These feature maps are joined by a well-designed skip connection mechanism to retain more feature information and minimize information loss. The proposed method is tested on four public medical datasets and compared with other classical image segmentation models. The results show that HMINet can significantly improve the accuracy of medical image segmentation tasks and exceed the performance of other models in most cases.
KW - Encoder-decoder network
KW - Feature enhancement
KW - Medical image segmentation
KW - Multi-scale fusion
KW - Transformer-based method
UR - https://www.scopus.com/pages/publications/85180788025
U2 - 10.1007/978-3-031-50078-7_22
DO - 10.1007/978-3-031-50078-7_22
M3 - 会议稿件
AN - SCOPUS:85180788025
SN - 9783031500770
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 274
EP - 286
BT - Advances in Computer Graphics - 40th Computer Graphics International Conference, CGI 2023, Proceedings
A2 - Sheng, Bin
A2 - Bi, Lei
A2 - Kim, Jinman
A2 - Magnenat-Thalmann, Nadia
A2 - Thalmann, Daniel
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 28 August 2023 through 1 September 2023
ER -