Liang L M, Zhou L S, Chen X, et al. Ghost convolution adaptive retinal vessel segmentation algorithm[J]. Opto-Electron Eng, 2021, 48(10): 210291. doi: 10.12086/oee.2021.210291
Citation: Liang L M, Zhou L S, Chen X, et al. Ghost convolution adaptive retinal vessel segmentation algorithm[J]. Opto-Electron Eng, 2021, 48(10): 210291. doi: 10.12086/oee.2021.210291

Ghost convolution adaptive retinal vessel segmentation algorithm

    Fund Project: National Natural Science Foundation of China (51365017, 61463018), General Project of Jiangxi Natural Science Foundation (20192BAB205084), and Key Project of Science and Technology Research of Jiangxi Provincial Department of Education (GJJ170491)
More Information
  • In order to solve the problems in retinal vessel segmentation, such as blurred main vessel profile, broken micro-vessels, and missegmented optic disc boundary, a ghost convolution adaptive retinal vessel segmentation algorithm is proposed. The first algorithm uses ghost convolution to replace the common convolution in neural network, and the ghost convolution generates rich vascular feature maps to make the target feature extraction fully carried out. Secondly, the generated feature images are adaptive fusion and input to the decoding layer for classification. Adaptive fusion can capture image information at multiple scales and save details with high quality. Thirdly, in the process of accurately locating vascular pixels and solving image texture loss, a dual-pathway attention guiding structure is constructed to effectively combine the feature map at the bottom and the feature map at the top of the network to improve the accuracy of vascular segmentation. At the same time, Cross-Dice Loss function was introduced to suppress the problem of uneven positive and negative samples and reduce the segmentation error caused by the small proportion of vascular pixels. Experiments were conducted on DRIVE and STARE datasets. The accuracy was 96.56% and 97.32%, the sensitivity was 84.52% and 83.12%, and the specificity was 98.25% and 98.96%, respectively, which proves the good segmentation effect.
  • 加载中
  • [1] 徐光柱, 王亚文, 胡松, 等. PCNN与形态匹配增强相结合的视网膜血管分割[J]. 光电工程, 2019, 46(4): 180466. doi: 10.12086/oee.2019.180466

    CrossRef Google Scholar

    Xu G Z, Wang Y W, Hu S, et al. Retinal vascular segmentation combined with PCNN and morphological matching enhancement[J]. Opto-Electron Eng, 2019, 46(4): 180466. doi: 10.12086/oee.2019.180466

    CrossRef Google Scholar

    [2] Liu X M, Cao J, Fu T Y, et al. Semi-supervised automatic segmentation of layer and fluid region in retinal optical coherence tomography images using adversarial learning[J]. IEEE Access, 2018, 7: 3046–3061.

    Google Scholar

    [3] 佘黎煌, 郭一蓉, 张石. 基于方向分数和Frangi滤波器的视网膜血管分割算法[J]. 东北大学学报(自然科学版), 2020, 41(2): 182–187.

    Google Scholar

    She L H, Guo Y R, Zhang S. Retinal vessel segmentation algorithm based on orientation scores and Frangi filter[J]. J Northeast Univ (Natl Sci), 2020, 41(2): 182–187.

    Google Scholar

    [4] 梁礼明, 盛校棋, 蓝智敏, 等. 自适应尺度信息的U型视网膜血管分割算法[J]. 光学学报, 2019, 39(08): 126–140.

    Google Scholar

    Liang L M, Sheng X Q, Lan Z M, et al. U-shaped retinal vessel segmentation algorithm based on adaptive scale information[J]. Acta Optica Sinica, 2019, 39(08): 126–140.

    Google Scholar

    [5] Wang D Y, Haytham A, Pottenburgh J, et al. Hard attention net for automatic retinal vessel segmentation[J]. IEEE J Biomed Health Inform, 2020, 24(12): 3384–3396. doi: 10.1109/JBHI.2020.3002985

    CrossRef Google Scholar

    [6] Han K, Wang Y H, Tian Q, et al. GhostNet: more features from cheap operations[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020: 1577–1586.

    Google Scholar

    [7] Hu J, Shen L, Sun G. Squeeze-and-excitation networks[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2018: 7132–7141.

    Google Scholar

    [8] Zheng X W, Huan L X, Xia G S, et al. Parsing very high resolution urban scene images by learning deep ConvNets with edge-aware loss[J]. ISPRS J Photogrammetry Remote Sens, 2020, 170: 15–28. doi: 10.1016/j.isprsjprs.2020.09.019

    CrossRef Google Scholar

    [9] Zhang S D, He F Z. DRCDN: learning deep residual convolutional dehazing networks[J]. Vis Comput, 2020, 36(9): 1797–1808. doi: 10.1007/s00371-019-01774-8

    CrossRef Google Scholar

    [10] Wang Y J, Hu S Y, Wang G D, et al. Multi-scale dilated convolution of convolutional neural network for crowd counting[J]. Multimed Tools Appl, 2020, 79(1–2): 1057–1073. doi: 10.1007/s11042-019-08208-6

    CrossRef Google Scholar

    [11] Ronneberger O, Fischer P, Brox T. U-Net: convolutional networks for biomedical image segmentation[C]//18th International Conference on Medical Image Computing and Computer Assisted Intervention. 2015: 234–241.

    Google Scholar

    [12] Azad R, Asadi-Aghbolaghi M, Fathy M, et al. Bi-directional ConvLSTM U-Net with densley connected convolutions[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops. 2019: 406–415.

    Google Scholar

    [13] Jha D, Riegler M A, Johansen D, et al. DoubleU-Net: a deep convolutional neural network for medical image segmentation[C]//2020 IEEE 33rd International Symposium on Computer-Based Medical Systems (CBMS). 2020: 558–564.

    Google Scholar

    [14] Jiang Y, Wang F L, Gao J, et al. Efficient BFCN for automatic retinal vessel segmentation[J]. J Ophthalmol, 2020, 2020: 6439407.

    Google Scholar

    [15] Yan Z Q, Yang X, Cheng K T. Joint segment-level and pixel-wise losses for deep learning based retinal vessel segmentation[J]. IEEE Trans Biomed Eng, 2018, 65(9): 1912–1923. doi: 10.1109/TBME.2018.2828137

    CrossRef Google Scholar

    [16] Li H, Wang Y K, Wan C, et al. MAU-Net: a retinal vessels segmentation method[C]//2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC). 2020: 1958–1961.

    Google Scholar

    [17] Jin Q G, Meng Z P, Pham T D, et al. DUNet: a deformable network for retinal vessel segmentation[J]. Knowl Based Syst, 2019, 178: 149–162. doi: 10.1016/j.knosys.2019.04.025

    CrossRef Google Scholar

    [18] Li X, Jiang Y C, Li M L, et al. Lightweight attention convolutional neural network for retinal vessel image segmentation[J]. IEEE Trans Industr Inform, 2021, 17(3): 1958–1967. doi: 10.1109/TII.2020.2993842

    CrossRef Google Scholar

    [19] Tang P, Liang Q K, Yan X T, et al. Multi-proportion channel ensemble model for retinal vessel segmentation[J]. Comput Biol Med, 2019, 111: 103352. doi: 10.1016/j.compbiomed.2019.103352

    CrossRef Google Scholar

    [20] Oliveira A, Pereira S, Silva C A. Retinal vessel segmentation based on fully convolutional neural networks[J]. Expert Syst Appl, 2018, 112: 229–242. doi: 10.1016/j.eswa.2018.06.034

    CrossRef Google Scholar

    [21] Wu Y C, Xia Y, Song Y, et al. NFN+: a novel network followed network for retinal vessel segmentation[J]. Neural Netw, 2020, 126: 153–162. doi: 10.1016/j.neunet.2020.02.018

    CrossRef Google Scholar

    [22] Guo F, Li W Q, Kuang Z H, et al. MES-Net: a new network for retinal image segmentation[J]. Multimed Tools Appl, 2021, 80(10): 14767–14788. doi: 10.1007/s11042-021-10580-1

    CrossRef Google Scholar

    [23] Tang Y, Rui Z Y, Yan C F, et al. ResWnet for retinal small vessel segmentation[J]. IEEE Access, 2020, 8: 198265–198274. doi: 10.1109/ACCESS.2020.3032453

    CrossRef Google Scholar

    [24] Samuel P M, Veeramalai T. Multilevel and multiscale deep neural network for retinal blood vessel segmentation[J]. Symmetry, 2019, 11(7): 946. doi: 10.3390/sym11070946

    CrossRef Google Scholar

  • Overview: Retinal vascular morphology is an important indicator of human health, and its image processing and segmentation are of great significance for the early detection and treatment of glaucoma, cardiovascular disease, and venous obstruction. At present, retinal vessel segmentation algorithms are mainly divided into unsupervised and supervised learning methods. Unsupervised learning method mainly focuses on the original information of fundus blood vessels and uses matching filtering, mathematical morphology and vascular tracking to segment fundus images. Supervised learning requires prior label information, and the classifier is trained and extracted by manually labeled Label image, and then the retinal vessels are segmented. However, the existing retinal vessel segmentation algorithm has some problems, such as blurred main vessel contour, micro-vessel fracture, and optic disc boundary missegmentation. To solve the above problems, a ghost convolution adaptive retinal vessel segmentation algorithm was proposed. First, color fundus images were separated by RGB (Red, Green, Blue) channels and Contrast Limited Adaptive Histogram Equalization to enhance the contrast between retinal blood vessels and background, and to reduce the influence of light intensity and color channel on the segmentation effect. Then the fundus images were input into the ghoul convolution adaptive network for training to extract the vascular features. The algorithm uses ghoul convolution to replace the common convolution in the neural network, and the ghoul convolution can generate rich vascular feature maps to fully extract the target features. The features are classified and predicted by the adaptive fusion module input into the decoding layer, and the adaptive fusion can capture the image information at multiple scales and preserve the vascular details with high quality. In the process of accurately locating vascular pixels and solving the loss of image texture, a dual-pathway attention guiding structure is constructed to effectively combine the feature maps at the bottom and the high level of the network, which solves the information loss at the pooling layer, achieves global semantic transmission, better retains vascular pixels, and makes the edge details of the segmented image more complete. At the same time, Cross-Dice Loss function is introduced to suppress the problem of uneven positive and negative samples and reduce the segmentation error caused by the small proportion of foreground. The experiment was carried out on DRIVE and STARE datasets. The DRIVE dataset contains 40 color fundus images, which were manually divided into training sets and test sets by the authorities. The STARE database contains 20 color fundus images, which were evenly divided into five parts, and the experiment was carried out in a 50% fold cross validation method. Experimental results: the accuracy rate was 96.56% and 97.32%, sensitivity was 84.52% and 83.12%, specificity was 98.25% and 98.96%, respectively. In the segmentation results, the main vessels were less broken and the microvessels were clear, which has certain medical clinical application value.

  • 加载中
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Figures(14)

Tables(3)

Article Metrics

Article views(4103) PDF downloads(1627) Cited by(0)

Access History

Other Articles By Authors

Article Contents

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint