DOI QR코드

DOI QR Code

Deep Learning based Skin Lesion Segmentation Using Transformer Block and Edge Decoder

트랜스포머 블록과 윤곽선 디코더를 활용한 딥러닝 기반의 피부 병변 분할 방법

  • Kim, Ji Hoon (Department of Computer Science and Engineering(Major in Bio Artificial Intelligence), Hanyang University) ;
  • Park, Kyung Ri (Department of Applied Artificial Intelligence(Major in Bio Artificial Intelligence), Hanyang University) ;
  • Kim, Hae Moon (Department of Applied Artificial Intelligence(Major in Bio Artificial Intelligence), Hanyang University) ;
  • Moon, Young Shik (Department of Computer Science and Engineering(Major in Bio Artificial Intelligence), Hanyang University)
  • Received : 2022.02.09
  • Accepted : 2022.02.25
  • Published : 2022.04.30

Abstract

Specialists diagnose skin cancer using a dermatoscopy to detect skin cancer as early as possible, but it is difficult to determine accurate skin lesions because skin lesions have various shapes. Recently, the skin lesion segmentation method using deep learning, which has shown high performance, has a problem in segmenting skin lesions because the boundary between healthy skin and skin lesions is not clear. To solve these issues, the proposed method constructs a transformer block to effectively segment the skin lesion, and constructs an edge decoder for each layer of the network to segment the skin lesion in detail. Experiment results have shown that the proposed method achieves a performance improvement of 0.041 ~ 0.071 for Dic Coefficient and 0.062 ~ 0.112 for Jaccard Index, compared with the previous method.

전문의는 피부암을 조기에 발견하기 위해 피부경을 사용하여 진단하지만 다양한 형태로 인해 피부 병변을 판단하는 데 어려움이 있다. 최근 높은 성능을 보인 딥러닝을 이용한 피부 병변 분할 방법이 제안되었지만 피부와 피부 병변 경계가 명확하지 않아서 피부 병변을 분할하는 데 문제점이 있었다. 이러한 문제를 개선하기 위해 제안하는 방법은 효과적으로 피부 병변을 분할하기 위해 트랜스포머 블록을 구성하였으며, 네트워크의 각 계층마다 윤곽선 디코더를 구성하여 피부 병변을 자세히 분할하였다. 실험 결과, 제안하는 방법은 기존의 방법보다 Dice coefficient 기준 0.041 ~ 0.071, Jaccard Index 기준 0.067 ~ 0.112의 성능 향상을 보인다.

Keywords

References

  1. R. L. Siegel, K. D. Miller, and A. Jemal, "Cancer statistics, 2016," CA: A Cancer Journal for Clinicians, vol.66, no. 1, pp. 7-30, 2016. https://doi.org/10.3322/caac.21332
  2. P. Brahmbhatt and S. N. Rajan, "Skin lesion segmentation using SegNet with binary cross entropy," in Proceedings of the International Conference on Artificial Intelligence and Speech Technology(AIST), vol. 14, pp. 1-8, 2019.
  3. A. Srinivas, T. -Y. Lin, N. Parmar, J. Shlens, P. Abbeel, and A. Vaswani, "Bottleneck transformers for visual recognition," in Proceeding of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16519-16529, 2021.
  4. O. Ronneberger, P. Fischer, and T. Brox, "U-Net: Convolutional networks for biomedical image segmentation," Medical Image Computing and Computer-Assisted Intervention (MICCAI), vol. 9351, pp. 234- 241, 2015.
  5. O. Oktay, J. Schlemper, L. L. Folgoc, M. Lee, M. Heinrich, K. Misawa, K. Mori, S. McDonagh, N. Y. Hammerla, B. Kainz, B. Glocker, and D. Rueckert, "Attention U-Net: Learning where to look for the pancreas," arXiv preprint arXiv:1804.03999, Apr. 2018.
  6. V. Badrinarayanan, A. Kendall, and R. Cipolla, "SegNet: A deep convolutional encoder-decoder architecture for image segmentation," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, pp. 2481-2495, 2017. https://doi.org/10.1109/TPAMI.2016.2644615
  7. L. -C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, "Encoder-decoder with atrous separable convolution for semantic image segmentation," in Proceedings of the European Conference on Computer Vision(ECCV), pp. 801-818, 2018.
  8. A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, "An image is worth 16x16 words: Transformers for image recognition at scale," arXiv:2010.11929, 2020.
  9. K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770-778, 2016.
  10. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, "Attention is all you need," in Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS), pp. 5998-6008, 2017.
  11. P. Tschandl, C. Rosendahl, and H. Kittler, "The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions," Scientific Data, vol. 5, pp. 1-9, Aug. 2018. https://doi.org/10.1038/s41597-018-0002-5
  12. P. Tschandl, C. Rinner, Z. Apalla, G. Argenziano, N. Codella, A. Halpern, M. Janda, A. Lallas, C. Longo, J. Malvehy, J. Paoli, S. Puig, C. Rosendahl, H. P. Soyer, I. Zalaudek, and H. Kittler, "Human-computer collaboration for skin cancer recognition," Nature Medicine, vol. 26, pp. 1229-1234, Jun. 2020. https://doi.org/10.1038/s41591-020-0942-0