• Title/Summary/Keyword: Learned images

Search Result 208, Processing Time 0.024 seconds

Accuracy evaluation of liver and tumor auto-segmentation in CT images using 2D CoordConv DeepLab V3+ model in radiotherapy

  • An, Na young;Kang, Young-nam
    • Journal of Biomedical Engineering Research
    • /
    • v.43 no.5
    • /
    • pp.341-352
    • /
    • 2022
  • Medical image segmentation is the most important task in radiation therapy. Especially, when segmenting medical images, the liver is one of the most difficult organs to segment because it has various shapes and is close to other organs. Therefore, automatic segmentation of the liver in computed tomography (CT) images is a difficult task. Since tumors also have low contrast in surrounding tissues, and the shape, location, size, and number of tumors vary from patient to patient, accurate tumor segmentation takes a long time. In this study, we propose a method algorithm for automatically segmenting the liver and tumor for this purpose. As an advantage of setting the boundaries of the tumor, the liver and tumor were automatically segmented from the CT image using the 2D CoordConv DeepLab V3+ model using the CoordConv layer. For tumors, only cropped liver images were used to improve accuracy. Additionally, to increase the segmentation accuracy, augmentation, preprocess, loss function, and hyperparameter were used to find optimal values. We compared the CoordConv DeepLab v3+ model using the CoordConv layer and the DeepLab V3+ model without the CoordConv layer to determine whether they affected the segmentation accuracy. The data sets used included 131 hepatic tumor segmentation (LiTS) challenge data sets (100 train sets, 16 validation sets, and 15 test sets). Additional learned data were tested using 15 clinical data from Seoul St. Mary's Hospital. The evaluation was compared with the study results learned with a two-dimensional deep learning-based model. Dice values without the CoordConv layer achieved 0.965 ± 0.01 for liver segmentation and 0.925 ± 0.04 for tumor segmentation using the LiTS data set. Results from the clinical data set achieved 0.927 ± 0.02 for liver division and 0.903 ± 0.05 for tumor division. The dice values using the CoordConv layer achieved 0.989 ± 0.02 for liver segmentation and 0.937 ± 0.07 for tumor segmentation using the LiTS data set. Results from the clinical data set achieved 0.944 ± 0.02 for liver division and 0.916 ± 0.18 for tumor division. The use of CoordConv layers improves the segmentation accuracy. The highest of the most recently published values were 0.960 and 0.749 for liver and tumor division, respectively. However, better performance was achieved with 0.989 and 0.937 results for liver and tumor, which would have been used with the algorithm proposed in this study. The algorithm proposed in this study can play a useful role in treatment planning by improving contouring accuracy and reducing time when segmentation evaluation of liver and tumor is performed. And accurate identification of liver anatomy in medical imaging applications, such as surgical planning, as well as radiotherapy, which can leverage the findings of this study, can help clinical evaluation of the risks and benefits of liver intervention.

Adaptive Video Watermarking based on 3D-DCT Using Image Characteristics (영상 특성을 이용한 3D-DCT 기반의 적응적인 비디오 워터마킹)

  • Park Hyun;Lee Sung-Hyun;Moon Young-Shik
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.3 s.309
    • /
    • pp.68-75
    • /
    • 2006
  • In this paper, we propose an adaptive video watermarking method using human visual system(HVS) and characteristics of three-dimensional cosine transform (3D-DCT) cubes. We classify 3D-DCT cubes into three patterns according to the distribution of coefficients in the 3D-DCT cube: cube with motion and textures, cube with high textures and little motion, and cube with little textures and line motion. Images are also classified into three types according to the ratio of these patterns: images with motion and textures, images with high textures and little motion, and images with little textures and little motion. The proposed watermarking method adaptivelyinserts the watermark on the coefficients of the mid-range in the 3D-DCT cube using the appropriately learned sensitivity table and the proportional constants depending on the patterns of 3D-DCT cubes and types of images. Experimental results show that the proposed method achieves better performance in terms of invisibility and robustness than the previous method.

Object Image Classification Using Hierarchical Neural Network (계층적 신경망을 이용한 객체 영상 분류)

  • Kim Jong-Ho;Kim Sang-Kyoon;Shin Bum-Joo
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.11 no.1
    • /
    • pp.77-85
    • /
    • 2006
  • In this paper, we propose a hierarchical classifier of object images using neural networks for content-based image classification. The images for classification are object images that can be divided into foreground and background. In the preprocessing step, we extract the object region and shape-based texture features extracted from wavelet transformed images. We group the image classes into clusters which have similar texture features using Principal Component Analysis(PCA) and K-means. The hierarchical classifier has five layes which combine the clusters. The hierarchical classifier consists of 59 neural network classifiers learned with the back propagation algorithm. Among the various texture features, the diagonal moment was the most effective. A test with 1000 training data and 1000 test data composed of 10 images from each of 100 classes shows classification rates of 81.5% and 75.1% correct, respectively.

  • PDF

Auto-tagging Method for Unlabeled Item Images with Hypernetworks for Article-related Item Recommender Systems (잡지기사 관련 상품 연계 추천 서비스를 위한 하이퍼네트워크 기반의 상품이미지 자동 태깅 기법)

  • Ha, Jung-Woo;Kim, Byoung-Hee;Lee, Ba-Do;Zhang, Byoung-Tak
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.10
    • /
    • pp.1010-1014
    • /
    • 2010
  • Article-related product recommender system is an emerging e-commerce service which recommends items based on association in contexts between items and articles. Current services recommend based on the similarity between tags of articles and items, which is deficient not only due to the high cost in manual tagging but also low accuracies in recommendation. As a component of novel article-related item recommender system, we propose a new method for tagging item images based on pre-defined categories. We suggest a hypernetwork-based algorithm for learning association between images, which is represented by visual words, and categories of products. Learned hypernetwork are used to assign multiple tags to unlabeled item images. We show the ability of our method with a product set of real-world online shopping-mall including 1,251 product images with 10 categories. Experimental results not only show that the proposed method has competitive tagging performance compared with other classifiers but also present that the proposed multi-tagging method based on hypernetworks improves the accuracy of tagging.

Performance Analysis of Exercise Gesture-Recognition Using Convolutional Block Attention Module (합성 블록 어텐션 모듈을 이용한 운동 동작 인식 성능 분석)

  • Kyeong, Chanuk;Jung, Wooyong;Seon, Joonho;Sun, Young-Ghyu;Kim, Jin-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.6
    • /
    • pp.155-161
    • /
    • 2021
  • Gesture recognition analytics through a camera in real time have been widely studied in recent years. Since a small number of features from human joints are extracted, low accuracy of classifying models is get in conventional gesture recognition studies. In this paper, CBAM (Convolutional Block Attention Module) with high accuracy for classifying images is proposed as a classification model and algorithm calculating the angle of joints depending on actions is presented to solve the issues. Employing five exercise gestures images from the fitness posture images provided by AI Hub, the images are applied to the classification model. Important 8-joint angles information for classifying the exercise gestures is extracted from the images by using MediaPipe, a graph-based framework provided by Google. Setting the features as input of the classification model, the classification model is learned. From the simulation results, it is confirmed that the exercise gestures are classified with high accuracy in the proposed model.

Application of CNN for Fish Species Classification (어종 분류를 위한 CNN의 적용)

  • Park, Jin-Hyun;Hwang, Kwang-Bok;Park, Hee-Mun;Choi, Young-Kiu
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.1
    • /
    • pp.39-46
    • /
    • 2019
  • In this study, before system development for the elimination of foreign fish species, we propose an algorithm to classify fish species by training fish images with CNN. The raw data for CNN learning were directly captured images for each species, Dataset 1 increases the number of images to improve the classification of fish species and Dataset 2 realizes images close to natural environment are constructed and used as training and test data. The classification performance of four CNNs are over 99.97% for dataset 1 and 99.5% for dataset 2, in particular, we confirm that the learned CNN using Data Set 2 has satisfactory performance for fish images similar to the natural environment. And among four CNNs, AlexNet achieves satisfactory performance, and this has also the shortest execution time and training time, we confirm that it is the most suitable structure to develop the system for the elimination of foreign fish species.

Analysis of Malignant Tumor Using Texture Characteristics in Breast Ultrasonography (유방 초음파 영상에서 질감 특성을 이용한 악성종양 분석)

  • Cho, Jin-Young;Ye, Soo-Young
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.20 no.2
    • /
    • pp.70-77
    • /
    • 2019
  • Breast ultrasound readings are very important to diagnose early breast cancer. In Ultrasonic inspection, it shows a significant difference in image quality depending on the ultrasonic equipment, and there is a large difference in diagnosis depending on the experience and skill of the inspector. Therefore, objective criteria are needed for accurate diagnosis and treatment. In this study, we analyzed texture characteristics by applying GLCM (Gray Level Co-occurrence Matrix) algorithm and extracted characteristic parameters and diagnosed breast cancer using neural network classifier. Breast ultrasound images were classified into normal, benign and malignant tumors and six texture parameters were extracted. Fourteen cases of normal, malignant and benign tumor diagnosed by mammography were studied by using the extracted six parameters and learning by multi - layer perceptron neural network back propagation learning method. As a result of classification using 51 normal images, 62 benign tumor images, and 74 malignant tumor images of the learned model, the classification rate was 95.2%.

A Study on Facial Wrinkle Detection using Active Appearance Models (AAM을 이용한 얼굴 주름 검출에 관한 연구)

  • Lee, Sang-Bum;Kim, Tae-Mook
    • Journal of Digital Convergence
    • /
    • v.12 no.7
    • /
    • pp.239-245
    • /
    • 2014
  • In this paper, a weighted value wrinkle detection method is suggested based on the analysis on the entire facial features such as face contour, face size, eyes and ears. Firstly, the main facial elements are detected with AAM method entirely from the input screen images. Such elements are mainly composed of shape-based and appearance methods. These are used for learning the facial model and for matching the face from new screen images based on the learned models. Secondly, the face and background are separated in the screen image. Four points with the biggest possibilities for wrinkling are selected from the face and high wrinkle weighted values are assigned to them. Finally, the wrinkles are detected by applying Canny edge algorithm for the interested points of weighted value. The suggested algorithm adopts various screen images for experiment. The experiments display the excellent results of face and wrinkle detection in the most of the screen images.

Facial Recognition Algorithm Based on Edge Detection and Discrete Wavelet Transform

  • Chang, Min-Hyuk;Oh, Mi-Suk;Lim, Chun-Hwan;Ahmad, Muhammad-Bilal;Park, Jong-An
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.3 no.4
    • /
    • pp.283-288
    • /
    • 2001
  • In this paper, we proposed a method for extracting facial characteristics of human being in an image. Given a pair of gray level sample images taken with and without human being, the face of human being is segmented from the image. Noise in the input images is removed with the help of Gaussian filters. Edge maps are found of the two input images. The binary edge differential image is obtained from the difference of the two input edge maps. A mask for face detection is made from the process of erosion followed by dilation on the resulting binary edge differential image. This mask is used to extract the human being from the two input image sequences. Features of face are extracted from the segmented image. An effective recognition system using the discrete wave let transform (DWT) is used for recognition. For extracting the facial features, such as eyebrows, eyes, nose and mouth, edge detector is applied on the segmented face image. The area of eye and the center of face are found from horizontal and vertical components of the edge map of the segmented image. other facial features are obtained from edge information of the image. The characteristic vectors are extrated from DWT of the segmented face image. These characteristic vectors are normalized between +1 and -1, and are used as input vectors for the neural network. Simulation results show recognition rate of 100% on the learned system, and about 92% on the test images.

  • PDF

21 Century Video Image Fashion Communication - Focusing on Prada Fashion Animation - (21세기 영상 패션 커뮤니케이션 - 프라다 패션 애니메이션을 중심으로 -)

  • Jang, Ra-Yoon;Yang, Sook-Hi
    • The Research Journal of the Costume Culture
    • /
    • v.18 no.6
    • /
    • pp.1318-1330
    • /
    • 2010
  • The 21st century is the age when a sensational image has more explanatory power and can deliver a more powerful message than a message consisting of logical thinking. Powerful visual images create a big impact on many people throughout the world, overcoming linguistic barriers and even replacing language as a means of communication. In the fashion field, the concept and power of visual images within the new multimedia of the 21st century are becoming increasingly important. In recent years, other than the above methods, videos, movies and animation features have been produced directly to enhance visual effects and attempts are increasing to use these new tools as communication methods. This study focuses on animation contents that have been used in the fashion industry to overcome prejudice of luxury international brands that feature images that emphasize value, quality and heritage. The purpose of this study is to focus on the specific character of fashion animation in order to overview the concept of 21st video fashion communication and to show how the collection concept that includes color and detail places an emphasis on visual images. Analysis of previous research, theoretical research through literature and case study on Prada fashion animation led to the following conclusion. The common features of two different Prada fashion animation show that both animation have the following features in common : realism, dramatic impact and convergence for expression methods, and creativeness, hybrid and a happy ending for contents. Beginning with this study, I believe that various angles of interest and concern about communication in the fashion world, which is a social and cultural phenomenon that changes rapidly, can be will be looked at and learned from.