• 제목/요약/키워드: deep similarity

검색결과 226건 처리시간 0.023초

Development and Validation of a Deep Learning System for Segmentation of Abdominal Muscle and Fat on Computed Tomography

  • Hyo Jung Park;Yongbin Shin;Jisuk Park;Hyosang Kim;In Seob Lee;Dong-Woo Seo;Jimi Huh;Tae Young Lee;TaeYong Park;Jeongjin Lee;Kyung Won Kim
    • Korean Journal of Radiology
    • /
    • 제21권1호
    • /
    • pp.88-100
    • /
    • 2020
  • Objective: We aimed to develop and validate a deep learning system for fully automated segmentation of abdominal muscle and fat areas on computed tomography (CT) images. Materials and Methods: A fully convolutional network-based segmentation system was developed using a training dataset of 883 CT scans from 467 subjects. Axial CT images obtained at the inferior endplate level of the 3rd lumbar vertebra were used for the analysis. Manually drawn segmentation maps of the skeletal muscle, visceral fat, and subcutaneous fat were created to serve as ground truth data. The performance of the fully convolutional network-based segmentation system was evaluated using the Dice similarity coefficient and cross-sectional area error, for both a separate internal validation dataset (426 CT scans from 308 subjects) and an external validation dataset (171 CT scans from 171 subjects from two outside hospitals). Results: The mean Dice similarity coefficients for muscle, subcutaneous fat, and visceral fat were high for both the internal (0.96, 0.97, and 0.97, respectively) and external (0.97, 0.97, and 0.97, respectively) validation datasets, while the mean cross-sectional area errors for muscle, subcutaneous fat, and visceral fat were low for both internal (2.1%, 3.8%, and 1.8%, respectively) and external (2.7%, 4.6%, and 2.3%, respectively) validation datasets. Conclusion: The fully convolutional network-based segmentation system exhibited high performance and accuracy in the automatic segmentation of abdominal muscle and fat on CT images.

Intelligent missing persons index system Implementation based on the OpenCV image processing and TensorFlow Deep-running Image Processing

  • Baek, Yeong-Tae;Lee, Se-Hoon;Kim, Ji-Seong
    • 한국컴퓨터정보학회논문지
    • /
    • 제22권1호
    • /
    • pp.15-21
    • /
    • 2017
  • In this paper, we present a solution to the problems caused by using only text - based information as an index element when a commercialized missing person indexing system indexes missing persons registered in the database. The existing system could not be used for the missing persons inquiry because it could not formalize the image of the missing person registered together when registering the missing person. To solve these problems, we propose a method to extract the similarity of images by using OpenCV image processing and TensorFlow deep - running image processing, and to process images of missing persons to process them into meaningful information. In order to verify the indexing method used in this paper, we constructed a Web server that operates to provide the information that is most likely to be needed to users first, using the image provided in the non - regular environment of the same subject as the search element.

Cody Recommendation System Using Deep Learning and User Preferences

  • Kwak, Naejoung;Kim, Doyun;kim, Minho;kim, Jongseo;Myung, Sangha;Yoon, Youngbin;Choi, Jihye
    • International Journal of Advanced Culture Technology
    • /
    • 제7권4호
    • /
    • pp.321-326
    • /
    • 2019
  • As AI technology is recently introduced into various fields, it is being applied to the fashion field. This paper proposes a system for recommending cody clothes suitable for a user's selected clothes. The proposed system consists of user app, cody recommendation module, and server interworking of each module and managing database data. Cody recommendation system classifies clothing images into 80 categories composed of feature combinations, selects multiple representative reference images for each category, and selects 3 full body cordy images for each representative reference image. Cody images of the representative reference image were determined by analyzing the user's preference using Google survey app. The proposed algorithm classifies categories the clothing image selected by the user into a category, recognizes the most similar image among the classification category reference images, and transmits the linked cody images to the user's app. The proposed system uses the ResNet-50 model to categorize the input image and measures similarity using ORB and HOG features to select a reference image in the category. We test the proposed algorithm in the Android app, and the result shows that the recommended system runs well.

스마트 홈 환경에서의 딥 러닝을 활용한 다중 거주자 행동 인식에 관한 연구 (A Study on Multiple Resident Activity Recognition using Deep Learning in Smart Home)

  • 지효상;장기영;어준선;양성봉
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2019년도 추계학술발표대회
    • /
    • pp.830-832
    • /
    • 2019
  • IoT 기술의 도래로 인하여 실생활에 사용되는 사물들에 Sensor가 부착되어 시간마다 Sensor data가 발생하는 세상이 열리게 되었다. 이러한 IoT Device들에 부착되어 있는 sensor를 통하여 수집이 된 data는 방대한 양을 가지기 때문에 Deep Learning에 적용하는데 충분하며 아주 중요한 역할을 한다. 이러한 IoT Device들은 우리의 실제 생활에 아주 가까이 다양한 환경으로 접할 수 있다. 예를 들어 스마트시티, 스마트팩토리, 스마트홈 등이 있다. 이러한 것들은 우리의 일상생활에 편리함과 직결되어 있다. 본 논문에서는 Smart home 환경에서의 Multi Resident Activity Recognition이다. Smart home의 가구에 부착되어 있는 센서에서 발생된 센서데이터를 활용하여 1) Training Similarity Network, 2) Embedding, 3) Clustering, 4) Recognizing 네 단계 프로세스를 거쳐 문제를 해결한다. 그 결과, 우리가 제안한 프로세스를 통하여 차원 축소 효과와 Un-seen data를 효과적으로 처리할수 있게 된다.

딥러닝 기반 3차원 라이다의 반사율 세기 신호를 이용한 흑백 영상 생성 기법 (Deep Learning Based Gray Image Generation from 3D LiDAR Reflection Intensity)

  • 김현구;유국열;박주현;정호열
    • 대한임베디드공학회논문지
    • /
    • 제14권1호
    • /
    • pp.1-9
    • /
    • 2019
  • In this paper, we propose a method of generating a 2D gray image from LiDAR 3D reflection intensity. The proposed method uses the Fully Convolutional Network (FCN) to generate the gray image from 2D reflection intensity which is projected from LiDAR 3D intensity. Both encoder and decoder of FCN are configured with several convolution blocks in the symmetric fashion. Each convolution block consists of a convolution layer with $3{\times}3$ filter, batch normalization layer and activation function. The performance of the proposed method architecture is empirically evaluated by varying depths of convolution blocks. The well-known KITTI data set for various scenarios is used for training and performance evaluation. The simulation results show that the proposed method produces the improvements of 8.56 dB in peak signal-to-noise ratio and 0.33 in structural similarity index measure compared with conventional interpolation methods such as inverse distance weighted and nearest neighbor. The proposed method can be possibly used as an assistance tool in the night-time driving system for autonomous vehicles.

Noise2Atom: unsupervised denoising for scanning transmission electron microscopy images

  • Feng Wang;Trond R. Henninen;Debora Keller;Rolf Erni
    • Applied Microscopy
    • /
    • 제50권
    • /
    • pp.23.1-23.9
    • /
    • 2020
  • We propose an effective deep learning model to denoise scanning transmission electron microscopy (STEM) image series, named Noise2Atom, to map images from a source domain 𝓢 to a target domain 𝓒, where 𝓢 is for our noisy experimental dataset, and 𝓒 is for the desired clear atomic images. Noise2Atom uses two external networks to apply additional constraints from the domain knowledge. This model requires no signal prior, no noise model estimation, and no paired training images. The only assumption is that the inputs are acquired with identical experimental configurations. To evaluate the restoration performance of our model, as it is impossible to obtain ground truth for our experimental dataset, we propose consecutive structural similarity (CSS) for image quality assessment, based on the fact that the structures remain much the same as the previous frame(s) within small scan intervals. We demonstrate the superiority of our model by providing evaluation in terms of CSS and visual quality on different experimental datasets.

목조 문화재 영상에서의 크랙을 감지하기 위한 임베딩 유사도 기반 딥러닝 모델 (An Embedding Similarity-based Deep Learning Model for Detecting Displacement in Cultural Asset Images)

  • 강재용;김인기;임현석;곽정환
    • 한국컴퓨터정보학회:학술대회논문집
    • /
    • 한국컴퓨터정보학회 2021년도 제64차 하계학술대회논문집 29권2호
    • /
    • pp.133-135
    • /
    • 2021
  • 본 논문에서는 목조 문화재 영상에서의 변위 현상 중 하나인 크랙이 발생하는 영역을 감지하기 위한 임베딩 유사도 기반 모델을 제안한다. 우선 변위가 존재하지 않는 정상으로만 구성된 학습 이미지는 사전 학습된 합성 곱 신경망을 통과하여 임베딩 벡터들을 추출한다. 그 이후 임베딩 벡터들을 가지고 정상 클래스에 대한 분포의 파라미터 값을 구한다. 실제 추론 과정에 사용되는 테스트 이미지에 대해서도 마찬가지로 임베딩 벡터를 구한다. 그런 다음 테스트 이미지의 임베딩 벡터와 이전에 구한 정상 클래스를 대표하는 가우시안 분포 정보와의 거리를 계산하여 이상치 맵을 생성하여 최종적으로 변위가 존재하는 영역을 감지한다. 데이터 셋으로는 충주시 근처의 문화재에 방문해서 수집한 목조 문화재 이미지를 가지고 정상 및 비정상으로 구분한 데이터 셋을 사용하였다. 실험 결과 우리가 제안한 임베딩 유사도 기반 모델이 목조 문화재에서 크랙이 발생하는 변위 영역을 잘 감지함을 확인하였다. 이러한 결과로부터 우리가 제안한 방법이 목재 문화재의 크랙 현상에 대한 변위 영역 검출에 있어서 매우 적합함을 보여준다.

  • PDF

요추 특징점 추출을 위한 영역 분할 모델의 성능 비교 분석 (A Comparative Performance Analysis of Segmentation Models for Lumbar Key-points Extraction)

  • 유승희;최민호 ;장준수
    • 대한의용생체공학회:의공학회지
    • /
    • 제44권5호
    • /
    • pp.354-361
    • /
    • 2023
  • Most of spinal diseases are diagnosed based on the subjective judgment of a specialist, so numerous studies have been conducted to find objectivity by automating the diagnosis process using deep learning. In this paper, we propose a method that combines segmentation and feature extraction, which are frequently used techniques for diagnosing spinal diseases. Four models, U-Net, U-Net++, DeepLabv3+, and M-Net were trained and compared using 1000 X-ray images, and key-points were derived using Douglas-Peucker algorithms. For evaluation, Dice Similarity Coefficient(DSC), Intersection over Union(IoU), precision, recall, and area under precision-recall curve evaluation metrics were used and U-Net++ showed the best performance in all metrics with an average DSC of 0.9724. For the average Euclidean distance between estimated key-points and ground truth, U-Net was the best, followed by U-Net++. However the difference in average distance was about 0.1 pixels, which is not significant. The results suggest that it is possible to extract key-points based on segmentation and that it can be used to accurately diagnose various spinal diseases, including spondylolisthesis, with consistent criteria.

Deep learning-based automatic segmentation of the mandibular canal on panoramic radiographs: A multi-device study

  • Moe Thu Zar Aung;Sang-Heon Lim;Jiyong Han;Su Yang;Ju-Hee Kang;Jo-Eun Kim;Kyung-Hoe Huh;Won-Jin Yi;Min-Suk Heo;Sam-Sun Lee
    • Imaging Science in Dentistry
    • /
    • 제54권1호
    • /
    • pp.81-91
    • /
    • 2024
  • Purpose: The objective of this study was to propose a deep-learning model for the detection of the mandibular canal on dental panoramic radiographs. Materials and Methods: A total of 2,100 panoramic radiographs (PANs) were collected from 3 different machines: RAYSCAN Alpha (n=700, PAN A), OP-100 (n=700, PAN B), and CS8100 (n=700, PAN C). Initially, an oral and maxillofacial radiologist coarsely annotated the mandibular canals. For deep learning analysis, convolutional neural networks (CNNs) utilizing U-Net architecture were employed for automated canal segmentation. Seven independent networks were trained using training sets representing all possible combinations of the 3 groups. These networks were then assessed using a hold-out test dataset. Results: Among the 7 networks evaluated, the network trained with all 3 available groups achieved an average precision of 90.6%, a recall of 87.4%, and a Dice similarity coefficient (DSC) of 88.9%. The 3 networks trained using each of the 3 possible 2-group combinations also demonstrated reliable performance for mandibular canal segmentation, as follows: 1) PAN A and B exhibited a mean DSC of 87.9%, 2) PAN A and C displayed a mean DSC of 87.8%, and 3) PAN B and C demonstrated a mean DSC of 88.4%. Conclusion: This multi-device study indicated that the examined CNN-based deep learning approach can achieve excellent canal segmentation performance, with a DSC exceeding 88%. Furthermore, the study highlighted the importance of considering the characteristics of panoramic radiographs when developing a robust deep-learning network, rather than depending solely on the size of the dataset.

텍스처 특징 기반 제어점 선택 알고리즘과 병렬 심층 컨볼루션 신경망을 이용한 새로운 얼굴 모핑 방법 (A New Face Morphing Method using Texture Feature-based Control Point Selection Algorithm and Parallel Deep Convolutional Neural Network)

  • 박진혁;;임선자;이석환;권기룡
    • 한국멀티미디어학회논문지
    • /
    • 제25권2호
    • /
    • pp.176-188
    • /
    • 2022
  • In this paper, we propose a compact method for anthropomorphism that uses Deep Convolutional Neural Networks (DCNN) to detect the similarities between a human face and an animal face. We also apply texture feature-based morphing between them. We propose a basic texture feature-based morphing system for morphing between human faces only. The entire anthropomorphism process starts with the creation of an animal face classifier using a parallel DCNN that determines the most similar animal face to a given human face. The significance of our network is that it contains four sets of convolutional functions that run in parallel, allowing it to extract more features than a linear DCNN network. Our employed texture feature algorithm-based automatic morphing system recognizes the facial features of the human face and takes the Control Points automatically, rather than the traditional human aiding manual morphing system, once the similarity was established. The simulation results show that our suggested DCNN surpasses its competitors with a 92.0% accuracy rate. It also ensures that the most similar animal classes are found, and the texture-based morphing technology automatically completes the morphing process, ensuring a smooth transition from one image to another.