• Title/Summary/Keyword: Segmentation model

Search Result 1,041, Processing Time 0.029 seconds

An Average Shape Model for Segmenting Prostate Boundary of TRUS Prostate Image (초음파 전립선 영상에서 전립선 경계 분할을 위한 평균 형상 모델)

  • Kim, Sang Bog;Chung, Joo Young;Seo, Yeong Geon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.5
    • /
    • pp.187-194
    • /
    • 2014
  • Prostate cancer is a malignant tumor occurring in the prostate. Recently, the repetition rate is increasing. Image inspection method which we can check the prostate structure the most correctly is MRI(Magnetic Resonance Imaging), but it is hard to apply it to all the patients because of the cost. So, they use mostly TRUS(Transrectal Ultrasound) images acquired from prostate ultrasound inspection and which are cheap and easy to inspect the prostate in the process of treating and diagnosing the prostate cancer. Traditionally, in the hospital the doctors saw the TRUS images by their eyes and manually segmented the boundary between the prostate and nonprostate. But the manually segmenting process not only needed too much time but also had different boundaries according to the doctor. To cope the problems, some automatic segmentations of the prostate have been studied to generate the constant segmentation results and get the belief from patients. In this study, we propose an average shape model to segment the prostate boundary in TRUS prostate image. The method has 3 steps. First, it finds the probe using edge distribution. Next, it finds two straight lines connected with the probe. Finally it puts the shape model to the image using the position of the probe and straight lines.

Asymmetric Diffusion Model for Protein Spot Matching in 2-DE Image (2차원 전기영동 영상의 단백질 반점 정합을 위한 비대칭 확산 모형)

  • Choi, Kwan-Deok;Yoon, Young-Woo
    • The KIPS Transactions:PartB
    • /
    • v.15B no.6
    • /
    • pp.561-574
    • /
    • 2008
  • The spot detection phase of the 2-DE image analysis program segments a gel image into spot regions by an image segmentation algorithm and fits the spot regions to a spot shape model and quantifies the spot informations for the next phases. Currently the watershed algorithm is generally used as the segmentation algorithm and there are the Gaussian model and the diffusion model for the shape model. The diffusion model is closer to real spot shapes than the Gaussian model however spots have very various shapes and especially an asymmetric formation in x-coordinate and y-coordinate. The reason for asymmetric formation of spots is known that a protein could not be diffused completely because the 2-DE could not be processed under the ideal environment usually. Accordingly we propose an asymmetric diffusion model in this paper. The asymmetric diffusion model assumes that a protein spot is diffused from a disc at initial time of diffusing process, but is diffused asymmetrically for x-axis and y-axis respectively as time goes on. In experiments we processed spot matching for 19 gel images by using three models respectively and evaluated averages of SNR for comparing three models. As averages of SNR we got 14.22dB for the Gaussian model, 20.72dB for the diffusion model and 22.85dB for the asymmetric diffusion model. By experimental results we could confirm the asymmetric diffusion model is more efficient and more adequate for spot matching than the Gaussian model and the diffusion model.

Color Image Segmentation and Textile Texture Mapping of 2D Virtual Wearing System (2D 가상 착의 시스템의 컬러 영상 분할 및 직물 텍스쳐 매핑)

  • Lee, Eun-Hwan;Kwak, No-Yoon
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.35 no.5
    • /
    • pp.213-222
    • /
    • 2008
  • This paper is related to color image segmentation and textile texture mapping for the 2D virtual wearing system. The proposed system is characterized as virtually wearing a new textile pattern selected by user to the clothing shape section, based on its intensity difference map, segmented from a 2D clothes model image using color image segmentation technique. Regardless of color or intensity of model clothes, the proposed system is possible to virtually change the textile pattern or color with holding the illumination and shading properties of the selected clothing shape section, and also to quickly and easily simulate, compare, and select multiple textile pattern combinations for individual styles or entire outfits. The proposed system can provide higher practicality and easy-to-use interface, as it makes real-time processing possible in various digital environment, and creates comparatively natural and realistic virtual wearing styles, and also makes semi-automatic processing possible to reduce the manual works to a minimum. According to the proposed system, it can motivate the creative activity of the designers with simulation results on the effect of textile pattern design on the appearance of clothes without manufacturing physical clothes and, as it can help the purchasers for decision-making with them, promote B2B or B2C e-commerce.

A Robust Object Detection and Tracking Method using RGB-D Model (RGB-D 모델을 이용한 강건한 객체 탐지 및 추적 방법)

  • Park, Seohee;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.18 no.4
    • /
    • pp.61-67
    • /
    • 2017
  • Recently, CCTV has been combined with areas such as big data, artificial intelligence, and image analysis to detect various abnormal behaviors and to detect and analyze the overall situation of objects such as people. Image analysis research for this intelligent video surveillance function is progressing actively. However, CCTV images using 2D information generally have limitations such as object misrecognition due to lack of topological information. This problem can be solved by adding the depth information of the object created by using two cameras to the image. In this paper, we perform background modeling using Mixture of Gaussian technique and detect whether there are moving objects by segmenting the foreground from the modeled background. In order to perform the depth information-based segmentation using the RGB information-based segmentation results, stereo-based depth maps are generated using two cameras. Next, the RGB-based segmented region is set as a domain for extracting depth information, and depth-based segmentation is performed within the domain. In order to detect the center point of a robustly segmented object and to track the direction, the movement of the object is tracked by applying the CAMShift technique, which is the most basic object tracking method. From the experiments, we prove the efficiency of the proposed object detection and tracking method using the RGB-D model.

Dilated convolution and gated linear unit based sound event detection and tagging algorithm using weak label (약한 레이블을 이용한 확장 합성곱 신경망과 게이트 선형 유닛 기반 음향 이벤트 검출 및 태깅 알고리즘)

  • Park, Chungho;Kim, Donghyun;Ko, Hanseok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.39 no.5
    • /
    • pp.414-423
    • /
    • 2020
  • In this paper, we propose a Dilated Convolution Gate Linear Unit (DCGLU) to mitigate the lack of sparsity and small receptive field problems caused by the segmentation map extraction process in sound event detection with weak labels. In the advent of deep learning framework, segmentation map extraction approaches have shown improved performance in noisy environments. However, these methods are forced to maintain the size of the feature map to extract the segmentation map as the model would be constructed without a pooling operation. As a result, the performance of these methods is deteriorated with a lack of sparsity and a small receptive field. To mitigate these problems, we utilize GLU to control the flow of information and Dilated Convolutional Neural Networks (DCNNs) to increase the receptive field without additional learning parameters. For the performance evaluation, we employ a URBAN-SED and self-organized bird sound dataset. The relevant experiments show that our proposed DCGLU model outperforms over other baselines. In particular, our method is shown to exhibit robustness against nature sound noises with three Signal to Noise Ratio (SNR) levels (20 dB, 10 dB and 0 dB).

Semantic Segmentation of the Habitats of Ecklonia Cava and Sargassum in Undersea Images Using HRNet-OCR and Swin-L Models (HRNet-OCR과 Swin-L 모델을 이용한 조식동물 서식지 수중영상의 의미론적 분할)

  • Kim, Hyungwoo;Jang, Seonwoong;Bak, Suho;Gong, Shinwoo;Kwak, Jiwoo;Kim, Jinsoo;Lee, Yangwon
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.5_3
    • /
    • pp.913-924
    • /
    • 2022
  • In this paper, we presented a database construction of undersea images for the Habitats of Ecklonia cava and Sargassum and conducted an experiment for semantic segmentation using state-of-the-art (SOTA) models such as High Resolution Network-Object Contextual Representation (HRNet-OCR) and Shifted Windows-L (Swin-L). The result showed that our segmentation models were superior to the existing experiments in terms of the 29% increased mean intersection over union (mIOU). Swin-L model produced better performance for every class. In particular, the information of the Ecklonia cava class that had small data were also appropriately extracted by Swin-L model. Target objects and the backgrounds were well distinguished owing to the Transformer backbone better than the legacy models. A bigger database under construction will ensure more accuracy improvement and can be utilized as deep learning database for undersea images.

An Object-Level Feature Representation Model for the Multi-target Retrieval of Remote Sensing Images

  • Zeng, Zhi;Du, Zhenhong;Liu, Renyi
    • Journal of Computing Science and Engineering
    • /
    • v.8 no.2
    • /
    • pp.65-77
    • /
    • 2014
  • To address the problem of multi-target retrieval (MTR) of remote sensing images, this study proposes a new object-level feature representation model. The model provides an enhanced application image representation that improves the efficiency of MTR. Generating the model in our scheme includes processes, such as object-oriented image segmentation, feature parameter calculation, and symbolic image database construction. The proposed model uses the spatial representation method of the extended nine-direction lower-triangular (9DLT) matrix to combine spatial relationships among objects, and organizes the image features according to MPEG-7 standards. A similarity metric method is proposed that improves the precision of similarity retrieval. Our method provides a trade-off strategy that supports flexible matching on the target features, or the spatial relationship between the query target and the image database. We implement this retrieval framework on a dataset of remote sensing images. Experimental results show that the proposed model achieves competitive and high-retrieval precision.

3D Building Modeling Using Aerial LiDAR Data (항공 LiDAR 데이터를 이용한 3차원 건물모델링)

  • Cho, Hong-Beom;Cho, Woo-Sug;Park, Jun-Ku;Song, Nak-Hyun
    • Korean Journal of Remote Sensing
    • /
    • v.24 no.2
    • /
    • pp.141-152
    • /
    • 2008
  • The 3D building modeling is one of crucial components in constructing 3D geospatial information. The existing methods for 3D building modeling depend mainly on manual photogrammetric processes, which indeed take great amount of time and efforts. In recent years, many researches on 3D building modeling using aerial LiDAR data have been actively performed to aim at overcoming the limitations of existing 3D building modeling methods. Either techniques with interpolated grid data or data fusion with digital map and images have been investigated in most of existing researches on 3D building modeling with aerial LiDAR data. The paper proposed a method of 3D building modeling with LiDAR data only. Firstly, octree-based segmentation is applied recursively to LiDAR data classified as buildings in 3D space until there are no more LiDAR points to be segmented. Once octree-based segmentation is completed, each segmented patch is thereafter merged together based on its geometric spatial characteristics. Secondly, building model components are created with merged patches. Finally, a 3D building model is generated and composed with building model components. The experimental results with real LiDAR data showed that the proposed method was capable of modeling various types of 3D buildings.

Development and Validation of AI Image Segmentation Model for CT Image-Based Sarcopenia Diagnosis (CT 영상 기반 근감소증 진단을 위한 AI 영상분할 모델 개발 및 검증)

  • Lee Chung-Sub;Lim Dong-Wook;Noh Si-Hyeong;Kim Tae-Hoon;Ko Yousun;Kim Kyung Won;Jeong Chang-Won
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.12 no.3
    • /
    • pp.119-126
    • /
    • 2023
  • Sarcopenia is not well known enough to be classified as a disease in 2021 in Korea, but it is recognized as a social problem in developed countries that have entered an aging society. The diagnosis of sarcopenia follows the international standard guidelines presented by the European Working Group for Sarcopenia in Older People (EWGSOP) and the d Asian Working Group for Sarcopenia (AWGS). Recently, it is recommended to evaluate muscle function by using physical performance evaluation, walking speed measurement, and standing test in addition to absolute muscle mass as a diagnostic method. As a representative method for measuring muscle mass, the body composition analysis method using DEXA has been formally implemented in clinical practice. In addition, various studies for measuring muscle mass using abdominal images of MRI or CT are being actively conducted. In this paper, we develop an AI image segmentation model based on abdominal images of CT with a relatively short imaging time for the diagnosis of sarcopenia and describe the multicenter validation. We developed an artificial intelligence model using U-Net that can automatically segment muscle, subcutaneous fat, and visceral fat by selecting the L3 region from the CT image. Also, to evaluate the performance of the model, internal verification was performed by calculating the intersection over union (IOU) of the partitioned area, and the results of external verification using data from other hospitals are shown. Based on the verification results, we tried to review and supplement the problems and solutions.

Application of Geo-Segment Anything Model (SAM) Scheme to Water Body Segmentation: An Experiment Study Using CAS500-1 Images (수체 추출을 위한 Geo-SAM 기법의 응용: 국토위성영상 적용 실험)

  • Hayoung Lee;Kwangseob Kim;Kiwon Lee
    • Korean Journal of Remote Sensing
    • /
    • v.40 no.4
    • /
    • pp.343-350
    • /
    • 2024
  • Since the release of Meta's Segment Anything Model (SAM), a large-scale vision transformer generation model with rapid image segmentation capabilities, several studies have been conducted to apply this technology in various fields. In this study, we aimed to investigate the applicability of SAM for water bodies detection and extraction using the QGIS Geo-SAM plugin, which enables the use of SAM with satellite imagery. The experimental data consisted of Compact Advanced Satellite 500 (CAS500)-1 images. The results obtained by applying SAM to these data were compared with manually digitized water objects, Open Street Map (OSM), and water body data from the National Geographic Information Institute (NGII)-based hydrological digital map. The mean Intersection over Union (mIoU) calculated for all features extracted using SAM and these three-comparison data were 0.7490, 0.5905, and 0.4921, respectively. For features commonly appeared or extracted in all datasets, the results were 0.9189, 0.8779, and 0.7715, respectively. Based on analysis of the spatial consistency between SAM results and other comparison data, SAM showed limitations in detecting small-scale or poorly defined streams but provided meaningful segmentation results for water body classification.