• Title/Summary/Keyword: Deep learning reconstruction

Search Result 100, Processing Time 0.025 seconds

Deep Learning Algorithm for Simultaneous Noise Reduction and Edge Sharpening in Low-Dose CT Images: A Pilot Study Using Lumbar Spine CT

  • Hyunjung Yeoh;Sung Hwan Hong;Chulkyun Ahn;Ja-Young Choi;Hee-Dong Chae;Hye Jin Yoo;Jong Hyo Kim
    • Korean Journal of Radiology
    • /
    • v.22 no.11
    • /
    • pp.1850-1857
    • /
    • 2021
  • Objective: The purpose of this study was to assess whether a deep learning (DL) algorithm could enable simultaneous noise reduction and edge sharpening in low-dose lumbar spine CT. Materials and Methods: This retrospective study included 52 patients (26 male and 26 female; median age, 60.5 years) who had undergone CT-guided lumbar bone biopsy between October 2015 and April 2020. Initial 100-mAs survey images and 50-mAs intraprocedural images were reconstructed by filtered back projection. Denoising was performed using a vendor-agnostic DL model (ClariCT.AITM, ClariPI) for the 50-mAS images, and the 50-mAs, denoised 50-mAs, and 100-mAs CT images were compared. Noise, signal-to-noise ratio (SNR), and edge rise distance (ERD) for image sharpness were measured. The data were summarized as the mean ± standard deviation for these parameters. Two musculoskeletal radiologists assessed the visibility of the normal anatomical structures. Results: Noise was lower in the denoised 50-mAs images (36.38 ± 7.03 Hounsfield unit [HU]) than the 50-mAs (93.33 ± 25.36 HU) and 100-mAs (63.33 ± 16.09 HU) images (p < 0.001). The SNRs for the images in descending order were as follows: denoised 50-mAs (1.46 ± 0.54), 100-mAs (0.99 ± 0.34), and 50-mAs (0.58 ± 0.18) images (p < 0.001). The denoised 50-mAs images had better edge sharpness than the 100-mAs images at the vertebral body (ERD; 0.94 ± 0.2 mm vs. 1.05 ± 0.24 mm, p = 0.036) and the psoas (ERD; 0.42 ± 0.09 mm vs. 0.50 ± 0.12 mm, p = 0.002). The denoised 50-mAs images significantly improved the visualization of the normal anatomical structures (p < 0.001). Conclusion: DL-based reconstruction may enable simultaneous noise reduction and improvement in image quality with the preservation of edge sharpness on low-dose lumbar spine CT. Investigations on further radiation dose reduction and the clinical applicability of this technique are warranted.

Lightweight Single Image Super-Resolution Convolution Neural Network in Portable Device

  • Wang, Jin;Wu, Yiming;He, Shiming;Sharma, Pradip Kumar;Yu, Xiaofeng;Alfarraj, Osama;Tolba, Amr
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.11
    • /
    • pp.4065-4083
    • /
    • 2021
  • Super-resolution can improve the clarity of low-resolution (LR) images, which can increase the accuracy of high-level compute vision tasks. Portable devices have low computing power and storage performance. Large-scale neural network super-resolution methods are not suitable for portable devices. In order to save the computational cost and the number of parameters, Lightweight image processing method can improve the processing speed of portable devices. Therefore, we propose the Enhanced Information Multiple Distillation Network (EIMDN) to adapt lower delay and cost. The EIMDN takes feedback mechanism as the framework and obtains low level features through high level features. Further, we replace the feature extraction convolution operation in Information Multiple Distillation Block (IMDB), with Ghost module, and propose the Enhanced Information Multiple Distillation Block (EIMDB) to reduce the amount of calculation and the number of parameters. Finally, coordinate attention (CA) is used at the end of IMDB and EIMDB to enhance the important information extraction from Spaces and channels. Experimental results show that our proposed can achieve convergence faster with fewer parameters and computation, compared with other lightweight super-resolution methods. Under the condition of higher peak signal-to-noise ratio (PSNR) and higher structural similarity (SSIM), the performance of network reconstruction image texture and target contour is significantly improved.

Performance Enhancement of Speech Declipping using Clipping Detector (클리핑 감지기를 이용한 음성 신호 클리핑 제거의 성능 향상)

  • Eunmi Seo;Jeongchan Yu;Yujin Lim;Hochong Park
    • Journal of Broadcast Engineering
    • /
    • v.28 no.1
    • /
    • pp.132-140
    • /
    • 2023
  • In this paper, we propose a method for performance enhancement of speech declipping using clipping detector. Clipping occurs when the input speech level exceeds the dynamic range of microphone, and it significantly degrades the speech quality. Recently, many methods for high-performance speech declipping based on machine learning have been developed. However, they often deteriorate the speech signal because of degradation in signal reconstruction process when the degree of clipping is not high. To solve this problem, we propose a new approach that combines the declipping network and clipping detector, which enables a selective declipping operation depending on the clipping level and provides high-quality speech in all clipping levels. We measured the declipping performance using various metrics and confirmed that the proposed method improves the average performance over all clipping levels, compared with the conventional methods, and greatly improves the performance when the clipping distortion is small.

Automated 3D scoring of fluorescence in situ hybridization (FISH) using a confocal whole slide imaging scanner

  • Ziv Frankenstein;Naohiro Uraoka;Umut Aypar;Ruth Aryeequaye;Mamta Rao;Meera Hameed;Yanming Zhang;Yukako Yagi
    • Applied Microscopy
    • /
    • v.51
    • /
    • pp.4.1-4.12
    • /
    • 2021
  • Fluorescence in situ hybridization (FISH) is a technique to visualize specific DNA/RNA sequences within the cell nuclei and provide the presence, location and structural integrity of genes on chromosomes. A confocal Whole Slide Imaging (WSI) scanner technology has superior depth resolution compared to wide-field fluorescence imaging. Confocal WSI has the ability to perform serial optical sections with specimen imaging, which is critical for 3D tissue reconstruction for volumetric spatial analysis. The standard clinical manual scoring for FISH is labor-intensive, time-consuming and subjective. Application of multi-gene FISH analysis alongside 3D imaging, significantly increase the level of complexity required for an accurate 3D analysis. Therefore, the purpose of this study is to establish automated 3D FISH scoring for z-stack images from confocal WSI scanner. The algorithm and the application we developed, SHIMARIS PAFQ, successfully employs 3D calculations for clear individual cell nuclei segmentation, gene signals detection and distribution of break-apart probes signal patterns, including standard break-apart, and variant patterns due to truncation, and deletion, etc. The analysis was accurate and precise when compared with ground truth clinical manual counting and scoring reported in ten lymphoma and solid tumors cases. The algorithm and the application we developed, SHIMARIS PAFQ, is objective and more efficient than the conventional procedure. It enables the automated counting of more nuclei, precisely detecting additional abnormal signal variations in nuclei patterns and analyzes gigabyte multi-layer stacking imaging data of tissue samples from patients. Currently, we are developing a deep learning algorithm for automated tumor area detection to be integrated with SHIMARIS PAFQ.

Evaluation and Prediction of Post-Hepatectomy Liver Failure Using Imaging Techniques: Value of Gadoxetic Acid-Enhanced Magnetic Resonance Imaging

  • Keitaro Sofue;Ryuji Shimada;Eisuke Ueshima;Shohei Komatsu;Takeru Yamaguchi;Shinji Yabe;Yoshiko Ueno;Masatoshi Hori;Takamichi Murakami
    • Korean Journal of Radiology
    • /
    • v.25 no.1
    • /
    • pp.24-32
    • /
    • 2024
  • Despite improvements in operative techniques and perioperative care, post-hepatectomy liver failure (PHLF) remains the most serious cause of morbidity and mortality after surgery, and several risk factors have been identified to predict PHLF. Although volumetric assessment using imaging contributes to surgical simulation by estimating the function of future liver remnants in predicting PHLF, liver function is assumed to be homogeneous throughout the liver. The combination of volumetric and functional analyses may be more useful for an accurate evaluation of liver function and prediction of PHLF than only volumetric analysis. Gadoxetic acid is a hepatocyte-specific magnetic resonance (MR) contrast agent that is taken up by hepatocytes via the OATP1 transporter after intravenous administration. Gadoxetic acid-enhanced MR imaging (MRI) offers information regarding both global and regional functions, leading to a more precise evaluation even in cases with heterogeneous liver function. Various indices, including signal intensity-based methods and MR relaxometry, have been proposed for the estimation of liver function and prediction of PHLF using gadoxetic acid-enhanced MRI. Recent developments in MR techniques, including high-resolution hepatobiliary phase images using deep learning image reconstruction and whole-liver T1 map acquisition, have enabled a more detailed and accurate estimation of liver function in gadoxetic acid-enhanced MRI.

Comparison of Data Reconstruction Methods for Missing Value Imputation (결측값 대체를 위한 데이터 재현 기법 비교)

  • Cheongho Kim;Kee-Hoon Kang
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.1
    • /
    • pp.603-608
    • /
    • 2024
  • Nonresponse and missing values are caused by sample dropouts and avoidance of answers to surveys. In this case, problems with the possibility of information loss and biased reasoning arise, and a replacement of missing values with appropriate values is required. In this paper, as an alternative to missing values imputation, we compare several replacement methods, which use mean, linear regression, random forest, K-nearest neighbor, autoencoder and denoising autoencoder based on deep learning. These methods of imputing missing values are explained, and each method is compared by using continuous simulation data and real data. The comparison results confirm that in most cases, the performance of the random forest imputation method and the denoising autoencoder imputation method are better than the others.

Abnormal Flight Detection Technique of UAV based on U-Net (U-Net을 이용한 무인항공기 비정상 비행 탐지 기법 연구)

  • Myeong Jae Song;Eun Ju Choi;Byoung Soo Kim;Yong Ho Moon
    • Journal of Aerospace System Engineering
    • /
    • v.18 no.3
    • /
    • pp.41-47
    • /
    • 2024
  • Recently, as the practical application and commercialization of unmanned aerial vehicles (UAVs) is pursued, interest in ensuring the safety of the UAV is increasing. Because UAV accidents can result in property damage and loss of life, it is important to develop technology to prevent accidents. For this reason, a technique to detect the abnormal flight state of UAVs has been developed based on the AutoEncoder model. However, the existing detection technique is limited in terms of performance and real-time processing. In this paper, we propose a U-Net based abnormal flight detection technique. In the proposed technique, abnormal flight is detected based on the increasing rate of Mahalanobis distance for the reconstruction error obtained from the U-Net model. Through simulation experiments, it can be shown that the proposed detection technique has superior detection performance compared to the existing detection technique, and can operate in real-time in an on-board environment.

Radiation Dose Reduction in Digital Mammography by Deep-Learning Algorithm Image Reconstruction: A Preliminary Study (딥러닝 알고리즘을 이용한 저선량 디지털 유방 촬영 영상의 복원: 예비 연구)

  • Su Min Ha;Hak Hee Kim;Eunhee Kang;Bo Kyoung Seo;Nami Choi;Tae Hee Kim;You Jin Ku;Jong Chul Ye
    • Journal of the Korean Society of Radiology
    • /
    • v.83 no.2
    • /
    • pp.344-359
    • /
    • 2022
  • Purpose To develop a denoising convolutional neural network-based image processing technique and investigate its efficacy in diagnosing breast cancer using low-dose mammography imaging. Materials and Methods A total of 6 breast radiologists were included in this prospective study. All radiologists independently evaluated low-dose images for lesion detection and rated them for diagnostic quality using a qualitative scale. After application of the denoising network, the same radiologists evaluated lesion detectability and image quality. For clinical application, a consensus on lesion type and localization on preoperative mammographic examinations of breast cancer patients was reached after discussion. Thereafter, coded low-dose, reconstructed full-dose, and full-dose images were presented and assessed in a random order. Results Lesions on 40% reconstructed full-dose images were better perceived when compared with low-dose images of mastectomy specimens as a reference. In clinical application, as compared to 40% reconstructed images, higher values were given on full-dose images for resolution (p < 0.001); diagnostic quality for calcifications (p < 0.001); and for masses, asymmetry, or architectural distortion (p = 0.037). The 40% reconstructed images showed comparable values to 100% full-dose images for overall quality (p = 0.547), lesion visibility (p = 0.120), and contrast (p = 0.083), without significant differences. Conclusion Effective denoising and image reconstruction processing techniques can enable breast cancer diagnosis with substantial radiation dose reduction.

AI-Based Object Recognition Research for Augmented Reality Character Implementation (증강현실 캐릭터 구현을 위한 AI기반 객체인식 연구)

  • Seok-Hwan Lee;Jung-Keum Lee;Hyun Sim
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.6
    • /
    • pp.1321-1330
    • /
    • 2023
  • This study attempts to address the problem of 3D pose estimation for multiple human objects through a single image generated during the character development process that can be used in augmented reality. In the existing top-down method, all objects in the image are first detected, and then each is reconstructed independently. The problem is that inconsistent results may occur due to overlap or depth order mismatch between the reconstructed objects. The goal of this study is to solve these problems and develop a single network that provides consistent 3D reconstruction of all humans in a scene. Integrating a human body model based on the SMPL parametric system into a top-down framework became an important choice. Through this, two types of collision loss based on distance field and loss that considers depth order were introduced. The first loss prevents overlap between reconstructed people, and the second loss adjusts the depth ordering of people to render occlusion inference and annotated instance segmentation consistently. This method allows depth information to be provided to the network without explicit 3D annotation of the image. Experimental results show that this study's methodology performs better than existing methods on standard 3D pose benchmarks, and the proposed losses enable more consistent reconstruction from natural images.

Estimation of Manhattan Coordinate System using Convolutional Neural Network (합성곱 신경망 기반 맨하탄 좌표계 추정)

  • Lee, Jinwoo;Lee, Hyunjoon;Kim, Junho
    • Journal of the Korea Computer Graphics Society
    • /
    • v.23 no.3
    • /
    • pp.31-38
    • /
    • 2017
  • In this paper, we propose a system which estimates Manhattan coordinate systems for urban scene images using a convolutional neural network (CNN). Estimating the Manhattan coordinate system from an image under the Manhattan world assumption is the basis for solving computer graphics and vision problems such as image adjustment and 3D scene reconstruction. We construct a CNN that estimates Manhattan coordinate systems based on GoogLeNet [1]. To train the CNN, we collect about 155,000 images under the Manhattan world assumption by using the Google Street View APIs and calculate Manhattan coordinate systems using existing calibration methods to generate dataset. In contrast to PoseNet [2] that trains per-scene CNNs, our method learns from images under the Manhattan world assumption and thus estimates Manhattan coordinate systems for new images that have not been learned. Experimental results show that our method estimates Manhattan coordinate systems with the median error of $3.157^{\circ}$ for the Google Street View images of non-trained scenes, as test set. In addition, compared to an existing calibration method [3], the proposed method shows lower intermediate errors for the test set.