• 제목/요약/키워드: DeepLab

검색결과 186건 처리시간 0.031초

Relighting 3D Scenes with a Continuously Moving Camera

  • Kim, Soon-Hyun;Kyung, Min-Ho;Lee, Joo-Haeng
    • ETRI Journal
    • /
    • 제31권4호
    • /
    • pp.429-437
    • /
    • 2009
  • This paper proposes a novel technique for 3D scene relighting with interactive viewpoint changes. The proposed technique is based on a deep framebuffer framework for fast relighting computation which adopts image-based techniques to provide arbitrary view-changing. In the preprocessing stage, the shading parameters required for the surface shaders, such as surface color, normal, depth, ambient/diffuse/specular coefficients, and roughness, are cached into multiple deep framebuffers generated by several caching cameras which are created in an automatic manner. When the user designs the lighting setup, the relighting renderer builds a map to connect a screen pixel for the current rendering camera to the corresponding deep framebuffer pixel and then computes illumination at each pixel with the cache values taken from the deep framebuffers. All the relighting computations except the deep framebuffer pre-computation are carried out at interactive rates by the GPU.

Statics variation analysis due to spatially moving of a full ocean depth autonomous underwater vehicle

  • Jiang, Yanqing;Li, Ye;Su, Yumin;Cao, Jian;Li, Yueming;Wang, Youkang;Sun, Yeyi
    • International Journal of Naval Architecture and Ocean Engineering
    • /
    • 제11권1호
    • /
    • pp.448-461
    • /
    • 2019
  • Changes in gravity and buoyancy of a Full Ocean Depth Autonomous Underwater Vehicle (FOD-AUV) during its descending and ascending process must be considered very carefully compared with a Human Occupied Vehicle (HOV) or a Remotely Pperated Vehicle (ROV) whose activities rely on human decision. We firstly designed a two-step weight dropping pattern to achieve a high descending and ascending efficiency and a gravity-buoyancy balance at designed depth. The static equations showed that gravity acceleration, seawater density and displacement are three key aspects affecting the balance. Secondly, we try our best to analysis the gravity and buoyancy changing according to the previous known scientific information, such as anomaly of gravity acceleration, changing of seawater states. Finally, we drew conclusion that gravity changes little (no more than 0.1kgf, it is impossible to give a accurate value). A density-depth relationship at the Challenger Deep was acquired and the displacement changing of the FOD-AUV was calculated preciously.

안개영상의 의미론적 분할 및 안개제거를 위한 심층 멀티태스크 네트워크 (Deep Multi-task Network for Simultaneous Hazy Image Semantic Segmentation and Dehazing)

  • 송태용;장현성;하남구;연윤모;권구용;손광훈
    • 한국멀티미디어학회논문지
    • /
    • 제22권9호
    • /
    • pp.1000-1010
    • /
    • 2019
  • Image semantic segmentation and dehazing are key tasks in the computer vision. In recent years, researches in both tasks have achieved substantial improvements in performance with the development of Convolutional Neural Network (CNN). However, most of the previous works for semantic segmentation assume the images are captured in clear weather and show degraded performance under hazy images with low contrast and faded color. Meanwhile, dehazing aims to recover clear image given observed hazy image, which is an ill-posed problem and can be alleviated with additional information about the image. In this work, we propose a deep multi-task network for simultaneous semantic segmentation and dehazing. The proposed network takes single haze image as input and predicts dense semantic segmentation map and clear image. The visual information getting refined during the dehazing process can help the recognition task of semantic segmentation. On the other hand, semantic features obtained during the semantic segmentation process can provide cues for color priors for objects, which can help dehazing process. Experimental results demonstrate the effectiveness of the proposed multi-task approach, showing improved performance compared to the separate networks.

Convolutional Neural Networks for Character-level Classification

  • Ko, Dae-Gun;Song, Su-Han;Kang, Ki-Min;Han, Seong-Wook
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제6권1호
    • /
    • pp.53-59
    • /
    • 2017
  • Optical character recognition (OCR) automatically recognizes text in an image. OCR is still a challenging problem in computer vision. A successful solution to OCR has important device applications, such as text-to-speech conversion and automatic document classification. In this work, we analyze character recognition performance using the current state-of-the-art deep-learning structures. One is the AlexNet structure, another is the LeNet structure, and the other one is the SPNet structure. For this, we have built our own dataset that contains digits and upper- and lower-case characters. We experiment in the presence of salt-and-pepper noise or Gaussian noise, and report the performance comparison in terms of recognition error. Experimental results indicate by five-fold cross-validation that the SPNet structure (our approach) outperforms AlexNet and LeNet in recognition error.

단일 영상 비균일 블러 제거를 위한 다중 학습 구조 (Multi-task Architecture for Singe Image Dynamic Blur Restoration and Motion Estimation)

  • 정형주;장현성;하남구;연윤모;권구용;손광훈
    • 한국멀티미디어학회논문지
    • /
    • 제22권10호
    • /
    • pp.1149-1159
    • /
    • 2019
  • We present a novel deep learning architecture for obtaining a latent image from a single blurry image, which contains dynamic motion blurs through object/camera movements. The proposed architecture consists of two sub-modules: blur image restoration and optical flow estimation. The tasks are highly related in that object/camera movements make cause blurry artifacts, whereas they are estimated through optical flow. The ablation study demonstrates that training multi-task architecture simultaneously improves both tasks compared to handling them separately. Objective and subjective evaluations show that our method outperforms the state-of-the-arts deep learning based techniques.

Accuracy evaluation of liver and tumor auto-segmentation in CT images using 2D CoordConv DeepLab V3+ model in radiotherapy

  • An, Na young;Kang, Young-nam
    • 대한의용생체공학회:의공학회지
    • /
    • 제43권5호
    • /
    • pp.341-352
    • /
    • 2022
  • Medical image segmentation is the most important task in radiation therapy. Especially, when segmenting medical images, the liver is one of the most difficult organs to segment because it has various shapes and is close to other organs. Therefore, automatic segmentation of the liver in computed tomography (CT) images is a difficult task. Since tumors also have low contrast in surrounding tissues, and the shape, location, size, and number of tumors vary from patient to patient, accurate tumor segmentation takes a long time. In this study, we propose a method algorithm for automatically segmenting the liver and tumor for this purpose. As an advantage of setting the boundaries of the tumor, the liver and tumor were automatically segmented from the CT image using the 2D CoordConv DeepLab V3+ model using the CoordConv layer. For tumors, only cropped liver images were used to improve accuracy. Additionally, to increase the segmentation accuracy, augmentation, preprocess, loss function, and hyperparameter were used to find optimal values. We compared the CoordConv DeepLab v3+ model using the CoordConv layer and the DeepLab V3+ model without the CoordConv layer to determine whether they affected the segmentation accuracy. The data sets used included 131 hepatic tumor segmentation (LiTS) challenge data sets (100 train sets, 16 validation sets, and 15 test sets). Additional learned data were tested using 15 clinical data from Seoul St. Mary's Hospital. The evaluation was compared with the study results learned with a two-dimensional deep learning-based model. Dice values without the CoordConv layer achieved 0.965 ± 0.01 for liver segmentation and 0.925 ± 0.04 for tumor segmentation using the LiTS data set. Results from the clinical data set achieved 0.927 ± 0.02 for liver division and 0.903 ± 0.05 for tumor division. The dice values using the CoordConv layer achieved 0.989 ± 0.02 for liver segmentation and 0.937 ± 0.07 for tumor segmentation using the LiTS data set. Results from the clinical data set achieved 0.944 ± 0.02 for liver division and 0.916 ± 0.18 for tumor division. The use of CoordConv layers improves the segmentation accuracy. The highest of the most recently published values were 0.960 and 0.749 for liver and tumor division, respectively. However, better performance was achieved with 0.989 and 0.937 results for liver and tumor, which would have been used with the algorithm proposed in this study. The algorithm proposed in this study can play a useful role in treatment planning by improving contouring accuracy and reducing time when segmentation evaluation of liver and tumor is performed. And accurate identification of liver anatomy in medical imaging applications, such as surgical planning, as well as radiotherapy, which can leverage the findings of this study, can help clinical evaluation of the risks and benefits of liver intervention.

DeepLabV3+를 이용한 이종 센서의 구름탐지 기법 연구 (A Study on the Cloud Detection Technique of Heterogeneous Sensors Using Modified DeepLabV3+)

  • 김미정;고윤호
    • 대한원격탐사학회지
    • /
    • 제38권5_1호
    • /
    • pp.511-521
    • /
    • 2022
  • 위성영상에서의 구름 탐지 및 제거는 지형관측과 분석을 위해 필수적인 과정이다. 임계값 기반의 구름탐지 기법은 구름의 물리적인 특성을 이용하여 탐지하므로 안정적인 성능을 보여주지만, 긴 연산시간과 모든 채널의 영상 및 메타데이터가 필요하다는 단점을 가지고 있다. 최근 활발히 연구되고 있는 딥러닝을 활용한 구름탐지 기법은 4개 이하의 채널(RGB, NIR) 영상만을 활용하고도 짧은 연산시간과 우수한 성능을 보여주고 있다. 본 논문에서는 해상도가 다른 이종 데이터 셋을 활용하여 학습데이터 셋에 따른 딥러닝 네트워크 성능 의존도를 확인하였다. 이를 위해 DeepLabV3+ 네트워크를 구름탐지의 채널 별 특징이 추출되도록 개선하고 공개된 두 이종 데이터 셋과 혼합 데이터로 각각 학습하였다. 실험결과 테스트 영상과 다른 종류의 영상으로만 학습한 네트워크에서는 낮은 Jaccard 지표를 보여주었다. 그러나 테스트 데이터와 동종의 데이터를 일부 추가한 혼합 데이터로 학습한 네트워크는 높은 Jaccard 지표를 나타내었다. 구름은 사물과 달리 형태가 구조화 되어 있지 않아 공간적인 특성보다 채널 별 특성을 학습에 반영하는 것이 구름 탐지에 효과적이므로 위성 센서의 채널 별 특징을 학습하는 것이 필요하기 때문이다. 본 연구를 통해 해상도가 다른 이종 센서의 구름탐지는 학습데이터 셋에 매우 의존적임을 확인하였다.

심층신경망을 이용한 스마트 양식장용 어류 크기 자동 측정 시스템 (Automatic Fish Size Measurement System for Smart Fish Farm Using a Deep Neural Network)

  • 이윤호;전주현;주문갑
    • 대한임베디드공학회논문지
    • /
    • 제17권3호
    • /
    • pp.177-183
    • /
    • 2022
  • To measure the size and weight of the fish, we developed an automatic fish size measurement system using a deep neural network, where the YOLO (You Only Look Once)v3 model was used. To detect fish, an IP camera with infrared function was installed over the fish pool to acquire image data and used as input data for the deep neural network. Using the bounding box information generated as a result of detecting the fish and the structure for which the actual length is known, the size of the fish can be obtained. A GUI (Graphical User Interface) program was implemented using LabVIEW and RTSP (Real-Time Streaming protocol). The automatic fish size measurement system shows the results and stores them in a database for future work.

Deeplab V3+를 활용한 kidney 탐색 (Kidney Search with Deeplab V3+)

  • 김성중;유재천
    • 한국컴퓨터정보학회:학술대회논문집
    • /
    • 한국컴퓨터정보학회 2020년도 제61차 동계학술대회논문집 28권1호
    • /
    • pp.57-58
    • /
    • 2020
  • 본 논문은 영상분할 기법 중 DeepLab V3+를 적용하여 초음파 영상속에서 특정 장기, 혹은 기관을 발견하고자한다. 그와 동시에 찾아진 Object의 area를 mIOU 기반으로 초음파 영상속에서의 DeepLab V3+의 성능을 확인하고자 한다.

  • PDF