• Title/Summary/Keyword: 2D Convolutional Neural Network

Search Result 99, Processing Time 0.024 seconds

Depth Map Estimation Model Using 3D Feature Volume (3차원 특징볼륨을 이용한 깊이영상 생성 모델)

  • Shin, Soo-Yeon;Kim, Dong-Myung;Suh, Jae-Won
    • The Journal of the Korea Contents Association
    • /
    • v.18 no.11
    • /
    • pp.447-454
    • /
    • 2018
  • This paper proposes a depth image generation algorithm of stereo images using a deep learning model composed of a CNN (convolutional neural network). The proposed algorithm consists of a feature extraction unit which extracts the main features of each parallax image and a depth learning unit which learns the parallax information using extracted features. First, the feature extraction unit extracts a feature map for each parallax image through the Xception module and the ASPP(Atrous spatial pyramid pooling) module, which are composed of 2D CNN layers. Then, the feature map for each parallax is accumulated in 3D form according to the time difference and the depth image is estimated after passing through the depth learning unit for learning the depth estimation weight through 3D CNN. The proposed algorithm estimates the depth of object region more accurately than other algorithms.

Design of a 1-D CRNN Model for Prediction of Fine Dust Risk Level (미세먼지 위험 단계 예측을 위한 1-D CRNN 모델 설계)

  • Lee, Ki-Hyeok;Hwang, Woo-Sung;Choi, Myung-Ryul
    • Journal of Digital Convergence
    • /
    • v.19 no.2
    • /
    • pp.215-220
    • /
    • 2021
  • In order to reduce the harmful effects on the human body caused by the recent increase in the generation of fine dust in Korea, there is a need for technology to help predict the level of fine dust and take precautions. In this paper, we propose a 1D Convolutional-Recurrent Neural Network (1-D CRNN) model to predict the level of fine dust in Korea. The proposed model is a structure that combines the CNN and the RNN, and uses domestic and foreign fine dust, wind direction, and wind speed data for data prediction. The proposed model achieved an accuracy of about 76%(Partial up to 84%). The proposed model aims to data prediction model for time series data sets that need to consider various data in the future.

LiDAR Sensor based Object Classification System for Delivery Robot Applications (배달 로봇 응용을 위한 LiDAR 센서 기반 객체 분류 시스템)

  • Woo-Jin Park;Jeong-Gyu Lee;Chae-woon Park;Yunho Jung
    • Journal of IKEEE
    • /
    • v.28 no.3
    • /
    • pp.375-381
    • /
    • 2024
  • In this paper, we propose a lightweight object classification system using a LiDAR sensor for delivery service robots. The 3D point cloud data is encoded into a 2D pseudo image using a Pillar Feature Network (PFN), and then passed through a lightweight classification network designed based on Depthwise Separable Convolutional Neural Networks (DS-CNN). The implementation results show that the designed classification network has 9.08K parameters and 3.49M Multiply-Accumulate (MAC) operations, while supporting a classification accuracy of 94.94%.

Recurrent Neural Network Based Spectrum Sensing Technique for Cognitive Radio Communications (인지 무선 통신을 위한 순환 신경망 기반 스펙트럼 센싱 기법)

  • Jung, Tae-Yun;Jeong, Eui-Rim
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.6
    • /
    • pp.759-767
    • /
    • 2020
  • This paper proposes a new Recurrent neural network (RNN) based spectrum sensing technique for cognitive radio communications. The proposed technique determines the existence of primary user's signal without any prior information of the primary users. The method performs high-speed sampling by considering the whole sensing bandwidth and then converts the signal into frequency spectrum via fast Fourier transform (FFT). This spectrum signal is cut in sensing channel bandwidth and entered into the RNN to determine the channel vacancy. The performance of the proposed technique is verified through computer simulations. According to the results, the proposed one is superior to more than 2 [dB] than the existing threshold-based technique and has similar performance to that of the existing Convolutional neural network (CNN) based method. In addition, experiments are carried out in indoor environments and the results show that the proposed technique performs more than 4 [dB] better than both the conventional threshold-based and the CNN based methods.

Fusion System of Time-of-Flight Sensor and Stereo Cameras Considering Single Photon Avalanche Diode and Convolutional Neural Network (SPAD과 CNN의 특성을 반영한 ToF 센서와 스테레오 카메라 융합 시스템)

  • Kim, Dong Yeop;Lee, Jae Min;Jun, Sewoong
    • The Journal of Korea Robotics Society
    • /
    • v.13 no.4
    • /
    • pp.230-236
    • /
    • 2018
  • 3D depth perception has played an important role in robotics, and many sensory methods have also proposed for it. As a photodetector for 3D sensing, single photon avalanche diode (SPAD) is suggested due to sensitivity and accuracy. We have researched for applying a SPAD chip in our fusion system of time-of-fight (ToF) sensor and stereo camera. Our goal is to upsample of SPAD resolution using RGB stereo camera. Currently, we have 64 x 32 resolution SPAD ToF Sensor, even though there are higher resolution depth sensors such as Kinect V2 and Cube-Eye. This may be a weak point of our system, however we exploit this gap using a transition of idea. A convolution neural network (CNN) is designed to upsample our low resolution depth map using the data of the higher resolution depth as label data. Then, the upsampled depth data using CNN and stereo camera depth data are fused using semi-global matching (SGM) algorithm. We proposed simplified fusion method created for the embedded system.

Deep Window Detection in Street Scenes

  • Ma, Wenguang;Ma, Wei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.2
    • /
    • pp.855-870
    • /
    • 2020
  • Windows are key components of building facades. Detecting windows, crucial to 3D semantic reconstruction and scene parsing, is a challenging task in computer vision. Early methods try to solve window detection by using hand-crafted features and traditional classifiers. However, these methods are unable to handle the diversity of window instances in real scenes and suffer from heavy computational costs. Recently, convolutional neural networks based object detection algorithms attract much attention due to their good performances. Unfortunately, directly training them for challenging window detection cannot achieve satisfying results. In this paper, we propose an approach for window detection. It involves an improved Faster R-CNN architecture for window detection, featuring in a window region proposal network, an RoI feature fusion and a context enhancement module. Besides, a post optimization process is designed by the regular distribution of windows to refine detection results obtained by the improved deep architecture. Furthermore, we present a newly collected dataset which is the largest one for window detection in real street scenes to date. Experimental results on both existing datasets and the new dataset show that the proposed method has outstanding performance.

A Multi-Scale Parallel Convolutional Neural Network Based Intelligent Human Identification Using Face Information

  • Li, Chen;Liang, Mengti;Song, Wei;Xiao, Ke
    • Journal of Information Processing Systems
    • /
    • v.14 no.6
    • /
    • pp.1494-1507
    • /
    • 2018
  • Intelligent human identification using face information has been the research hotspot ranging from Internet of Things (IoT) application, intelligent self-service bank, intelligent surveillance to public safety and intelligent access control. Since 2D face images are usually captured from a long distance in an unconstrained environment, to fully exploit this advantage and make human recognition appropriate for wider intelligent applications with higher security and convenience, the key difficulties here include gray scale change caused by illumination variance, occlusion caused by glasses, hair or scarf, self-occlusion and deformation caused by pose or expression variation. To conquer these, many solutions have been proposed. However, most of them only improve recognition performance under one influence factor, which still cannot meet the real face recognition scenario. In this paper we propose a multi-scale parallel convolutional neural network architecture to extract deep robust facial features with high discriminative ability. Abundant experiments are conducted on CMU-PIE, extended FERET and AR database. And the experiment results show that the proposed algorithm exhibits excellent discriminative ability compared with other existing algorithms.

Prediction of aerodynamics using VGG16 and U-Net (VGG16 과 U-Net 구조를 이용한 공력특성 예측)

  • Bo Ra, Kim;Seung Hun, Lee;Seung Hyun, Jang;Gwang Il, Hwang;Min, Yoon
    • Journal of the Korean Society of Visualization
    • /
    • v.20 no.3
    • /
    • pp.109-116
    • /
    • 2022
  • The optimized design of airfoils is essential to increase the performance and efficiency of wind turbines. The aerodynamic characteristics of airfoils near the stall show large deviation from experiments and numerical simulations. Hence, it is needed to perform repetitive analysis of various shapes near the stall. To overcome this, the artificial intelligence is used and combined with numerical simulations. In this study, three types of airfoils are chosen, which are S809, S822 and SD7062 used in wind turbines. A convolutional neural network model is proposed in the combination of VGG16 and U-Net. Learning data are constructed by extracting pressure fields and aerodynamic characteristics through numerical analysis of 2D shape. Based on these data, the pressure field and lift coefficient of untrained airfoils are predicted. As a result, even in untrained airfoils, the pressure field is accurately predicted with an error of within 0.04%.

Deep learning-based target distance and velocity estimation technique for OFDM radars (OFDM 레이다를 위한 딥러닝 기반 표적의 거리 및 속도 추정 기법)

  • Choi, Jae-Woong;Jeong, Eui-Rim
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.1
    • /
    • pp.104-113
    • /
    • 2022
  • In this paper, we propose deep learning-based target distance and velocity estimation technique for OFDM radar systems. In the proposed technique, the 2D periodogram is obtained via 2D fast Fourier transform (FFT) from the reflected signal after removing the modulation effect. The periodogram is the input to the conventional and proposed estimators. The peak of the 2D periodogram represents the target, and the constant false alarm rate (CFAR) algorithm is the most popular conventional technique for the target's distance and speed estimation. In contrast, the proposed method is designed using the multiple output convolutional neural network (CNN). Unlike the conventional CFAR, the proposed estimator is easier to use because it does not require any additional information such as noise power. According to the simulation results, the proposed CNN improves the mean square error (MSE) by more than 5 times compared with the conventional CFAR, and the proposed estimator becomes more accurate as the number of transmitted OFDM symbols increases.

DP-LinkNet: A convolutional network for historical document image binarization

  • Xiong, Wei;Jia, Xiuhong;Yang, Dichun;Ai, Meihui;Li, Lirong;Wang, Song
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.5
    • /
    • pp.1778-1797
    • /
    • 2021
  • Document image binarization is an important pre-processing step in document analysis and archiving. The state-of-the-art models for document image binarization are variants of encoder-decoder architectures, such as FCN (fully convolutional network) and U-Net. Despite their success, they still suffer from three limitations: (1) reduced feature map resolution due to consecutive strided pooling or convolutions, (2) multiple scales of target objects, and (3) reduced localization accuracy due to the built-in invariance of deep convolutional neural networks (DCNNs). To overcome these three challenges, we propose an improved semantic segmentation model, referred to as DP-LinkNet, which adopts the D-LinkNet architecture as its backbone, with the proposed hybrid dilated convolution (HDC) and spatial pyramid pooling (SPP) modules between the encoder and the decoder. Extensive experiments are conducted on recent document image binarization competition (DIBCO) and handwritten document image binarization competition (H-DIBCO) benchmark datasets. Results show that our proposed DP-LinkNet outperforms other state-of-the-art techniques by a large margin. Our implementation and the pre-trained models are available at https://github.com/beargolden/DP-LinkNet.