• Title/Summary/Keyword: Embedded Training

Search Result 128, Processing Time 0.022 seconds

Hard Example Generation by Novel View Synthesis for 3-D Pose Estimation (3차원 자세 추정 기법의 성능 향상을 위한 임의 시점 합성 기반의 고난도 예제 생성)

  • Minji Kim;Sungchan Kim
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.19 no.1
    • /
    • pp.9-17
    • /
    • 2024
  • It is widely recognized that for 3D human pose estimation (HPE), dataset acquisition is expensive and the effectiveness of augmentation techniques of conventional visual recognition tasks is limited. We address these difficulties by presenting a simple but effective method that augments input images in terms of viewpoints when training a 3D human pose estimation (HPE) model. Our intuition is that meaningful variants of the input images for HPE could be obtained by viewing a human instance in the images from an arbitrary viewpoint different from that in the original images. The core idea is to synthesize new images that have self-occlusion and thus are difficult to predict at different viewpoints even with the same pose of the original example. We incorporate this idea into the training procedure of the 3D HPE model as an augmentation stage of the input samples. We show that a strategy for augmenting the synthesized example should be carefully designed in terms of the frequency of performing the augmentation and the selection of viewpoints for synthesizing the samples. To this end, we propose a new metric to measure the prediction difficulty of input images for 3D HPE in terms of the distance between corresponding keypoints on both sides of a human body. Extensive exploration of the space of augmentation probability choices and example selection according to the proposed distance metric leads to a performance gain of up to 6.2% on Human3.6M, the well-known pose estimation dataset.

Feature Selection for Abnormal Driving Behavior Recognition Based on Variance Distribution of Power Spectral Density

  • Nassuna, Hellen;Kim, Jaehoon;Eyobu, Odongo Steven;Lee, Dongik
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.15 no.3
    • /
    • pp.119-127
    • /
    • 2020
  • The detection and recognition of abnormal driving becomes crucial for achieving safety in Intelligent Transportation Systems (ITS). This paper presents a feature extraction method based on spectral data to train a neural network model for driving behavior recognition. The proposed method uses a two stage signal processing approach to derive time-saving and efficient feature vectors. For the first stage, the feature vector set is obtained by calculating variances from each frequency bin containing the power spectrum data. The feature set is further reduced in the second stage where an intersection method is used to select more significant features that are finally applied for training a neural network model. A stream of live signals are fed to the trained model which recognizes the abnormal driving behaviors. The driving behaviors considered in this study are weaving, sudden braking and normal driving. The effectiveness of the proposed method is demonstrated by comparing with existing methods, which are Particle Swarm Optimization (PSO) and Convolution Neural Network (CNN). The experiments show that the proposed approach achieves satisfactory results with less computational complexity.

Lane Detection System using CNN (CNN을 사용한 차선검출 시스템)

  • Kim, Jihun;Lee, Daesik;Lee, Minho
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.11 no.3
    • /
    • pp.163-171
    • /
    • 2016
  • Lane detection is a widely researched topic. Although simple road detection is easily achieved by previous methods, lane detection becomes very difficult in several complex cases involving noisy edges. To address this, we use a Convolution neural network (CNN) for image enhancement. CNN is a deep learning method that has been very successfully applied in object detection and recognition. In this paper, we introduce a robust lane detection method based on a CNN combined with random sample consensus (RANSAC) algorithm. Initially, we calculate edges in an image using a hat shaped kernel, then we detect lanes using the CNN combined with the RANSAC. In the training process of the CNN, input data consists of edge images and target data is images that have real white color lanes on an otherwise black background. The CNN structure consists of 8 layers with 3 convolutional layers, 2 subsampling layers and multi-layer perceptron (MLP) of 3 fully-connected layers. Convolutional and subsampling layers are hierarchically arranged to form a deep structure. Our proposed lane detection algorithm successfully eliminates noise lines and was found to perform better than other formal line detection algorithms such as RANSAC

Deep Learning Based Gray Image Generation from 3D LiDAR Reflection Intensity (딥러닝 기반 3차원 라이다의 반사율 세기 신호를 이용한 흑백 영상 생성 기법)

  • Kim, Hyun-Koo;Yoo, Kook-Yeol;Park, Ju H.;Jung, Ho-Youl
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.14 no.1
    • /
    • pp.1-9
    • /
    • 2019
  • In this paper, we propose a method of generating a 2D gray image from LiDAR 3D reflection intensity. The proposed method uses the Fully Convolutional Network (FCN) to generate the gray image from 2D reflection intensity which is projected from LiDAR 3D intensity. Both encoder and decoder of FCN are configured with several convolution blocks in the symmetric fashion. Each convolution block consists of a convolution layer with $3{\times}3$ filter, batch normalization layer and activation function. The performance of the proposed method architecture is empirically evaluated by varying depths of convolution blocks. The well-known KITTI data set for various scenarios is used for training and performance evaluation. The simulation results show that the proposed method produces the improvements of 8.56 dB in peak signal-to-noise ratio and 0.33 in structural similarity index measure compared with conventional interpolation methods such as inverse distance weighted and nearest neighbor. The proposed method can be possibly used as an assistance tool in the night-time driving system for autonomous vehicles.

Deep Learning-Based Roundabout Traffic Analysis System Using Unmanned Aerial Vehicle Videos (드론 영상을 이용한 딥러닝 기반 회전 교차로 교통 분석 시스템)

  • Janghoon Lee;Yoonho Hwang;Heejeong Kwon;Ji-Won Choi;Jong Taek Lee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.3
    • /
    • pp.125-132
    • /
    • 2023
  • Roundabouts have strengths in traffic flow and safety but can present difficulties for inexperienced drivers. Demand to acquire and analyze drone images has increased to enhance a traffic environment allowing drivers to deal with roundabouts easily. In this paper, we propose a roundabout traffic analysis system that detects, tracks, and analyzes vehicles using a deep learning-based object detection model (YOLOv7) in drone images. About 3600 images for object detection model learning and testing were extracted and labeled from 1 hour of drone video. Through training diverse conditions and evaluating the performance of object detection models, we achieved an average precision (AP) of up to 97.2%. In addition, we utilized SORT (Simple Online and Realtime Tracking) and OC-SORT (Observation-Centric SORT), a real-time object tracking algorithm, which resulted in an average MOTA (Multiple Object Tracking Accuracy) of up to 89.2%. By implementing a method for measuring roundabout entry speed, we achieved an accuracy of 94.5%.

Deep-Learning-Based Water Shield Automation System by Predicting River Overflow and Vehicle Flooding Possibility (하천 범람 및 차량 침수 가능성 예측을 통한 딥러닝 기반 차수막 자동화 시스템)

  • Seung-Jae Ham;Min-Su Kang;Seong-Woo Jeong;Joonhyuk Yoo
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.3
    • /
    • pp.133-139
    • /
    • 2023
  • This paper proposes a two-stage Water Shield Automation System (WSAS) to predict the possibility of river overflow and vehicle flooding due to sudden rainfall. The WSAS uses a two-stage Deep Neural Network (DNN) model. First, a river overflow prediction module is designed with LSTM to decide whether the river is flooded by predicting the river's water level rise. Second, a vehicle flooding prediction module predicts flooding of underground parking lots by detecting flooded tires with YOLOv5 from CCTV images. Finally, the WSAS automatically installs the water barrier whenever the river overflow and vehicle flooding events happen in the underground parking lots. The only constraint to implementing is that collecting training data for flooded vehicle tires is challenging. This paper exploits the Image C&S data augmentation technique to synthesize flooded tire images. Experimental results validate the superiority of WSAS by showing that the river overflow prediction module can reduce RMSE by three times compared with the previous method, and the vehicle flooding detection module can increase mAP by 20% compared with the naive detection method, respectively.

Artificial Intelligence-Based CW Radar Signal Processing Method for Improving Non-contact Heart Rate Measurement (비접촉형 심박수 측정 정확도 향상을 위한 인공지능 기반 CW 레이더 신호처리)

  • Won Yeol Yoon;Nam Kyu Kwon
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.6
    • /
    • pp.277-283
    • /
    • 2023
  • Vital signals provide essential information regarding the health status of individuals, thereby contributing to health management and medical research. Present monitoring methods, such as ECGs (Electrocardiograms) and smartwatches, demand proximity and fixed postures, which limit their applicability. To address this, Non-contact vital signal measurement methods, such as CW (Continuous-Wave) radar, have emerged as a solution. However, unwanted signal components and a stepwise processing approach lead to errors and limitations in heart rate detection. To overcome these issues, this study introduces an integrated neural network approach that combines noise removal, demodulation, and dominant-frequency detection into a unified process. The neural network employed for signal processing in this research adopts a MLP (Multi-Layer Perceptron) architecture, which analyzes the in-phase and quadrature signals collected within a specified time window, using two distinct input layers. The training of the neural network utilizes CW radar signals and reference heart rates obtained from the ECG. In the experimental evaluation, networks trained on different datasets were compared, and their performance was assessed based on loss and frequency accuracy. The proposed methodology exhibits substantial potential for achieving precise vital signals through non-contact measurements, effectively mitigating the limitations of existing methodologies.

Hyperparameter optimization for Lightweight and Resource-Efficient Deep Learning Model in Human Activity Recognition using Short-range mmWave Radar (mmWave 레이더 기반 사람 행동 인식 딥러닝 모델의 경량화와 자원 효율성을 위한 하이퍼파라미터 최적화 기법)

  • Jiheon Kang
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.6
    • /
    • pp.319-325
    • /
    • 2023
  • In this study, we proposed a method for hyperparameter optimization in the building and training of a deep learning model designed to process point cloud data collected by a millimeter-wave radar system. The primary aim of this study is to facilitate the deployment of a baseline model in resource-constrained IoT devices. We evaluated a RadHAR baseline deep learning model trained on a public dataset composed of point clouds representing five distinct human activities. Additionally, we introduced a coarse-to-fine hyperparameter optimization procedure, showing substantial potential to enhance model efficiency without compromising predictive performance. Experimental results show the feasibility of significantly reducing model size without adversely impacting performance. Specifically, the optimized model demonstrated a 3.3% improvement in classification accuracy despite a 16.8% reduction in number of parameters compared th the baseline model. In conclusion, this research offers valuable insights for the development of deep learning models for resource-constrained IoT devices, underscoring the potential of hyperparameter optimization and model size reduction strategies. This work contributes to enhancing the practicality and usability of deep learning models in real-world environments, where high levels of accuracy and efficiency in data processing and classification tasks are required.

Obstacle Avoidance System for Autonomous CTVs in Offshore Wind Farms Based on Deep Reinforcement Learning (심층 강화학습 기반 자율운항 CTV의 해상풍력발전단지 내 장애물 회피 시스템)

  • Jingyun Kim;Haemyung Chon;Jackyou Noh
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.19 no.3
    • /
    • pp.131-139
    • /
    • 2024
  • Crew Transfer Vessels (CTVs) are primarily used for the maintenance of offshore wind farms. Despite being manually operated by professional captains and crew, collisions with other ships and marine structures still occur. To prevent this, the introduction of autonomous navigation systems to CTVs is necessary. In this study, research on the obstacle avoidance system of the autonomous navigation system for CTVs was conducted. In particular, research on obstacle avoidance simulation for CTVs using deep reinforcement learning was carried out, taking into account the currents and wind loads in offshore wind farms. For this purpose, 3 degrees of freedom ship maneuvering modeling for CTVs considering the currents and wind loads in offshore wind farms was performed, and a simulation environment for offshore wind farms was implemented to train and test the deep reinforcement learning agent. Specifically, this study conducted research on obstacle avoidance maneuvers using MATD3 within deep reinforcement learning, and as a result, it was confirmed that the model, which underwent training over 10,000 episodes, could successfully avoid both static and moving obstacles. This confirms the conclusion that the application of the methods proposed in this study can successfully facilitate obstacle avoidance for autonomous navigation CTVs within offshore wind farms.

Scene Generation of CNC Tools Utilizing Instant NGP and Rendering Performance Evaluation (Instant NGP를 활용한 CNC Tool의 장면 생성 및 렌더링 성능 평가)

  • Taeyeong Jung;Youngjun Yoo
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.19 no.2
    • /
    • pp.83-90
    • /
    • 2024
  • CNC tools contribute to the production of high-precision and consistent results. However, employing damaged CNC tools or utilizing compromised numerical control can lead to significant issues, including equipment damage, overheating, and system-wide errors. Typically, the assessment of external damage to CNC tools involves capturing a single viewpoint through a camera to evaluate tool wear. This study aims to enhance existing methods by using only a single manually focused Microscope camera to enable comprehensive external analysis from multiple perspectives. Applying the NeRF (Neural Radiance Fields) algorithm to images captured with a single manual focus microscope camera, we construct a 3D rendering system. Through this system, it is possible to generate scenes of areas that cannot be captured even with a fixed camera setup, thereby assisting in the analysis of exterior features. However, the NeRF model requires considerable training time, ranging from several hours to over two days. To overcome these limitations of NeRF, various subsequent models have been developed. Therefore, this study aims to compare and apply the performance of Instant NGP, Mip-NeRF, and DS-NeRF, which have garnered attention following NeRF.