• Title/Summary/Keyword: 3D LiDAR sensor

Search Result 54, Processing Time 0.026 seconds

Utilizing Airborne LiDAR Data for Building Extraction and Superstructure Analysis for Modeling (항공 LiDAR 데이터를 이용한 건물추출과 상부구조물 특성분석 및 모델링)

  • Jung, Hyung-Sup;Lim, Sae-Bom;Lee, Dong-Cheon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.26 no.3
    • /
    • pp.227-239
    • /
    • 2008
  • Processing LiDAR (Light Detection And Ranging) data obtained from ALS (Airborne Laser Scanning) systems mainly involves organization and segmentation of the data for 3D object modeling and mapping purposes. The ALS systems are viable and becoming more mature technology in various applications. ALS technology requires complex integration of optics, opto-mechanics and electronics in the multi-sensor components, Le. data captured from GPS, INS and laser scanner. In this study, digital image processing techniques mainly were implemented to gray level coded image of the LiDAR data for building extraction and superstructures segmentation. One of the advantages to use gray level image is easy to apply various existing digital image processing algorithms. Gridding and quantization of the raw LiDAR data into limited gray level might introduce smoothing effect and loss of the detail information. However, smoothed surface data that are more suitable for surface patch segmentation and modeling could be obtained by the quantization of the height values. The building boundaries were precisely extracted by the robust edge detection operator and regularized with shape constraints. As for segmentation of the roof structures, basically region growing based and gap filling segmentation methods were implemented. The results present that various image processing methods are applicable to extract buildings and to segment surface patches of the superstructures on the roofs. Finally, conceptual methodology for extracting characteristic information to reconstruct roof shapes was proposed. Statistical and geometric properties were utilized to segment and model superstructures. The simulation results show that segmentation of the roof surface patches and modeling were possible with the proposed method.

Analysis of Optimal Pathways for Terrestrial LiDAR Scanning for the Establishment of Digital Inventory of Forest Resources (디지털 산림자원정보 구축을 위한 최적의 지상LiDAR 스캔 경로 분석)

  • Ko, Chi-Ung;Yim, Jong-Su;Kim, Dong-Geun;Kang, Jin-Taek
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.2
    • /
    • pp.245-256
    • /
    • 2021
  • This study was conducted to identify the applicability of a LiDAR sensor to forest resources inventories by comparing data on a tree's position, height, and DBH obtained by the sensor with those by existing forest inventory methods, for the tree species of Criptomeria japonica in Jeolmul forest in Jeju, South Korea. To this end, a backpack personal LiDAR (Greenvalley International, Model D50) was employed. To facilitate the process of the data collection, patterns of collecting the data by the sensor were divided into seven ones, considering the density of sample plots and the work efficiency. Then, the accuracy of estimating the variables of each tree was assessed. The amount of time spent on acquiring and processing the data by each method was compared to evaluate the efficiency. The findings showed that the rate of detecting standing trees by the LiDAR was 100%. Also, the high statistical accuracy was observed in both Pattern 5 (DBH: RMSE 1.07 cm, Bias -0.79 cm, Height: RMSE 0.95 m, Bias -3.2 m), and Pattern 7 (DBH: RMSE 1.18 cm, Bias -0.82 cm, Height: RMSE 1.13 m, Bias -2.62 m), compared to the results drawn in the typical inventory manner. Concerning the time issue, 115 to 135 minutes per 1ha were taken to process the data by utilizing the LiDAR, while 375 to 1,115 spent in the existing way, proving the higher efficiency of the device. It can thus be concluded that using a backpack personal LiDAR helps increase efficiency in conducting a forest resources inventory in an planted coniferous forest with understory vegetation, implying a need for further research in a variety of forests.

Object Detection and Localization on Map using Multiple Camera and Lidar Point Cloud

  • Pansipansi, Leonardo John;Jang, Minseok;Lee, Yonsik
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.422-424
    • /
    • 2021
  • In this paper, it leads the approach of fusing multiple RGB cameras for visual objects recognition based on deep learning with convolution neural network and 3D Light Detection and Ranging (LiDAR) to observe the environment and match into a 3D world in estimating the distance and position in a form of point cloud map. The goal of perception in multiple cameras are to extract the crucial static and dynamic objects around the autonomous vehicle, especially the blind spot which assists the AV to navigate according to the goal. Numerous cameras with object detection might tend slow-going the computer process in real-time. The computer vision convolution neural network algorithm to use for eradicating this problem use must suitable also to the capacity of the hardware. The localization of classified detected objects comes from the bases of a 3D point cloud environment. But first, the LiDAR point cloud data undergo parsing, and the used algorithm is based on the 3D Euclidean clustering method which gives an accurate on localizing the objects. We evaluated the method using our dataset that comes from VLP-16 and multiple cameras and the results show the completion of the method and multi-sensor fusion strategy.

  • PDF

GENERATION OF AIRBORNE LIDAR INTENSITY IMAGE BY NORMALIZAING RANGE DIFFERENCES

  • Shin, Jung-Il;Yoon, Jong-Suk;Lee, Kyu-Sung
    • Proceedings of the KSRS Conference
    • /
    • v.1
    • /
    • pp.504-507
    • /
    • 2006
  • Airborn Lidar technology has been applied to diverse applications with the advantages of accurate 3D information. Further, Lidar intensity, backscattered signal power, can provid us additional information regarding target's characteristics. Lidar intensity varies by the target reflectance, moisture condition, range, and viewing geometry. This study purposes to generate normalized airborne LiDAR intensity image considering those influential factors such as reflectance, range and geometric/topographic factors (scan angle, ground height, aspect, slope, local incidence angle: LIA). Laser points from one flight line were extracted to simplify the geometric conditions. Laser intensities of sample plots, selected by using a set of reference data and ground survey, werethen statistically analyzed with independent variables. Target reflectance, range between sensor and target, and surface slope were main factors to influence the laser intensity. Intensity of laser points was initially normalized by removing range effect only. However, microsite topographic factor, such as slope angle, was not normalized due to difficulty of automatic calculation.

  • PDF

Building Dataset of Sensor-only Facilities for Autonomous Cooperative Driving

  • Hyung Lee;Chulwoo Park;Handong Lee;Junhyuk Lee
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.1
    • /
    • pp.21-30
    • /
    • 2024
  • In this paper, we propose a method to build a sample dataset of the features of eight sensor-only facilities built as infrastructure for autonomous cooperative driving. The feature extracted from point cloud data acquired by LiDAR and build them into the sample dataset for recognizing the facilities. In order to build the dataset, eight sensor-only facilities with high-brightness reflector sheets and a sensor acquisition system were developed. To extract the features of facilities located within a certain measurement distance from the acquired point cloud data, a cylindrical projection method was applied to the extracted points after applying DBSCAN method for points and then a modified OTSU method for reflected intensity. Coordinates of 3D points, projected coordinates of 2D, and reflection intensity were set as the features of the facility, and the dataset was built along with labels. In order to check the effectiveness of the facility dataset built based on LiDAR data, a common CNN model was selected and tested after training, showing an accuracy of about 90% or more, confirming the possibility of facility recognition. Through continuous experiments, we will improve the feature extraction algorithm for building the proposed dataset and improve its performance, and develop a dedicated model for recognizing sensor-only facilities for autonomous cooperative driving.

Georeferencing of GPR image data using HD map construction method (정밀 도로 지도 구축 방법을 이용한 GPR 영상 데이터 지오레퍼런싱)

  • Shin, Jinsoo;Won, Jonghyun;Lee, Seeyoung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.39 no.6
    • /
    • pp.507-513
    • /
    • 2021
  • GPR (Ground Penetrating RADAR) is a sensor that inspects the pavement state of roads, sinkholes, and underground pipes. It is widely used in road management. MMS (Mobile Mapping System) creates a detailed and accurate road map of the road surface and its surroundings. If both types of data are built in the same area, it is efficient to construct both ground and underground spatial information at the same time. In addition, since it is possible to grasp the road and important facilities around the road, the location of underground pipelines, etc. without special technology, an intuitive understanding of the site is also possible, which is a useful tool in managing the road or facilities. However, overseas equipment to which this latest technology is applied is expensive and does not fit the domestic situation. LiDAR (Light Detection And Raging) and GNSS/INS (Global Navigation Satellite System / Inertial Navigation System) were synchronized in order to replace overseas developed equipment and to secure original technology to develop domestic equipment in the future, and GPR data was also synchronized to the same GNSS/INS. We developed software that performs georeferencing using the location and attitude information from GNSS/INS at the time of acquiring synchronized GPR data. The experiments were conducted on the road site by dividing the open sky and the non-open sky. The road and surrounding facilities on the ground could be easily checked through the 3D point cloud data acquired through LiDAR. Georeferenced GPR data could also be viewed with a 3D viewer along with point cloud data, and the location of underground facilities could be easily and quickly confirmed through GPR data.

Development of the Program for Reconnaissance and Exploratory Drones based on Open Source (오픈 소스 기반의 정찰 및 탐색용 드론 프로그램 개발)

  • Chae, Bum-sug;Kim, Jung-hwan
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.17 no.1
    • /
    • pp.33-40
    • /
    • 2022
  • With the recent increase in the development of military drones, they are adopted and used as the combat system of battalion level or higher. However, it is difficult to use drones that can be used in battles below the platoon level due to the current conditions for the formation of units in the Korean military. In this paper, therefore, we developed a program drones equipped with a thermal imaging camera and LiDAR sensor for reconnaissance and exploration that can be applied in battles below the platoon level. Using these drones, we studied the possibility and feasibility of drones for small-scale combats that can find hidden enemies, search for an appropriate detour through image processing and conduct reconnaissance and search for battlefields, hiding and cover-up through image processing. In addition to the purpose of using the proposed drone to search for an enemies lying in ambush in the battlefield, it can be used as a function to check the optimal movement path when a combat unit is moving, or as a function to check the optimal place for cover-up or hiding. In particular, it is possible to check another route other than the route recommended by the program because the features of the terrain can be checked from various viewpoints through 3D modeling. We verified the possiblity of flying by designing and assembling in a form of adding LiDAR and thermal imaging camera module to a drone assembled based on racing drone parts, which are open source hardware, and developed autonomous flight and search functions which can be used even by non-professional drone operators based on open source software, and then installed them to verify their feasibility.

A CPU-GPU Hybrid System of Environment Perception and 3D Terrain Reconstruction for Unmanned Ground Vehicle

  • Song, Wei;Zou, Shuanghui;Tian, Yifei;Sun, Su;Fong, Simon;Cho, Kyungeun;Qiu, Lvyang
    • Journal of Information Processing Systems
    • /
    • v.14 no.6
    • /
    • pp.1445-1456
    • /
    • 2018
  • Environment perception and three-dimensional (3D) reconstruction tasks are used to provide unmanned ground vehicle (UGV) with driving awareness interfaces. The speed of obstacle segmentation and surrounding terrain reconstruction crucially influences decision making in UGVs. To increase the processing speed of environment information analysis, we develop a CPU-GPU hybrid system of automatic environment perception and 3D terrain reconstruction based on the integration of multiple sensors. The system consists of three functional modules, namely, multi-sensor data collection and pre-processing, environment perception, and 3D reconstruction. To integrate individual datasets collected from different sensors, the pre-processing function registers the sensed LiDAR (light detection and ranging) point clouds, video sequences, and motion information into a global terrain model after filtering redundant and noise data according to the redundancy removal principle. In the environment perception module, the registered discrete points are clustered into ground surface and individual objects by using a ground segmentation method and a connected component labeling algorithm. The estimated ground surface and non-ground objects indicate the terrain to be traversed and obstacles in the environment, thus creating driving awareness. The 3D reconstruction module calibrates the projection matrix between the mounted LiDAR and cameras to map the local point clouds onto the captured video images. Texture meshes and color particle models are used to reconstruct the ground surface and objects of the 3D terrain model, respectively. To accelerate the proposed system, we apply the GPU parallel computation method to implement the applied computer graphics and image processing algorithms in parallel.

Development and Performance Evaluation of Multi-sensor Module for Use in Disaster Sites of Mobile Robot (조사로봇의 재난현장 활용을 위한 다중센서모듈 개발 및 성능평가에 관한 연구)

  • Jung, Yonghan;Hong, Junwooh;Han, Soohee;Shin, Dongyoon;Lim, Eontaek;Kim, Seongsam
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_3
    • /
    • pp.1827-1836
    • /
    • 2022
  • Disasters that occur unexpectedly are difficult to predict. In addition, the scale and damage are increasing compared to the past. Sometimes one disaster can develop into another disaster. Among the four stages of disaster management, search and rescue are carried out in the response stage when an emergency occurs. Therefore, personnel such as firefighters who are put into the scene are put in at a lot of risk. In this respect, in the initial response process at the disaster site, robots are a technology with high potential to reduce damage to human life and property. In addition, Light Detection And Ranging (LiDAR) can acquire a relatively wide range of 3D information using a laser. Due to its high accuracy and precision, it is a very useful sensor when considering the characteristics of a disaster site. Therefore, in this study, development and experiments were conducted so that the robot could perform real-time monitoring at the disaster site. Multi-sensor module was developed by combining LiDAR, Inertial Measurement Unit (IMU) sensor, and computing board. Then, this module was mounted on the robot, and a customized Simultaneous Localization and Mapping (SLAM) algorithm was developed. A method for stably mounting a multi-sensor module to a robot to maintain optimal accuracy at disaster sites was studied. And to check the performance of the module, SLAM was tested inside the disaster building, and various SLAM algorithms and distance comparisons were performed. As a result, PackSLAM developed in this study showed lower error compared to other algorithms, showing the possibility of application in disaster sites. In the future, in order to further enhance usability at disaster sites, various experiments will be conducted by establishing a rough terrain environment with many obstacles.

Collision Avoidance Sensor System for Mobile Crane (전지형 크레인의 인양물 충돌방지를 위한 환경탐지 센서 시스템 개발)

  • Kim, Ji-Chul;Kim, Young Jea;Kim, Mingeuk;Lee, Hanmin
    • Journal of Drive and Control
    • /
    • v.19 no.4
    • /
    • pp.62-69
    • /
    • 2022
  • Construction machinery is exposed to accidents such as collisions, narrowness, and overturns during operation. In particular, mobile crane is operated only with the driver's vision and limited information of the assistant worker. Thus, there is a high risk of an accident. Recently, some collision avoidance device using sensors such as cameras and LiDAR have been applied. However, they are still insufficient to prevent collisions in the omnidirectional 3D space. In this study, a rotating LiDAR device was developed and applied to a 250-ton crane to obtain a full-space point cloud. An algorithm that could provide distance information and safety status to the driver was developed. Also, deep-learning segmentation algorithm was used to classify human-worker. The developed device could recognize obstacles within 100m of a 360-degree range. In the experiment, a safety distance was calculated with an error of 10.3cm at 30m to give the operator an accurate distance and collision alarm.