• Title/Summary/Keyword: automated object recognition

Search Result 40, Processing Time 0.022 seconds

Equipment and Worker Recognition of Construction Site with Vision Feature Detection

  • Qi, Shaowen;Shan, Jiazeng;Xu, Lei
    • International Journal of High-Rise Buildings
    • /
    • v.9 no.4
    • /
    • pp.335-342
    • /
    • 2020
  • This article comes up with a new method which is based on the visual characteristic of the objects and machine learning technology to achieve semi-automated recognition of the personnel, machine & materials of the construction sites. Balancing the real-time performance and accuracy, using Faster RCNN (Faster Region-based Convolutional Neural Networks) with transfer learning method appears to be a rational choice. After fine-tuning an ImageNet pre-trained Faster RCNN and testing with it, the result shows that the precision ratio (mAP) has so far reached 67.62%, while the recall ratio (AR) has reached 56.23%. In other word, this recognizing method has achieved rational performance. Further inference with the video of the construction of Huoshenshan Hospital also indicates preliminary success.

A Study on Human-Robot Interface based on Imitative Learning using Computational Model of Mirror Neuron System (Mirror Neuron System 계산 모델을 이용한 모방학습 기반 인간-로봇 인터페이스에 관한 연구)

  • Ko, Kwang-Enu;Sim, Kwee-Bo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.23 no.6
    • /
    • pp.565-570
    • /
    • 2013
  • The mirror neuron regions which are distributed in cortical area handled a functionality of intention recognition on the basis of imitative learning of an observed action which is acquired from visual-information of a goal-directed action. In this paper an automated intention recognition system is proposed by applying computational model of mirror neuron system to the human-robot interaction system. The computational model of mirror neuron system is designed by using dynamic neural networks which have model input which includes sequential feature vector set from the behaviors from the target object and actor and produce results as a form of motor data which can be used to perform the corresponding intentional action through the imitative learning and estimation procedures of the proposed computational model. The intention recognition framework is designed by a system which has a model input from KINECT sensor and has a model output by calculating the corresponding motor data within a virtual robot simulation environment on the basis of intention-related scenario with the limited experimental space and specified target object.

Construction of Database for Deep Learning-based Occlusion Area Detection in the Virtual Environment (가상 환경에서의 딥러닝 기반 폐색영역 검출을 위한 데이터베이스 구축)

  • Kim, Kyeong Su;Lee, Jae In;Gwak, Seok Woo;Kang, Won Yul;Shin, Dae Young;Hwang, Sung Ho
    • Journal of Drive and Control
    • /
    • v.19 no.3
    • /
    • pp.9-15
    • /
    • 2022
  • This paper proposes a method for constructing and verifying datasets used in deep learning technology, to prevent safety accidents in automated construction machinery or autonomous vehicles. Although open datasets for developing image recognition technologies are challenging to meet requirements desired by users, this study proposes the interface of virtual simulators to facilitate the creation of training datasets desired by users. The pixel-level training image dataset was verified by creating scenarios, including various road types and objects in a virtual environment. Detecting an object from an image may interfere with the accurate path determination due to occlusion areas covered by another object. Thus, we construct a database, for developing an occlusion area detection algorithm in a virtual environment. Additionally, we present the possibility of its use as a deep learning dataset to calculate a grid map, that enables path search considering occlusion areas. Custom datasets are built using the RDBMS system.

Automated Construction Progress Management Using Computer Vision-based CNN Model and BIM (이미지 기반 기계 학습과 BIM을 활용한 자동화된 시공 진도 관리 - 합성곱 신경망 모델(CNN)과 실내측위기술, 4D BIM을 기반으로 -)

  • Rho, Juhee;Park, Moonseo;Lee, Hyun-Soo
    • Korean Journal of Construction Engineering and Management
    • /
    • v.21 no.5
    • /
    • pp.11-19
    • /
    • 2020
  • A daily progress monitoring and further schedule management of a construction project have a significant impact on the construction manager's decision making in schedule change and controlling field operation. However, a current site monitoring method highly relies on the manually recorded daily-log book by the person in charge of the work. For this reason, it is difficult to take a detached view and sometimes human error such as omission of contents may occur. In order to resolve these problems, previous researches have developed automated site monitoring method with the object recognition-based visualization or BIM data creation. Despite of the research results along with the related technology development, there are limitations in application targeting the practical construction projects due to the constraints in the experimental methods that assume the fixed equipment at a specific location. To overcome these limitations, some smart devices carried by the field workers can be employed as a medium for data creation. Specifically, the extracted information from the site picture by object recognition technology of CNN model, and positional information by GIPS are applied to update 4D BIM data. A standard CNN model is developed and BIM data modification experiments are conducted with the collected data to validate the research suggestion. Based on the experimental results, it is confirmed that the methods and performance are applicable to the construction site management and further it is expected to contribute speedy and precise data creation with the application of automated progress monitoring methods.

Development of System for Real-Time Object Recognition and Matching using Deep Learning at Simulated Lunar Surface Environment (딥러닝 기반 달 표면 모사 환경 실시간 객체 인식 및 매칭 시스템 개발)

  • Jong-Ho Na;Jun-Ho Gong;Su-Deuk Lee;Hyu-Soung Shin
    • Tunnel and Underground Space
    • /
    • v.33 no.4
    • /
    • pp.281-298
    • /
    • 2023
  • Continuous research efforts are being devoted to unmanned mobile platforms for lunar exploration. There is an ongoing demand for real-time information processing to accurately determine the positioning and mapping of areas of interest on the lunar surface. To apply deep learning processing and analysis techniques to practical rovers, research on software integration and optimization is imperative. In this study, a foundational investigation has been conducted on real-time analysis of virtual lunar base construction site images, aimed at automatically quantifying spatial information of key objects. This study involved transitioning from an existing region-based object recognition algorithm to a boundary box-based algorithm, thus enhancing object recognition accuracy and inference speed. To facilitate extensive data-based object matching training, the Batch Hard Triplet Mining technique was introduced, and research was conducted to optimize both training and inference processes. Furthermore, an improved software system for object recognition and identical object matching was integrated, accompanied by the development of visualization software for the automatic matching of identical objects within input images. Leveraging satellite simulative captured video data for training objects and moving object-captured video data for inference, training and inference for identical object matching were successfully executed. The outcomes of this research suggest the feasibility of implementing 3D spatial information based on continuous-capture video data of mobile platforms and utilizing it for positioning objects within regions of interest. As a result, these findings are expected to contribute to the integration of an automated on-site system for video-based construction monitoring and control of significant target objects within future lunar base construction sites.

Exploring Image Processing and Image Restoration Techniques

  • Omarov, Batyrkhan Sultanovich;Altayeva, Aigerim Bakatkaliyevna;Cho, Young Im
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.15 no.3
    • /
    • pp.172-179
    • /
    • 2015
  • Because of the development of computers and high-technology applications, all devices that we use have become more intelligent. In recent years, security and surveillance systems have become more complicated as well. Before new technologies included video surveillance systems, security cameras were used only for recording events as they occurred, and a human had to analyze the recorded data. Nowadays, computers are used for video analytics, and video surveillance systems have become more autonomous and automated. The types of security cameras have also changed, and the market offers different kinds of cameras with integrated software. Even though there is a variety of hardware, their capabilities leave a lot to be desired. Therefore, this drawback is trying to compensate by dint of computer program solutions. Image processing is a very important part of video surveillance and security systems. Capturing an image exactly as it appears in the real world is difficult if not impossible. There is always noise to deal with. This is caused by the graininess of the emulsion, low resolution of the camera sensors, motion blur caused by movements and drag, focus problems, depth-of-field issues, or the imperfect nature of the camera lens. This paper reviews image processing, pattern recognition, and image digitization techniques, which will be useful in security services, to analyze bio-images, for image restoration, and for object classification.

Building DSMs Generation Integrating Three Line Scanner (TLS) and LiDAR

  • Suh, Yong-Cheol;Nakagawa , Masafumi
    • Korean Journal of Remote Sensing
    • /
    • v.21 no.3
    • /
    • pp.229-242
    • /
    • 2005
  • Photogrammetry is a current method of GIS data acquisition. However, as a matter of fact, a large manpower and expenditure for making detailed 3D spatial information is required especially in urban areas where various buildings exist. There are no photogrammetric systems which can automate a process of spatial information acquisition completely. On the other hand, LiDAR has high potential of automating 3D spatial data acquisition because it can directly measure 3D coordinates of objects, but it is rather difficult to recognize the object with only LiDAR data, for its low resolution at this moment. With this background, we believe that it is very advantageous to integrate LiDAR data and stereo CCD images for more efficient and automated acquisition of the 3D spatial data with higher resolution. In this research, the automatic urban object recognition methodology was proposed by integrating ultra highresolution stereo images and LiDAR data. Moreover, a method to enable more reliable and detailed stereo matching method for CCD images was examined by using LiDAR data as an initial 3D data to determine the search range and to detect possibility of occlusions. Finally, intellectual DSMs, which were identified urban features with high resolution, were generated with high speed processing.

Design of Vehicle-mounted Loading and Unloading Equipment and Autonomous Control Method using Deep Learning Object Detection (차량 탑재형 상·하역 장비의 설계와 딥러닝 객체 인식을 이용한 자동제어 방법)

  • Soon-Kyo Lee;Sunmok Kim;Hyowon Woo;Suk Lee;Ki-Baek Lee
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.1
    • /
    • pp.79-91
    • /
    • 2024
  • Large warehouses are building automation systems to increase efficiency. However, small warehouses, military bases, and local stores are unable to introduce automated logistics systems due to lack of space and budget, and are handling tasks manually, failing to improve efficiency. To solve this problem, this study designed small loading and unloading equipment that can be mounted on transportation vehicles. The equipment can be controlled remotely and is automatically controlled from the point where pallets loaded with cargo are visible using real-time video from an attached camera. Cargo recognition and control command generation for automatic control are achieved through a newly designed deep learning model. This model is designed to be optimized for loading and unloading equipment and mission environments based on the YOLOv3 structure. The trained model recognized 10 types of palettes with different shapes and colors with an average accuracy of 100% and estimated the state with an accuracy of 99.47%. In addition, control commands were created to insert forks into pallets without failure in 14 scenarios assuming actual loading and unloading situations.

Automatic Surface Matching for the Registration of LIDAR Data and MR Imagery

  • Habib, Ayman F.;Cheng, Rita W.T.;Kim, Eui-Myoung;Mitishita, Edson A.;Frayne, Richard;Ronsky, Janet L.
    • ETRI Journal
    • /
    • v.28 no.2
    • /
    • pp.162-174
    • /
    • 2006
  • Several photogrammetric and geographic information system applications such as surface matching, object recognition, city modeling, environmental monitoring, and change detection deal with multiple versions of the same surface that have been derived from different sources and/or at different times. Surface registration is a necessary procedure prior to the manipulation of these 3D datasets. This need is also applicable in the field of medical imaging, where imaging modalities such as magnetic resonance imaging (MRI) can provide temporal 3D imagery for monitoring disease progression. This paper will present a general automated surface registration procedure that can establish correspondences between conjugate surface elements. Experimental results using light detection and ranging (LIDAR) and MRI data will verify the feasibility, robustness, and accuracy of this approach.

  • PDF

Automated Vinyl Green House Identification Method Using Spatial Pattern in High Spatial Resolution Imagery (공간패턴을 이용한 자동 비닐하우스 추출방법)

  • Lee, Jong-Yeol;Kim, Byoung-Sun
    • Korean Journal of Remote Sensing
    • /
    • v.24 no.2
    • /
    • pp.117-124
    • /
    • 2008
  • This paper introduces a novel approach for automated mapping of a map feature that is vinyl green house in high spatial resolution imagery Some map features have their unique spatial patterns. These patterns are normally detected in high spatial resolution remotely sensed data by human recognition system. When spatial patterns can be applied to map feature identification, it will improve image classification accuracy and will be contributed a lot to feature identification. In this study, an automated feature identification approach using spatial aucorrelation is developed, specifically for the vinyl green house that has distinctive spatial pattern in its array. The algorithm aimed to develop the method without any human intervention such as digitizing. The method can investigate the characteristics of repeated spatial pattern of vinyl green house. The repeated spatial pattern comes from the orderly array of vinyl green house. For this, object-based approaches are essential because the pattern is recognized when the shapes that are consists of the groups of pixels are involved. The experimental result shows very effective vinyl house extraction. The targeted three vinyl green houses were exactly identified in the IKONOS image for a part of Jeju area.