• Title/Summary/Keyword: RCNN

Search Result 35, Processing Time 0.024 seconds

Development of Mask-RCNN Based Axle Control Violation Detection Method for Enforcement on Overload Trucks (과적 화물차 단속을 위한 Mask-RCNN기반 축조작 검지 기술 개발)

  • Park, Hyun suk;Cho, Yong sung;Kim, Young Nam;Kim, Jin pyung
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.21 no.5
    • /
    • pp.57-66
    • /
    • 2022
  • The Road Management Administration is cracking down on overloaded vehicles by installing low-speed or high-speed WIMs at toll gates and main lines on expressways. However, in recent years, the act of intelligently evading the overloaded-vehicle control system of the Road Management Administration by illegally manipulating the variable axle of an overloaded truck is increasing. In this manipulation, when entering the overloaded-vehicle checkpoint, all axles of the vehicle are lowered to pass normally, and when driving on the main road, the variable axle of the vehicle is illegally lifted with the axle load exceeding 10 tons alarmingly. Therefore, this study developed a technology to detect the state of the variable axle of a truck driving on the road using roadside camera images. In particular, this technology formed the basis for cracking down on overloaded vehicles by lifting the variable axle after entering the checkpoint and linking the vehicle with the account information of the checkpoint. Fundamentally, in this study, the tires of the vehicle were recognized using the Mask RCNN algorithm, the recognized tires were virtually arranged before and after the checkpoint, and the height difference of the vehicle was measured from the arrangement to determine whether the variable axle was lifted after the vehicle left the checkpoint.

Vehicle Manufacturer Recognition using Deep Learning and Perspective Transformation

  • Ansari, Israfil;Shim, Jaechang
    • Journal of Multimedia Information System
    • /
    • v.6 no.4
    • /
    • pp.235-238
    • /
    • 2019
  • In real world object detection is an active research topic for understanding different objects from images. There are different models presented in past and had significant results. In this paper we are presenting vehicle logo detection using previous object detection models such as You only look once (YOLO) and Faster Region-based CNN (F-RCNN). Both the front and rear view of the vehicles were used for training and testing the proposed method. Along with deep learning an image pre-processing algorithm called perspective transformation is proposed for all the test images. Using perspective transformation, the top view images were transformed into front view images. This algorithm has higher detection rate as compared to raw images. Furthermore, YOLO model has better result as compare to F-RCNN model.

Equipment and Worker Recognition of Construction Site with Vision Feature Detection

  • Qi, Shaowen;Shan, Jiazeng;Xu, Lei
    • International Journal of High-Rise Buildings
    • /
    • v.9 no.4
    • /
    • pp.335-342
    • /
    • 2020
  • This article comes up with a new method which is based on the visual characteristic of the objects and machine learning technology to achieve semi-automated recognition of the personnel, machine & materials of the construction sites. Balancing the real-time performance and accuracy, using Faster RCNN (Faster Region-based Convolutional Neural Networks) with transfer learning method appears to be a rational choice. After fine-tuning an ImageNet pre-trained Faster RCNN and testing with it, the result shows that the precision ratio (mAP) has so far reached 67.62%, while the recall ratio (AR) has reached 56.23%. In other word, this recognizing method has achieved rational performance. Further inference with the video of the construction of Huoshenshan Hospital also indicates preliminary success.

Design of disease diagnosis system for pets (반려동물의 질병 진단 시스템)

  • Go, Jun-Hyeok;O, Dong-Hyeop;Lee, Ji-Won;Baek, Chan-Young;Kim, Woo-Sung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.05a
    • /
    • pp.635-636
    • /
    • 2022
  • 본 논문은 딥러닝을 이용해 개인이나, 수의사가 반려동물의 피부병을 특정 하는데 있어서 도움을 줄 수 있는 시스템을 설계하였다. 이 시스템은 사용자가 사용하는 모바일 어플리케이션을 통해 이미지를 수집하고 Mask_RCNN 모델을 사용하여 '구진 플라크','비듬 각질 상피성잔고리', '태선화 과다색소침착', 미란 궤양', '결정 종괴', 농포 여드름'의 6 가지 상태로 분류한 다음 사용자에게 대처법과 병명을 알려주는 반려동물 질병 진단 시스템을 설계하였다.

Deep Learning-based Rice Seed Segmentation for Phynotyping (표현체 연구를 위한 심화학습 기반 벼 종자 분할)

  • Jeong, Yu Seok;Lee, Hong Ro;Baek, Jeong Ho;Kim, Kyung Hwan;Chung, Young Suk;Lee, Chang Woo
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.25 no.5
    • /
    • pp.23-29
    • /
    • 2020
  • The National Institute of Agricultural Sciences of the Rural Developement Administration (NAS, RDA) is conducting various studies on various crops, such as monitoring the cultivation environment and analyzing harvested seeds for high-throughput phenotyping. In this paper, we propose a deep learning-based rice seed segmentation method to analyze the seeds of various crops owned by the NAS. Using Mask-RCNN deep learning model, we perform the rice seed segmentation from manually taken images under specific environment (constant lighting, white background) for analyzing the seed characteristics. For this purpose, we perform the parameter tuning process of the Mask-RCNN model. By the proposed method, the results of the test on seed object detection showed that the accuracy was 82% for rice stem image and 97% for rice grain image, respectively. As a future study, we are planning to researches of more reliable seeds extraction from cluttered seed images by a deep learning-based approach and selection of high-throughput phenotype through precise data analysis such as length, width, and thickness from the detected seed objects.

A Stock Price Prediction Based on Recurrent Convolution Neural Network with Weighted Loss Function (가중치 손실 함수를 가지는 순환 컨볼루션 신경망 기반 주가 예측)

  • Kim, HyunJin;Jung, Yeon Sung
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.8 no.3
    • /
    • pp.123-128
    • /
    • 2019
  • This paper proposes the stock price prediction based on the artificial intelligence, where the model with recurrent convolution neural network (RCNN) layers is adopted. In the motivation of this prediction, long short-term memory model (LSTM)-based neural network can make the output of the time series prediction. On the other hand, the convolution neural network provides the data filtering, averaging, and augmentation. By combining the advantages mentioned above, the proposed technique predicts the estimated stock price of next day. In addition, in order to emphasize the recent time series, a custom weighted loss function is adopted. Moreover, stock data related to the stock price index are adopted to consider the market trends. In the experiments, the proposed stock price prediction reduces the test error by 3.19%, which is over other techniques by about 19%.

Development of Deep Learning-Based Damage Detection Prototype for Concrete Bridge Condition Evaluation (콘크리트 교량 상태평가를 위한 딥러닝 기반 손상 탐지 프로토타입 개발)

  • Nam, Woo-Suk;Jung, Hyunjun;Park, Kyung-Han;Kim, Cheol-Min;Kim, Gyu-Seon
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.42 no.1
    • /
    • pp.107-116
    • /
    • 2022
  • Recently, research has been actively conducted on the technology of inspection facilities through image-based analysis assessment of human-inaccessible facilities. This research was conducted to study the conditions of deep learning-based imaging data on bridges and to develop an evaluation prototype program for bridges. To develop a deep learning-based bridge damage detection prototype, the Semantic Segmentation model, which enables damage detection and quantification among deep learning models, applied Mask-RCNN and constructed learning data 5,140 (including open-data) and labeling suitable for damage types. As a result of performance modeling verification, precision and reproduction rate analysis of concrete cracks, stripping/slapping, rebar exposure and paint stripping showed that the precision was 95.2 %, and the recall was 93.8 %. A 2nd performance verification was performed on onsite data of crack concrete using damage rate of bridge members.

Development of Mask-RCNN Model for Detecting Greenhouses Based on Satellite Image (위성이미지 기반 시설하우스 판별 Mask-RCNN 모델 개발)

  • Kim, Yun Seok;Heo, Seong;Yoon, Seong Uk;Ahn, Jinhyun;Choi, Inchan;Chang, Sungyul;Lee, Seung-Jae;Chung, Yong Suk
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.23 no.3
    • /
    • pp.156-162
    • /
    • 2021
  • The number of smart farms has increased to save labor in agricultural production as the subsidy become available from central and local governments. The number of illegal greenhouses has also increased, which causes serious issues for the local governments. In the present study, we developed Mask-RCNN model to detect greenhouses based on satellite images. Greenhouses in the satellite images were labeled for training and validation of the model. The Mask-RC NN model had the average precision (AP) of 75.6%. The average precision values for 50% and 75% of overlapping area were 91.1% and 81.8%, respectively. This results indicated that the Mask-RC NN model would be useful to detect the greenhouses recently built without proper permission using a periodical screening procedure based on satellite images. Furthermore, the model can be connected with GIS to establish unified management system for greenhouses. It can also be applied to the statistical analysis of the number and total area of greenhouses.

Attention based Feature-Fusion Network for 3D Object Detection (3차원 객체 탐지를 위한 어텐션 기반 특징 융합 네트워크)

  • Sang-Hyun Ryoo;Dae-Yeol Kang;Seung-Jun Hwang;Sung-Jun Park;Joong-Hwan Baek
    • Journal of Advanced Navigation Technology
    • /
    • v.27 no.2
    • /
    • pp.190-196
    • /
    • 2023
  • Recently, following the development of LIDAR technology which can detect distance from the object, the interest for LIDAR based 3D object detection network is getting higher. Previous networks generate inaccurate localization results due to spatial information loss during voxelization and downsampling. In this study, we propose an attention-based convergence method and a camera-LIDAR convergence system to acquire high-level features and high positional accuracy. First, by introducing the attention method into the Voxel-RCNN structure, which is a grid-based 3D object detection network, the multi-scale sparse 3D convolution feature is effectively fused to improve the performance of 3D object detection. Additionally, we propose the late-fusion mechanism for fusing outcomes in 3D object detection network and 2D object detection network to delete false positive. Comparative experiments with existing algorithms are performed using the KITTI data set, which is widely used in the field of autonomous driving. The proposed method showed performance improvement in both 2D object detection on BEV and 3D object detection. In particular, the precision was improved by about 0.54% for the car moderate class compared to Voxel-RCNN.

A method based on Multi-Convolution layers Joint and Generative Adversarial Networks for Vehicle Detection

  • Han, Guang;Su, Jinpeng;Zhang, Chengwei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.4
    • /
    • pp.1795-1811
    • /
    • 2019
  • In order to achieve rapid and accurate detection of vehicle objects in complex traffic conditions, we propose a novel vehicle detection method. Firstly, more contextual and small-object vehicle information can be obtained by our Joint Feature Network (JFN). Secondly, our Evolved Region Proposal Network (EPRN) generates initial anchor boxes by adding an improved version of the region proposal network in this network, and at the same time filters out a large number of false vehicle boxes by soft-Non Maximum Suppression (NMS). Then, our Mask Network (MaskN) generates an example that includes the vehicle occlusion, the generator and discriminator can learn from each other in order to further improve the vehicle object detection capability. Finally, these candidate vehicle detection boxes are optimized to obtain the final vehicle detection boxes by the Fine-Tuning Network(FTN). Through the evaluation experiment on the DETRAC benchmark dataset, we find that in terms of mAP, our method exceeds Faster-RCNN by 11.15%, YOLO by 11.88%, and EB by 1.64%. Besides, our algorithm also has achieved top2 comaring with MS-CNN, YOLO-v3, RefineNet, RetinaNet, Faster-rcnn, DSSD and YOLO-v2 of vehicle category in KITTI dataset.