• Title/Summary/Keyword: mask R-CNN

Search Result 75, Processing Time 0.023 seconds

Automatic Dataset Generation of Object Detection and Instance Segmentation using Mask R-CNN (Mask R-CNN을 이용한 물체인식 및 개체분할의 학습 데이터셋 자동 생성)

  • Jo, HyunJun;Kim, Dawit;Song, Jae-Bok
    • The Journal of Korea Robotics Society
    • /
    • v.14 no.1
    • /
    • pp.31-39
    • /
    • 2019
  • A robot usually adopts ANN (artificial neural network)-based object detection and instance segmentation algorithms to recognize objects but creating datasets for these algorithms requires high labeling costs because the dataset should be manually labeled. In order to lower the labeling cost, a new scheme is proposed that can automatically generate a training images and label them for specific objects. This scheme uses an instance segmentation algorithm trained to give the masks of unknown objects, so that they can be obtained in a simple environment. The RGB images of objects can be obtained by using these masks, and it is necessary to label the classes of objects through a human supervision. After obtaining object images, they are synthesized with various background images to create new images. Labeling the synthesized images is performed automatically using the masks and previously input object classes. In addition, human intervention is further reduced by using the robot arm to collect object images. The experiments show that the performance of instance segmentation trained through the proposed method is equivalent to that of the real dataset and that the time required to generate the dataset can be significantly reduced.

Separation of Occluding Pigs using Deep Learning-based Image Processing Techniques (딥 러닝 기반의 영상처리 기법을 이용한 겹침 돼지 분리)

  • Lee, Hanhaesol;Sa, Jaewon;Shin, Hyunjun;Chung, Youngwha;Park, Daihee;Kim, Hakjae
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.2
    • /
    • pp.136-145
    • /
    • 2019
  • The crowded environment of a domestic pig farm is highly vulnerable to the spread of infectious diseases such as foot-and-mouth disease, and studies have been conducted to automatically analyze behavior of pigs in a crowded pig farm through a video surveillance system using a camera. Although it is required to correctly separate occluding pigs for tracking each individual pigs, extracting the boundaries of the occluding pigs fast and accurately is a challenging issue due to the complicated occlusion patterns such as X shape and T shape. In this study, we propose a fast and accurate method to separate occluding pigs not only by exploiting the characteristics (i.e., one of the fast deep learning-based object detectors) of You Only Look Once, YOLO, but also by overcoming the limitation (i.e., the bounding box-based object detector) of YOLO with the test-time data augmentation of rotation. Experimental results with two-pigs occlusion patterns show that the proposed method can provide better accuracy and processing speed than one of the state-of-the-art widely used deep learning-based segmentation techniques such as Mask R-CNN (i.e., the performance improvement over Mask R-CNN was about 11 times, in terms of the accuracy/processing speed performance metrics).

Car detection area segmentation using deep learning system

  • Dong-Jin Kwon;Sang-hoon Lee
    • International journal of advanced smart convergence
    • /
    • v.12 no.4
    • /
    • pp.182-189
    • /
    • 2023
  • A recently research, object detection and segmentation have emerged as crucial technologies widely utilized in various fields such as autonomous driving systems, surveillance and image editing. This paper proposes a program that utilizes the QT framework to perform real-time object detection and precise instance segmentation by integrating YOLO(You Only Look Once) and Mask R CNN. This system provides users with a diverse image editing environment, offering features such as selecting specific modes, drawing masks, inspecting detailed image information and employing various image processing techniques, including those based on deep learning. The program advantage the efficiency of YOLO to enable fast and accurate object detection, providing information about bounding boxes. Additionally, it performs precise segmentation using the functionalities of Mask R CNN, allowing users to accurately distinguish and edit objects within images. The QT interface ensures an intuitive and user-friendly environment for program control and enhancing accessibility. Through experiments and evaluations, our proposed system has been demonstrated to be effective in various scenarios. This program provides convenience and powerful image processing and editing capabilities to both beginners and experts, smoothly integrating computer vision technology. This paper contributes to the growth of the computer vision application field and showing the potential to integrate various image processing algorithms on a user-friendly platform

Artificial Intelligence Image Segmentation for Extracting Construction Formwork Elements (거푸집 부재 인식을 위한 인공지능 이미지 분할)

  • Ayesha Munira, Chowdhury;Moon, Sung-Woo
    • Journal of KIBIM
    • /
    • v.12 no.1
    • /
    • pp.1-9
    • /
    • 2022
  • Concrete formwork is a crucial component for any construction project. Artificial intelligence offers great potential to automate formwork design by offering various design options and under different criteria depending on the requirements. This study applied image segmentation in 2D formwork drawings to extract sheathing, strut and pipe support formwork elements. The proposed artificial intelligence model can recognize, classify, and extract formwork elements from 2D CAD drawing image and training and test results confirmed the model performed very well at formwork element recognition with average precision and recall better than 80%. Recognition systems for each formwork element can be implemented later to generate 3D BIM models.

Parking information service system using LoRa network (LoRa 네트워크를 활용한 주차정보 서비스 시스템)

  • Kim, yuchan;Moon, Nammee
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.11a
    • /
    • pp.273-276
    • /
    • 2020
  • 기존의 물리 센서를 활용한 주차 감지는 주차장 규모가 클수록 큰 비용이 필요하고 이미지 기반의 분석은 개별 주차장에 대한 데이터 라벨링과 학습의 노력이 필요했다. 본 논문은 LoRa(Long Range) 네트워크와 마이크로프로세서를 활용한 IoT기반의 시스템으로 영상데이터를 서버로 전송하고 COCO(Common Object in context) 데이터셋으로 학습된 Mask R-CNN 기반의 모델을 활용한 주차장 내 차량점유 감지 알고리즘을 통해 개별 주차장에 대한 학습 또는 라벨링 없이 주차장 내 주차상태를 식별하고 사용자에게 인터페이스를 통해 실시간으로 주차정보를 제공하는 시스템을 구현한다.

  • PDF

Emergency Situation Recognition System Using CCTV and Deep Learning (CCTV와 딥러닝을 이용한 응급 상황 인식 시스템)

  • Park, SeJun;Jeong, Beom-jin;Lee, Jeong-joon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2020.11a
    • /
    • pp.807-809
    • /
    • 2020
  • 기존의 CCTV 관리 체계는 사건·사고에 대한 신속한 조치가 불가능하고 정황 파악이나 증거자료 확보 등 사후조치의 성격이 강하다. 본 논문에서는 Mask R-CNN(Regions with CNN)을 이용하여 CCTV가 읽어 들이는 객체가 응급상황인지 판단하는 방법을 제시한다. 사람으로 인식되는 영역을 다층 퍼셉트론(MLP, Multi-Layer Perceptron)으로 학습시켜 해당 대상이 처한 상황을 인지하고 응급상황으로 인식되는 상황이 지속될 경우 관리 모니터를 통해 사용자에게 알림을 준다. 본 연구를 통해 실시간 상호작용적인 CCTV 관리 체계를 구축하여 도움이 필요한 사람의 골든타임을 놓치지 않게 될 것으로 기대한다.

Fall Situation Recognition by Body Centerline Detection using Deep Learning

  • Kim, Dong-hyeon;Lee, Dong-seok;Kwon, Soon-kak
    • Journal of Multimedia Information System
    • /
    • v.7 no.4
    • /
    • pp.257-262
    • /
    • 2020
  • In this paper, a method of detecting the emergency situations such as body fall is proposed by using color images. We detect body areas and key parts of a body through a pre-learned Mask R-CNN in the images captured by a camera. Then we find the centerline of the body through the joint points of both shoulders and feet. Also, we calculate an angle to the center line and then calculate the amount of change in the angle per hour. If the angle change is more than a certain value, then it is decided as a suspected fall. Also, if the suspected fall state persists for more than a certain frame, then it is determined as a fall situation. Simulation results show that the proposed method can detect body fall situation accurately.

A Dataset of Ground Vehicle Targets from Satellite SAR Images and Its Application to Detection and Instance Segmentation (위성 SAR 영상의 지상차량 표적 데이터 셋 및 탐지와 객체분할로의 적용)

  • Park, Ji-Hoon;Choi, Yeo-Reum;Chae, Dae-Young;Lim, Ho;Yoo, Ji Hee
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.25 no.1
    • /
    • pp.30-44
    • /
    • 2022
  • The advent of deep learning-based algorithms has facilitated researches on target detection from synthetic aperture radar(SAR) imagery. While most of them concentrate on detection tasks for ships with open SAR ship datasets and for aircraft from SAR scenes of airports, there is relatively scarce researches on the detection of SAR ground vehicle targets where several adverse factors such as high false alarm rates, low signal-to-clutter ratios, and multiple targets in close proximity are predicted to degrade the performances. In this paper, a dataset of ground vehicle targets acquired from TerraSAR-X(TSX) satellite SAR images is presented. Then, both detection and instance segmentation are simultaneously carried out on this dataset based on the deep learning-based Mask R-CNN. Finally, this paper shows the future research directions to further improve the performances of detecting the SAR ground vehicle targets.

3D Clothes Modeling of Virtual Human for Metaverse (메타버스를 위한 가상 휴먼의 3차원 의상 모델링)

  • Kim, Hyun Woo;Kim, Dong Eon;Kim, Yujin;Park, In Kyu
    • Journal of Broadcast Engineering
    • /
    • v.27 no.5
    • /
    • pp.638-653
    • /
    • 2022
  • In this paper, we propose the new method of creating 3D virtual-human reflecting the pattern of clothes worn by the person in the high-resolution whole body front image and the body shape data about the person. To get the pattern of clothes, we proceed Instance Segmentation and clothes parsing using Cascade Mask R-CNN. After, we use Pix2Pix to blur the boundaries and estimate the background color and can get UV-Map of 3D clothes mesh proceeding UV-Map base warping. Also, we get the body shape data using SMPL-X and deform the original clothes and body mesh. With UV-Map of clothes and deformed clothes and body mesh, user finally can see the animation of 3D virtual-human reflecting user's appearance by rendering with the state-of-the game engine, i.e. Unreal Engine.

Class Classification and Type of Learning Data by Object for Smart Autonomous Delivery (스마트 자율배송을 위한 클래스 분류와 객체별 학습데이터 유형)

  • Young-Jin Kang;;Jeong, Seok Chan
    • The Journal of Bigdata
    • /
    • v.7 no.1
    • /
    • pp.37-47
    • /
    • 2022
  • Autonomous delivery operation data is the key to driving a paradigm shift for last-mile delivery in the Corona era. To bridge the technological gap between domestic autonomous delivery robots and overseas technology-leading countries, large-scale data collection and verification that can be used for artificial intelligence training is required as the top priority. Therefore, overseas technology-leading countries are contributing to verification and technological development by opening AI training data in public data that anyone can use. In this paper, 326 objects were collected to trainn autonomous delivery robots, and artificial intelligence models such as Mask r-CNN and Yolo v3 were trained and verified. In addition, the two models were compared based on comparison and the elements required for future autonomous delivery robot research were considered.