• Title/Summary/Keyword: camera image

Search Result 4,917, Processing Time 0.151 seconds

Deep Video Stabilization via Optical Flow in Unstable Scenes (동영상 안정화를 위한 옵티컬 플로우의 비지도 학습 방법)

  • Bohee Lee;Kwangsu Kim
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.2
    • /
    • pp.115-127
    • /
    • 2023
  • Video stabilization is one of the camera technologies that the importance is gradually increasing as the personal media market has recently become huge. For deep learning-based video stabilization, existing methods collect pairs of video datas before and after stabilization, but it takes a lot of time and effort to create synchronized datas. Recently, to solve this problem, unsupervised learning method using only unstable video data has been proposed. In this paper, we propose a network structure that learns the stabilized trajectory only with the unstable video image without the pair of unstable and stable video pair using the Convolutional Auto Encoder structure, one of the unsupervised learning methods. Optical flow data is used as network input and output, and optical flow data was mapped into grid units to simplify the network and minimize noise. In addition, to generate a stabilized trajectory with an unsupervised learning method, we define the loss function that smoothing the input optical flow data. And through comparison of the results, we confirmed that the network is learned as intended by the loss function.

Untact-based elevator operating system design using deep learning of private buildings (프라이빗 건물의 딥러닝을 활용한 언택트 기반 엘리베이터 운영시스템 설계)

  • Lee, Min-hye;Kang, Sun-kyoung;Shin, Seong-yoon;Mun, Hyung-jin
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.161-163
    • /
    • 2021
  • In an apartment or private building, it is difficult for the user to operate the elevator button in a similar situation with luggage in both hands. In an environment where human contact must be minimized due to a highly infectious virus such as COVID-19, it is inevitable to operate an elevator based on untact. This paper proposes an operating system capable of operating the elevator by using the user's voice and image processing through the user's face without pressing the elevator button. The elevator can be operated to a designated floor without pressing a button by detecting the face of a person entering the elevator by detecting the person's face from the camera installed in the elevator, matching the information registered in advance. When it is difficult to recognize a person's face, it is intended to enhance the convenience of elevator use in an untouched environment by controlling the floor of the elevator using the user's voice through a microphone and automatically recording access information.

  • PDF

Design and development of non-contact locks including face recognition function based on machine learning (머신러닝 기반 안면인식 기능을 포함한 비접촉 잠금장치 설계 및 개발)

  • Yeo Hoon Yoon;Ki Chang Kim;Whi Jin Jo;Hongjun Kim
    • Convergence Security Journal
    • /
    • v.22 no.1
    • /
    • pp.29-38
    • /
    • 2022
  • The importance of prevention of epidemics is increasing due to the serious spread of infectious diseases. For prevention of epidemics, we need to focus on the non-contact industry. Therefore, in this paper, a face recognition door lock that controls access through non-contact is designed and developed. First very simple features are combined to find objects and face recognition is performed using Haar-based cascade algorithm. Then the texture of the image is binarized to find features using LBPH. An non-contact door lock system which composed of Raspberry PI 3B+ board, an ultrasonic sensor, a camera module, a motor, etc. are suggested. To verify actual performance and ascertain the impact of light sources, various experiment were conducted. As experimental results, the maximum value of the recognition rate was about 85.7%.

Vehicle Type Classification Model based on Deep Learning for Smart Traffic Control Systems (스마트 교통 단속 시스템을 위한 딥러닝 기반 차종 분류 모델)

  • Kim, Doyeong;Jang, Sungjin;Jang, Jongwook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.469-472
    • /
    • 2022
  • With the recent development of intelligent transportation systems, various technologies applying deep learning technology are being used. To crackdown on illegal vehicles and criminal vehicles driving on the road, a vehicle type classification system capable of accurately determining the type of vehicle is required. This study proposes a vehicle type classification system optimized for mobile traffic control systems using YOLO(You Only Look Once). The system uses a one-stage object detection algorithm YOLOv5 to detect vehicles into six classes: passenger cars, subcompact, compact, and midsize vans, full-size vans, trucks, motorcycles, special vehicles, and construction machinery. About 5,000 pieces of domestic vehicle image data built by the Korea Institute of Science and Technology for the development of artificial intelligence technology were used as learning data. It proposes a lane designation control system that applies a vehicle type classification algorithm capable of recognizing both front and side angles with one camera.

  • PDF

Development of CanSat System for Vehicle Tracking based on Jetson Nano (젯슨 나노 기반의 차량 추적 캔위성 시스템 개발)

  • Lee, Younggun;Lee, Sanghyun;You, Seunghoon;Lee, Sangku
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.556-558
    • /
    • 2022
  • This paper proposes a CanSat system with a vehicle tracking function based on Jetson Nano, a high-performance small computer capable of operating artificial intelligence algorithms. The CanSat system consists of a CanSat and a ground station. The CanSat falls in the atmosphere and transmits the data obtained through the installed sensors to the ground station using wireless communication. The existing CanSat is limited to the mission of simply transmitting the collected information to the ground station, and there is a limit to efficiently performing the mission due to the limited fall time and bandwidth limitation of wireless communication. The Jetson Nano based CanSat proposed in this paper uses a pre-trained neural network model to detect the location of a vehicle in each image taken from the air in real time, and then uses a 2-axis motor to move the camera to track the vehicle.

  • PDF

Efficient Video Service Providing Methods for Mobile of Indoor AP Terminals (실내 AP간 단말 이동에 따른 효율적인 동영상 서비스 제공 방안)

  • Hong, Sung-Hwa;Kim, Byoung-Kug
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.585-587
    • /
    • 2022
  • The visual motivation between AP devices is NTP-based visual motivation through the access of the Internet through the internal wired LAN, but this has several seconds of visual difference in hundreds of milliseconds (msec) depending on the network. The frame for the output of the video will vary depending on the application, but usually 24 (image) frames are output to the screen in one second. Therefore, the visual synchronization between peripheral devices can be performed through the adjacent moving camera device, not the wired method. The programming method of generating API for synchronization command when creating an application for visual synchronization and delivering it to AP through MAC may differ from the time in synchronization command according to the environment of the operating system at the transmission side and the situation of the buffer queue of the MAC. Therefore, as a method to solve this problem, the renewal of visual information in the device driver terminal controlling MAC can be much more effective.

  • PDF

EPAR V2.0: AUTOMATED MONITORING AND VISUALIZATION OF POTENTIAL AREAS FOR BUILDING RETROFIT USING THERMAL CAMERAS AND COMPUTATIONAL FLUID DYNAMICS (CFD) MODELS

  • Youngjib Ham;Mani Golparvar-Fard
    • International conference on construction engineering and project management
    • /
    • 2013.01a
    • /
    • pp.279-286
    • /
    • 2013
  • This paper introduces a new method for identification of building energy performance problems. The presented method is based on automated analysis and visualization of deviations between actual and expected energy performance of the building using EPAR (Energy Performance Augmented Reality) models. For generating EPAR models, during building inspections, energy auditors collect a large number of digital and thermal imagery using a consumer-level single thermal camera that has a built-in digital lens. Based on a pipeline of image-based 3D reconstruction algorithms built on GPU and multi-core CPU architecture, 3D geometrical and thermal point cloud models of the building under inspection are automatically generated and integrated. Then, the resulting actual 3D spatio-thermal model and the expected energy performance model simulated using computational fluid dynamics (CFD) analysis are superimposed within an augmented reality environment. Based on the resulting EPAR models which jointly visualize the actual and expected energy performance of the building under inspection, two new algorithms are introduced for quick and reliable identification of potential performance problems: 1) 3D thermal mesh modeling using k-d trees and nearest neighbor searching to automate calculation of temperature deviations; and 2) automated visualization of performance deviations using a metaphor based on traffic light colors. The proposed EPAR v2.0 modeling method is validated on several interior locations of a residential building and an instructional facility. Our empirical observations show that the automated energy performance analysis using EPAR models enables performance deviations to be rapidly and accurately identified. The visualization of performance deviations in 3D enables auditors to easily identify potential building performance problems. Rather than manually analyzing thermal imagery, auditors can focus on other important tasks such as evaluating possible remedial alternatives.

  • PDF

Implementation of ROS-Based Intelligent Unmanned Delivery Robot System (ROS 기반 지능형 무인 배송 로봇 시스템의 구현)

  • Seong-Jin Kong;Won-Chang Lee
    • Journal of IKEEE
    • /
    • v.27 no.4
    • /
    • pp.610-616
    • /
    • 2023
  • In this paper, we implement an unmanned delivery robot system with Robot Operating System(ROS)-based mobile manipulator, and introduce the technologies employed for the system implementation. The robot consists of a mobile robot capable of autonomous navigation inside the building using an elevator and a Selective Compliance Assembly Robot Arm(SCARA)-Type manipulator equipped with a vacuum pump. The robot can determines the position and orientation for picking up a package through image segmentation and corner detection using the camera on the manipulator. The proposed system has a user interface implemented to check the delivery status and determine the real-time location of the robot through a web server linked to the application and ROS, and recognizes the shipment and address at the delivery station through You Only Look Once(YOLO) and Optical Character Recognition(OCR). The effectiveness of the system is validated through delivery experiments conducted within a 4-story building.

Application for Workout and Diet Assistant using Image Processing and Machine Learning Skills (영상처리 및 머신러닝 기술을 이용하는 운동 및 식단 보조 애플리케이션)

  • Chi-Ho Lee;Dong-Hyun Kim;Seung-Ho Choi;In-Woong Hwang;Kyung-Sook Han
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.5
    • /
    • pp.83-88
    • /
    • 2023
  • In this paper, we developed a workout and diet assistance application to meet the growing demand for workout and dietary support services due to the increase in the home training population. The application analyzes the user's workout posture in real-time through the camera and guides the correct posture using guiding lines and voice feedback. It also classifies the foods included in the captured photos, estimates the amount of each food, and calculates and provides nutritional information such as calories. Nutritional information calculations are executed on the server, which then transmits the results back to the application. Once received, this data is presented visually to the user. Additionally, workout results and nutritional information are saved and organized by date for users to review.

A Method of Extracting Features of Sensor-only Facilities for Autonomous Cooperative Driving

  • Hyung Lee;Chulwoo Park;Handong Lee;Sanyeon Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.12
    • /
    • pp.191-199
    • /
    • 2023
  • In this paper, we propose a method to extract the features of five sensor-only facilities built as infrastructure for autonomous cooperative driving, which are from point cloud data acquired by LiDAR. In the case of image acquisition sensors installed in autonomous vehicles, the acquisition data is inconsistent due to the climatic environment and camera characteristics, so LiDAR sensor was applied to replace them. In addition, high-intensity reflectors were designed and attached to each facility to make it easier to distinguish it from other existing facilities with LiDAR. From the five sensor-only facilities developed and the point cloud data acquired by the data acquisition system, feature points were extracted based on the average reflective intensity of the high-intensity reflective paper attached to the facility, clustered by the DBSCAN method, and changed to two-dimensional coordinates by a projection method. The features of the facility at each distance consist of three-dimensional point coordinates, two-dimensional projected coordinates, and reflection intensity, and will be used as training data for a model for facility recognition to be developed in the future.