• Title/Summary/Keyword: video-based recognition system

Search Result 192, Processing Time 0.044 seconds

A Study on the Copyright Protection and Improvement of Digital Image (디지털 이미지의 저작권 보호와 개선방안에 관한 연구)

  • Song, Ho-Jin;Jeong, Eui-Tae
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.1
    • /
    • pp.303-308
    • /
    • 2018
  • In copyright rights of the digital image, copyright holders need to protect copyrights and improve the improvement of copyright laws through accurate understanding of copyright laws, and also to improve the effectiveness of copyright rights of copyright holders and the effectiveness of users' rights. First of all, the scope of the study is based on the case of recognition of the scope of the identity of the same scope and the copyright on the copyright law of the digital image, and the application of the image of the image of the image-writing image and the use of image artifacts based on the image copyright classification system suitable for the user's application.

A Study on Radar Video Fusion Systems for Pedestrian and Vehicle Detection (보행자 및 차량 검지를 위한 레이더 영상 융복합 시스템 연구)

  • Sung-Youn Cho;Yeo-Hwan Yoon
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.24 no.1
    • /
    • pp.197-205
    • /
    • 2024
  • Development of AI and big data-based algorithms to advance and optimize the recognition and detection performance of various static/dynamic vehicles in front and around the vehicle at a time when securing driving safety is the most important point in the development and commercialization of autonomous vehicles. etc. are being studied. However, there are many research cases for recognizing the same vehicle by using the unique advantages of radar and camera, but deep learning image processing technology is not used, or only a short distance is detected as the same target due to radar performance problems. Therefore, there is a need for a convergence-based vehicle recognition method that configures a dataset that can be collected from radar equipment and camera equipment, calculates the error of the dataset, and recognizes it as the same target. In this paper, we aim to develop a technology that can link location information according to the installation location because data errors occur because it is judged as the same object depending on the installation location of the radar and CCTV (video).

A Study on Swarm Robot-Based Invader-Enclosing Technique on Multiple Distributed Object Environments

  • Ko, Kwang-Eun;Park, Seung-Min;Park, Jun-Heong;Sim, Kwee-Bo
    • Journal of Electrical Engineering and Technology
    • /
    • v.6 no.6
    • /
    • pp.806-816
    • /
    • 2011
  • Interest about social security has recently increased in favor of safety for infrastructure. In addition, advances in computer vision and pattern recognition research are leading to video-based surveillance systems with improved scene analysis capabilities. However, such video surveillance systems, which are controlled by human operators, cannot actively cope with dynamic and anomalous events, such as having an invader in the corporate, commercial, or public sectors. For this reason, intelligent surveillance systems are increasingly needed to provide active social security services. In this study, we propose a core technique for intelligent surveillance system that is based on swarm robot technology. We present techniques for invader enclosing using swarm robots based on multiple distributed object environment. The proposed methods are composed of three main stages: location estimation of the object, specified object tracking, and decision of the cooperative behavior of the swarm robots. By using particle filter, object tracking and location estimation procedures are performed and a specified enclosing point for the swarm robots is located on the interactive positions in their coordinate system. Furthermore, the cooperative behaviors of the swarm robots are determined via the result of path navigation based on the combination of potential field and wall-following methods. The results of each stage are combined into the swarm robot-based invader-enclosing technique on multiple distributed object environments. Finally, several simulation results are provided to further discuss and verify the accuracy and effectiveness of the proposed techniques.

Hand Gesture based Manipulation of Meeting Data in Teleconference (핸드제스처를 이용한 원격미팅 자료 인터페이스)

  • Song, Je-Hoon;Choi, Ki-Ho;Kim, Jong-Won;Lee, Yong-Gu
    • Korean Journal of Computational Design and Engineering
    • /
    • v.12 no.2
    • /
    • pp.126-136
    • /
    • 2007
  • Teleconferences have been used in business sectors to reduce traveling costs. Traditionally, specialized telephones that enabled multiparty conversations were used. With the introduction of high speed networks, we now have high definition videos that add more realism in the presence of counterparts who could be thousands of miles away. This paper presents a new technology that adds even more realism by telecommunicating with hand gestures. This technology is part of a teleconference system named SMS (Smart Meeting Space). In SMS, a person can use hand gestures to manipulate meeting data that could be in the form of text, audio, video or 3D shapes. Fer detecting hand gestures, a machine learning algorithm called SVM (Support Vector Machine) has been used. For the prototype system, a 3D interaction environment has been implemented with $OpenGL^{TM}$, where a 3D human skull model can be grasped and moved in 6-DOF during a remote conversation between distant persons.

Automatic Person Identification using Multiple Cues

  • Swangpol, Danuwat;Chalidabhongse, Thanarat
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1202-1205
    • /
    • 2005
  • This paper describes a method for vision-based person identification that can detect, track, and recognize person from video using multiple cues: height and dressing colors. The method does not require constrained target's pose or fully frontal face image to identify the person. First, the system, which is connected to a pan-tilt-zoom camera, detects target using motion detection and human cardboard model. The system keeps tracking the moving target while it is trying to identify whether it is a human and identify who it is among the registered persons in the database. To segment the moving target from the background scene, we employ a version of background subtraction technique and some spatial filtering. Once the target is segmented, we then align the target with the generic human cardboard model to verify whether the detected target is a human. If the target is identified as a human, the card board model is also used to segment the body parts to obtain some salient features such as head, torso, and legs. The whole body silhouette is also analyzed to obtain the target's shape information such as height and slimness. We then use these multiple cues (at present, we uses shirt color, trousers color, and body height) to recognize the target using a supervised self-organization process. We preliminary tested the system on a set of 5 subjects with multiple clothes. The recognition rate is 100% if the person is wearing the clothes that were learned before. In case a person wears new dresses the system fail to identify. This means height is not enough to classify persons. We plan to extend the work by adding more cues such as skin color, and face recognition by utilizing the zoom capability of the camera to obtain high resolution view of face; then, evaluate the system with more subjects.

  • PDF

Development of Lane and Vehicle Headway Direction Recognition System for Military Heavy Equipment's Safe Transport - Based on Kalman Filter and Neural Network - (안전한 군용 중장비 수송을 위한 차선 및 차량 진행 방향 인식 시스템 개발 - 칼만 필터와 신경망을 기반으로 -)

  • Choi, Yeong-Yoon;Choi, Kwang-Mo;Moon, Ho-Seok
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.10 no.3
    • /
    • pp.139-147
    • /
    • 2007
  • In military transportation, the use of wide trailer for transporting the large and heavy weight equipments such as tank, armoured vehicle, and mobile gunnery is quite common. So, the vulnerability of causing traffic accidents for these wide military trailer to bump or collide with another car in adjacent lane is very high due to its broad width in excess of its own lane's width. Also, the possibility of these strayed accidents can be increased especially by the careless driver. In this paper, the recognition system of lane and vehicle headway direction is developed to detect the possible collision and warn the driver to prevent the fatal accident. In the system development, Kalman filtering is used first to extract the border of driving lane from the video images supplied by the CCD camera attached to the vehicle and the driving lane detection is completed with regression analysis. Next, the vehicle headway direction is recognized by using neural network scheme with the extracted parameters of the detected driving lane feature. The practical experiments for the developed system are also carried out in the real traffic road of Seoul city area and the results show us the more than 90% accuracy in recognizing the driving lane and vehicle headway direction.

Real-Time Vehicle License Plate Detection Based on Background Subtraction and Cascade of Boosted Classifiers

  • Sarker, Md. Mostafa Kamal;Song, Moon Kyou
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.39C no.10
    • /
    • pp.909-919
    • /
    • 2014
  • License plate (LP) detection is the most imperative part of an automatic LP recognition (LPR) system. Typical LPR contains two steps, namely LP detection (LPD) and character recognition. In this paper, we propose an efficient Vehicle-to-LP detection framework which combines with an adaptive GMM (Gaussian Mixture Model) and a cascade of boosted classifiers to make a faster vehicle LP detector. To develop a background model by using a GMM is possible in the circumstance of a fixed camera and extracts the motions using background subtraction. Firstly, an adaptive GMM is used to find the region of interest (ROI) on which motion detectors are running to detect the vehicle area as blobs ROIs. Secondly, a cascade of boosted classifiers is executed on the blobs ROIs to detect a LP. The experimental results on our test video with the resolution of $720{\times}576$ show that the LPD rate of the proposed system is 99.14% and the average computational time is approximately 42ms.

Vision based Traffic Light Detection and Recognition Methods for Daytime LED Traffic Light (비전 기반 주간 LED 교통 신호등 인식 및 신호등 패턴 판단에 관한 연구)

  • Kim, Hyun-Koo;Park, Ju H.;Jung, Ho-Youl
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.9 no.3
    • /
    • pp.145-150
    • /
    • 2014
  • This paper presents an effective vision based method for LED traffic light detection at the daytime. First, the proposed method calculates horizontal coordinates to set region of interest (ROI) on input sequence images. Second, the proposed uses color segmentation method to extract region of green and red traffic light. Next, to classify traffic light and another noise, shape filter and haar-like feature value are used. Finally, temporal delay filter with weight is applied to remove blinking effect of LED traffic light, and state and weight of traffic light detection are used to classify types of traffic light. For simulations, the proposed method is implemented through Intel Core CPU with 2.80 GHz and 4 GB RAM, and tested on the urban and rural road video. Average detection rate of traffic light is 94.50 % and average recognition rate of traffic type is 90.24 %. Average computing time of the proposed method is 11 ms.

A Collaborative Video Annotation and Browsing System using Linked Data (링크드 데이터를 이용한 협업적 비디오 어노테이션 및 브라우징 시스템)

  • Lee, Yeon-Ho;Oh, Kyeong-Jin;Sean, Vi-Sal;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.3
    • /
    • pp.203-219
    • /
    • 2011
  • Previously common users just want to watch the video contents without any specific requirements or purposes. However, in today's life while watching video user attempts to know and discover more about things that appear on the video. Therefore, the requirements for finding multimedia or browsing information of objects that users want, are spreading with the increasing use of multimedia such as videos which are not only available on the internet-capable devices such as computers but also on smart TV and smart phone. In order to meet the users. requirements, labor-intensive annotation of objects in video contents is inevitable. For this reason, many researchers have actively studied about methods of annotating the object that appear on the video. In keyword-based annotation related information of the object that appeared on the video content is immediately added and annotation data including all related information about the object must be individually managed. Users will have to directly input all related information to the object. Consequently, when a user browses for information that related to the object, user can only find and get limited resources that solely exists in annotated data. Also, in order to place annotation for objects user's huge workload is required. To cope with reducing user's workload and to minimize the work involved in annotation, in existing object-based annotation automatic annotation is being attempted using computer vision techniques like object detection, recognition and tracking. By using such computer vision techniques a wide variety of objects that appears on the video content must be all detected and recognized. But until now it is still a problem facing some difficulties which have to deal with automated annotation. To overcome these difficulties, we propose a system which consists of two modules. The first module is the annotation module that enables many annotators to collaboratively annotate the objects in the video content in order to access the semantic data using Linked Data. Annotation data managed by annotation server is represented using ontology so that the information can easily be shared and extended. Since annotation data does not include all the relevant information of the object, existing objects in Linked Data and objects that appear in the video content simply connect with each other to get all the related information of the object. In other words, annotation data which contains only URI and metadata like position, time and size are stored on the annotation sever. So when user needs other related information about the object, all of that information is retrieved from Linked Data through its relevant URI. The second module enables viewers to browse interesting information about the object using annotation data which is collaboratively generated by many users while watching video. With this system, through simple user interaction the query is automatically generated and all the related information is retrieved from Linked Data and finally all the additional information of the object is offered to the user. With this study, in the future of Semantic Web environment our proposed system is expected to establish a better video content service environment by offering users relevant information about the objects that appear on the screen of any internet-capable devices such as PC, smart TV or smart phone.

Development of the Dripping Speed Measurement System of Medical Liquid using Heuristic (휴리스틱을 이용한 의료 수액 낙하 속도 측정 시스템 개발)

  • Kim, Jung-Sook;Jeong, Junho
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.24 no.5
    • /
    • pp.542-547
    • /
    • 2014
  • This paper describes the medical and IT convergence system using a smart phone and a heuristic method for the measurement of the dripping speed of the liquid in a drip chamber, which can estimate the remaining time using pattern recognition and difference image from video frame information based on Android technology. The video frames were first made using a smartphone camera and we calculated the difference image between the n image and the (n-1) image and then changed into binary images using the threshold value. At this point, it is very important to find an optimal threshold value using heuristic method to recognize the dripping of the liquids. In addition, the user can adjust the dripping speed according to the doctor's prescription, exactly like watching the progress bar of a mobile application. The experiment results show that our method using video processing technique accurately measures the dripping speed for a wide range of speeds that are sufficient for ordinary practice.