• Title/Summary/Keyword: Video recognition

Search Result 683, Processing Time 0.022 seconds

A Study on Rotational Alignment Algorithm for Improving Character Recognition (문자 인식 향상을 위한 회전 정렬 알고리즘에 관한 연구)

  • Jin, Go-Whan
    • Journal of the Korea Convergence Society
    • /
    • v.10 no.11
    • /
    • pp.79-84
    • /
    • 2019
  • Video image based technology is being used in various fields with continuous development. The demand for vision system technology that analyzes and discriminates image objects acquired through cameras is rapidly increasing. Image processing is one of the core technologies of vision systems, and is used for defect inspection in the semiconductor manufacturing field, object recognition inspection such as the number of tire surfaces and symbols. In addition, research into license plate recognition is ongoing, and it is necessary to recognize objects quickly and accurately. In this paper, propose a recognition model through the rotational alignment of objects after checking the angle value of the tilt of the object in the input video image for the recognition of inclined objects such as numbers or symbols marked on the surface. The proposed model can perform object recognition of the rotationally sorted image after extracting the object region and calculating the angle of the object based on the contour algorithm. The proposed model extracts the object region based on the contour algorithm, calculates the angle of the object, and then performs object recognition on the rotationally aligned image. In future research, it is necessary to study template matching through machine learning.

Context-aware Video Surveillance System

  • An, Tae-Ki;Kim, Moon-Hyun
    • Journal of Electrical Engineering and Technology
    • /
    • v.7 no.1
    • /
    • pp.115-123
    • /
    • 2012
  • A video analysis system used to detect events in video streams generally has several processes, including object detection, object trajectories analysis, and recognition of the trajectories by comparison with an a priori trained model. However, these processes do not work well in a complex environment that has many occlusions, mirror effects, and/or shadow effects. We propose a new approach to a context-aware video surveillance system to detect predefined contexts in video streams. The proposed system consists of two modules: a feature extractor and a context recognizer. The feature extractor calculates the moving energy that represents the amount of moving objects in a video stream and the stationary energy that represents the amount of still objects in a video stream. We represent situations and events as motion changes and stationary energy in video streams. The context recognizer determines whether predefined contexts are included in video streams using the extracted moving and stationary energies from a feature extractor. To train each context model and recognize predefined contexts in video streams, we propose and use a new ensemble classifier based on the AdaBoost algorithm, DAdaBoost, which is one of the most famous ensemble classifier algorithms. Our proposed approach is expected to be a robust method in more complex environments that have a mirror effect and/or a shadow effect.

Comparison of satisfaction, interest, and experience awareness of 360° virtual reality video and first-person video in non-face-to-face practical lectures in medical emergency departments (응급구조학과 비대면 실습 강의에서 360° 가상현실 영상과 1인칭 시점 영상의 만족도, 흥미도, 경험인식 비교)

  • Lee, Hyo-Ju;Shin, Sang-Yol;Jung, Eun-Kyung
    • The Korean Journal of Emergency Medical Services
    • /
    • v.24 no.3
    • /
    • pp.55-63
    • /
    • 2020
  • Purpose: This study aimed to establish effective training strategies and methods by comparing the effects of 360° virtual reality video and first-person video in non-face-to-face practical lectures. Methods: This crossover study, implemented May 18-31, 2020, included 27 participants. We compared 360° virtual reality video and first-person video. SPSS version 25.0 was used for statistical analysis. Results: The 360° virtual reality video had a higher score of experience recognition (p=.039), vividness (p=.045), presence (p=.000), fantasy factor (p=.000) than the first-person video, but no significant difference was indicated for satisfaction (p=.348) or interest (p=.441). Conclusion: 360° virtual reality video and first-person video can be used as training alternatives to achieve the standard educational objectives in non-face-to-face practical lectures.

Viewer's Affective Feedback for Video Summarization

  • Dammak, Majdi;Wali, Ali;Alimi, Adel M.
    • Journal of Information Processing Systems
    • /
    • v.11 no.1
    • /
    • pp.76-94
    • /
    • 2015
  • For different reasons, many viewers like to watch a summary of films without having to waste their time. Traditionally, video film was analyzed manually to provide a summary of it, but this costs an important amount of work time. Therefore, it has become urgent to propose a tool for the automatic video summarization job. The automatic video summarization aims at extracting all of the important moments in which viewers might be interested. All summarization criteria can differ from one video to another. This paper presents how the emotional dimensions issued from real viewers can be used as an important input for computing which part is the most interesting in the total time of a film. Our results, which are based on lab experiments that were carried out, are significant and promising.

Low-Light Invariant Video Enhancement Scheme Using Zero Reference Deep Curve Estimation (Zero Deep Curve 추정방식을 이용한 저조도에 강인한 비디오 개선 방법)

  • Choi, Hyeong-Seok;Yang, Yoon Gi
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.8
    • /
    • pp.991-998
    • /
    • 2022
  • Recently, object recognition using image/video signals is rapidly spreading on autonomous driving and mobile phones. However, the actual input image/video signals are easily exposed to a poor illuminance environment. A recent researches for improving illumination enable to estimate and compensate the illumination parameters. In this study, we propose VE-DCE (video enhancement zero-reference deep curve estimation) to improve the illumination of low-light images. The proposed VE-DCE uses unsupervised learning-based zero-reference deep curve, which is one of the latest among learning based estimation techniques. Experimental results show that the proposed method can achieve the quality of low-light video as well as images compared to the previous method. In addition, it can reduce the computational complexity with respect to the existing method.

Human Activity Pattern Recognition Using Motion Information and Joints of Human Body (인체의 조인트와 움직임 정보를 이용한 인간의 행동패턴 인식)

  • Kwak, Nae-Joung;Song, Teuk-Seob
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.6
    • /
    • pp.1179-1186
    • /
    • 2012
  • In this paper, we propose an algorithm that recognizes human activity patterns using the human body's joints and the information of the joints. The proposed method extracts the object from inputted video, automatically extracts joints using the ratio of the human body, applies block-matching algorithm for each joint and gets the motion information of joints. The proposed method uses the joints to move, the directional vector of motions of joints, and the sign to represent the increase or decrease of x and y coordinates of joints as basic parameters for human recognition of activity. The proposed method was tested for 8 human activities of inputted video from a web camera and had the good result for the ration of recognition of the human activities.

Spatial-temporal Ensemble Method for Action Recognition (행동 인식을 위한 시공간 앙상블 기법)

  • Seo, Minseok;Lee, Sangwoo;Choi, Dong-Geol
    • The Journal of Korea Robotics Society
    • /
    • v.15 no.4
    • /
    • pp.385-391
    • /
    • 2020
  • As deep learning technology has been developed and applied to various fields, it is gradually changing from an existing single image based application to a video based application having a time base in order to recognize human behavior. However, unlike 2D CNN in a single image, 3D CNN in a video has a very high amount of computation and parameter increase due to the addition of a time axis, so improving accuracy in action recognition technology is more difficult than in a single image. To solve this problem, we investigate and analyze various techniques to improve performance in 3D CNN-based image recognition without additional training time and parameter increase. We propose a time base ensemble using the time axis that exists only in the videos and an ensemble in the input frame. We have achieved an accuracy improvement of up to 7.1% compared to the existing performance with a combination of techniques. It also revealed the trade-off relationship between computational and accuracy.

Silhouette-Edge-Based Descriptor for Human Action Representation and Recognition

  • Odoyo, Wilfred O.;Choi, Jae-Ho;Moon, In-Kyu;Cho, Beom-Joon
    • Journal of information and communication convergence engineering
    • /
    • v.11 no.2
    • /
    • pp.124-131
    • /
    • 2013
  • Extraction and representation of postures and/or gestures from human activities in videos have been a focus of research in this area of action recognition. With various applications cropping up from different fields, this paper seeks to improve the performance of these action recognition machines by proposing a shape-based silhouette-edge descriptor for the human body. Information entropy, a method to measure the randomness of a sequence of symbols, is used to aid the selection of vital key postures from video frames. Morphological operations are applied to extract and stack edges to uniquely represent different actions shape-wise. To classify an action from a new input video, a Hausdorff distance measure is applied between the gallery representations and the query images formed from the proposed procedure. The method is tested on known public databases for its validation. An effective method of human action annotation and description has been effectively achieved.

Multiple Object Tracking and Identification System Using CCTV and RFID (감시 카메라와 RFID를 활용한 다수 객체 추적 및 식별 시스템)

  • Kim, Jin-Ah;Moon, Nammee
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.6 no.2
    • /
    • pp.51-58
    • /
    • 2017
  • Because of safety and security, Surveillance camera market is growing. Accordingly, Study on video recognition and tracking is also actively in progress, but There is a limit to identify object by obtaining the information of object identified and tracked. Especially, It is more difficult to identify multiple objects in open space like shopping mall, airport and others utilized surveillance camera. Therefore, This paper proposed adding object identification function by using RFID to existing video-based object recognition and tracking system. Also, We tried to complement each other to solve the problem of video and RFID based. Thus, through the interaction of system modules We propose a solution to the problems of failing video-based object recognize and tracking and the problems that could be cased by the recognition error of RFID. The system designed to identify the object by classifying the identification of object in four steps so that the data reliability of the identified object can be maintained. To judge the efficiency of this system, this demonstrated by implementing the simulation program.

A Search Model Using Time Interval Variation to Identify Face Recognition Results

  • Choi, Yun-seok;Lee, Wan Yeon
    • International journal of advanced smart convergence
    • /
    • v.11 no.3
    • /
    • pp.64-71
    • /
    • 2022
  • Various types of attendance management systems are being introduced in a remote working environment and research on using face recognition is in progress. To ensure accurate worker's attendance, a face recognition-based attendance management system must analyze every frame of video, but face recognition is a heavy task, the number of the task should be minimized without affecting accuracy. In this paper, we proposed a search model using time interval variation to minimize the number of face recognition task of recorded videos for attendance management system. The proposed model performs face recognition by changing the interval of the frame identification time when there is no change in the attendance status for a certain period. When a change in the face recognition status occurs, it moves in the reverse direction and performs frame checks to more accurate attendance time checking. The implementation of proposed model performed at least 4.5 times faster than all frame identification and showed at least 97% accuracy.