• Title/Summary/Keyword: Frame Camera

Search Result 613, Processing Time 0.026 seconds

Back-Propagation Neural Network Based Face Detection and Pose Estimation (오류-역전파 신경망 기반의 얼굴 검출 및 포즈 추정)

  • Lee, Jae-Hoon;Jun, In-Ja;Lee, Jung-Hoon;Rhee, Phill-Kyu
    • The KIPS Transactions:PartB
    • /
    • v.9B no.6
    • /
    • pp.853-862
    • /
    • 2002
  • Face Detection can be defined as follows : Given a digitalized arbitrary or image sequence, the goal of face detection is to determine whether or not there is any human face in the image, and if present, return its location, direction, size, and so on. This technique is based on many applications such face recognition facial expression, head gesture and so on, and is one of important qualify factors. But face in an given image is considerably difficult because facial expression, pose, facial size, light conditions and so on change the overall appearance of faces, thereby making it difficult to detect them rapidly and exactly. Therefore, this paper proposes fast and exact face detection which overcomes some restrictions by using neural network. The proposed system can be face detection irrelevant to facial expression, background and pose rapidily. For this. face detection is performed by neural network and detection response time is shortened by reducing search region and decreasing calculation time of neural network. Reduced search region is accomplished by using skin color segment and frame difference. And neural network calculation time is decreased by reducing input vector sire of neural network. Principle Component Analysis (PCA) can reduce the dimension of data. Also, pose estimates in extracted facial image and eye region is located. This result enables to us more informations about face. The experiment measured success rate and process time using the Squared Mahalanobis distance. Both of still images and sequence images was experimented and in case of skin color segment, the result shows different success rate whether or not camera setting. Pose estimation experiments was carried out under same conditions and existence or nonexistence glasses shows different result in eye region detection. The experiment results show satisfactory detection rate and process time for real time system.

Fast Natural Feature Tracking Using Optical Flow (광류를 사용한 빠른 자연특징 추적)

  • Bae, Byung-Jo;Park, Jong-Seung
    • The KIPS Transactions:PartB
    • /
    • v.17B no.5
    • /
    • pp.345-354
    • /
    • 2010
  • Visual tracking techniques for Augmented Reality are classified as either a marker tracking approach or a natural feature tracking approach. Marker-based tracking algorithms can be efficiently implemented sufficient to work in real-time on mobile devices. On the other hand, natural feature tracking methods require a lot of computationally expensive procedures. Most previous natural feature tracking methods include heavy feature extraction and pattern matching procedures for each of the input image frame. It is difficult to implement real-time augmented reality applications including the capability of natural feature tracking on low performance devices. The required computational time cost is also in proportion to the number of patterns to be matched. To speed up the natural feature tracking process, we propose a novel fast tracking method based on optical flow. We implemented the proposed method on mobile devices to run in real-time and be appropriately used with mobile augmented reality applications. Moreover, during tracking, we keep up the total number of feature points by inserting new feature points proportional to the number of vanished feature points. Experimental results showed that the proposed method reduces the computational cost and also stabilizes the camera pose estimation results.

A Brazing Defect Detection Using an Ultrasonic Infrared Imaging Inspection (초음파 열 영상 검사를 이용한 브레이징 접합 결함 검출)

  • Cho, Jai-Wan;Choi, Young-Soo;Jung, Seung-Ho;Jung, Hyun-Kyu
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.27 no.5
    • /
    • pp.426-431
    • /
    • 2007
  • When a high-energy ultrasound propagates through a solid body that contains a crack or a delamination, the two faces of the defect do not ordinarily vibrate in unison, and dissipative phenomena such as friction, rubbing and clapping between the faces will convert some of the vibrational energy to heat. By combining this heating effect with infrared imaging, one can detect a subsurface defect in material in real time. In this paper a realtime detection of the brazing defect of thin Inconel plates using the UIR (ultrasonic infrared imaging) technology is described. A low frequency (23 kHz) ultrasonic transducer was used to infuse the welded Inconel plates with a short pulse of sound for 280 ms. The ultrasonic source has a maximum power of 2 kW. The surface temperature of the area under inspection is imaged by an infrared camera that is coupled to a fast frame grabber in a computer. The hot spots, which are a small area around the bound between the two faces of the Inconel plates near the defective brazing point and heated up highly, are observed. And the weak thermal signal is observed at the defect position of brazed plate also. Using the image processing technology such as background subtraction average and image enhancement using histogram equalization, the position of defective brazing regions in the thin Inconel plates can be located certainly.

Vehicle Detection and Tracking using Billboard Sweep Stereo Matching Algorithm (빌보드 스윕 스테레오 시차정합 알고리즘을 이용한 차량 검출 및 추적)

  • Park, Min Woo;Won, Kwang Hee;Jung, Soon Ki
    • Journal of Korea Multimedia Society
    • /
    • v.16 no.6
    • /
    • pp.764-781
    • /
    • 2013
  • In this paper, we propose a highly precise vehicle detection method with low false alarm using billboard sweep stereo matching and multi-stage hypothesis generation. First, we capture stereo images from cameras established in front of the vehicle and obtain the disparity map in which the regions of ground plane or background are removed using billboard sweep stereo matching algorithm. And then, we perform the vehicle detection and tracking on the labeled disparity map. The vehicle detection and tracking consists of three steps. In the learning step, the SVM(support vector machine) classifier is obtained using the features extracted from the gabor filter. The second step is the vehicle detection which performs the sobel edge detection in the image of the left camera and extracts candidates of the vehicle using edge image and billboard sweep stereo disparity map. The final step is the vehicle tracking using template matching in the next frame. Removal process of the tracking regions improves the system performance in the candidate region of the vehicle on the succeeding frames.

Raising Visual Experience of Soccer Video for Mobile Viewers (이동형 단말기 사용자를 위한 축구경기 비디오의 시청경험 향상 방법)

  • Ahn, Il-Koo;Ko, Jae-Seung;Kim, Won-Jun;Kim, Chang-Ick
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.13 no.3
    • /
    • pp.165-178
    • /
    • 2007
  • The recent progress in multimedia signal processing and transmission technologies has contributed to the extensive use of multimedia devices to watch sports games with small LCD panel. However, the most of video sequences are captured for normal viewing on standard TV or HDTV, for cost reasons, merely resized and delivered without additional editing. This may give the small-display-viewers uncomfortable experiences in understanding what is happening in a scene. For instance, in a soccer video sequence taken by a long-shot camera techniques, the tiny objects (e.g., soccer ball and players) may not be clearly viewed on the small LCD panel. Moreover, it is also difficult to recognize the contents of the scorebox which contains the elapsed time and scores. This renuires intelligent display technique to provide small-display-viewers with better experience. To this end, one of the key technologies is to determine region of interest (ROI) and display the magnified ROI on the screen, where ROI is a part of the scene that viewers pay more attention to than other regions. Examples include a region surrounding a ball in long-shot and a scorebox located in the comer of each frame. In this paper, we propose a scheme for raising viewing experiences of multimedia mobile device users. Instead of taking generic approaches utilizing visually salient features for extraction of ROI in a scene, we take domain-specific approach to exploit unique attributes of the soccer video. The proposed scheme consists of two modules: ROI determination and scorebox extraction. The experimental results show that the proposed scheme offers useful tools for intelligent video display on multimedia mobile devices.

Development of a CNN-based Cross Point Detection Algorithm for an Air Duct Cleaning Robot (CNN 기반 공조 덕트 청소 로봇의 교차점 검출 알고리듬 개발)

  • Yi, Sarang;Noh, Eunsol;Hong, Seokmoo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.8
    • /
    • pp.1-8
    • /
    • 2020
  • Air ducts installed for ventilation inside buildings accumulate contaminants during their service life. Robots are installed to clean the air duct at low cost, but they are still not fully automated and depend on manpower. In this study, an intersection detection algorithm for autonomous driving was applied to an air duct cleaning robot. Autonomous driving of the robot was achieved by calculating the distance and angle between the extracted point and the center point through the intersection detection algorithm from the camera image mounted on the robot. The training data consisted of CAD images of the duct interior as well as the cross-point coordinates and angles between the two boundary lines. The deep learning-based CNN model was applied as a detection algorithm. For training, the cross-point coordinates were obtained from CAD images. The accuracy was determined based on the differences in the actual and predicted areas and distances. A cleaning robot prototype was designed, consisting of a frame, a Raspberry Pi computer, a control unit and a drive unit. The algorithm was validated by video imagery of the robot in operation. The algorithm can be applied to vehicles operating in similar environments.

Real-time Hand Region Detection and Tracking using Depth Information (깊이정보를 이용한 실시간 손 영역 검출 및 추적)

  • Joo, SungIl;Weon, SunHee;Choi, HyungIl
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.3
    • /
    • pp.177-186
    • /
    • 2012
  • In this paper, we propose a real-time approach for detecting and tracking a hand region by analyzing depth images. We build a hand model in advance. The model has the shape information of a hand. The detecting process extracts out moving areas in an image, which are possibly caused by moving a hand in front of a camera. The moving areas can be identified by analyzing accumulated difference images and applying the region growing technique. The extracted moving areas are compared against a hand model to get justified as a hand region. The tracking process keeps the track of center points of hand regions of successive frames. For this purpose, it involves three steps. The first step is to determine a seed point that is the closest point to the center point of a previous frame. The second step is to perform region growing to form a candidate region of a hand. The third step is to determine the center point of a hand to be tracked. This point is searched by the mean-shift algorithm within a confined area whose size varies adaptively according to the depth information. To verify the effectiveness of our approach, we have evaluated the performance of our approach while changing the shape and position of a hand as well as the velocity of hand movement.

A Design on Face Recognition System Based on pRBFNNs by Obtaining Real Time Image (실시간 이미지 획득을 통한 pRBFNNs 기반 얼굴인식 시스템 설계)

  • Oh, Sung-Kwun;Seok, Jin-Wook;Kim, Ki-Sang;Kim, Hyun-Ki
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.12
    • /
    • pp.1150-1158
    • /
    • 2010
  • In this study, the Polynomial-based Radial Basis Function Neural Networks is proposed as one of the recognition part of overall face recognition system that consists of two parts such as the preprocessing part and recognition part. The design methodology and procedure of the proposed pRBFNNs are presented to obtain the solution to high-dimensional pattern recognition problem. First, in preprocessing part, we use a CCD camera to obtain a picture frame in real-time. By using histogram equalization method, we can partially enhance the distorted image influenced by natural as well as artificial illumination. We use an AdaBoost algorithm proposed by Viola and Jones, which is exploited for the detection of facial image area between face and non-facial image area. As the feature extraction algorithm, PCA method is used. In this study, the PCA method, which is a feature extraction algorithm, is used to carry out the dimension reduction of facial image area formed by high-dimensional information. Secondly, we use pRBFNNs to identify the ID by recognizing unique pattern of each person. The proposed pRBFNNs architecture consists of three functional modules such as the condition part, the conclusion part, and the inference part as fuzzy rules formed in 'If-then' format. In the condition part of fuzzy rules, input space is partitioned with Fuzzy C-Means clustering. In the conclusion part of rules, the connection weight of pRBFNNs is represented as three kinds of polynomials such as constant, linear, and quadratic. Coefficients of connection weight identified with back-propagation using gradient descent method. The output of pRBFNNs model is obtained by fuzzy inference method in the inference part of fuzzy rules. The essential design parameters (including learning rate, momentum coefficient and fuzzification coefficient) of the networks are optimized by means of the Particle Swarm Optimization. The proposed pRBFNNs are applied to real-time face recognition system and then demonstrated from the viewpoint of output performance and recognition rate.

Study on Close-Up Shots in Film (2015) (영화 <사도>(2015)의 클로즈업 쇼트 연구)

  • Lee, A-Young
    • The Journal of the Korea Contents Association
    • /
    • v.16 no.7
    • /
    • pp.609-621
    • /
    • 2016
  • A close-up shot, capturing all the fine details of an actor's face by filling up the frame, expresses an actor's performance more realistically than a stage where a living actor performs on. This is because a close-up generates an impact with its specific images and meanings seizing the attention of the audience and conjures up a psychological effect as if an actor's face is directly communicating with the audience at a minimum distance. Therefore, this study analyzed the film to examine the photographic effect and acting effect of close-up shots and recommend the need for acting training thereof. The film was selected since Song Kang-ho, to add more realism to his character acted going back and forth 20 years of age with a special makeup on face, his facial expressions, gestures, props in close-up successfully helped deliver the actor's performance by revealing the character's personality and emotions of the film, and generated an array of linguistic, visual and emotional meanings which are the key to film acting. This study is expected to contribute to helping actors to learn about the effect of close-ups and the key to film acting and find effective ways to express themselves in front of the camera.

A Fast and Accurate Face Detection and Tracking Method by using Depth Information (깊이정보를 이용한 고속 고정밀 얼굴검출 및 추적 방법)

  • Bae, Yun-Jin;Choi, Hyun-Jun;Seo, Young-Ho;Kim, Dong-Wook
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.7A
    • /
    • pp.586-599
    • /
    • 2012
  • This paper proposes a fast face detection and tracking method which uses depth images as well as RGB images. It consists of the face detection procedure and the face tracking procedure. The face detection method basically uses an existing method, Adaboost, but it reduces the size of the search area by using the depth image. The proposed face tracking method uses a template matching technique and incorporates an early-termination scheme to reduce the execution time further. The results from implementing and experimenting the proposed methods showed that the proposed face detection method takes only about 39% of the execution time of the existing method. The proposed tracking method takes only 2.48ms per frame with $640{\times}480$ resolution. For the exactness, the proposed detection method showed a little lower in detection ratio but in the error ratio, which is for the cases when a detected one as a face is not really a face, the proposed method showed only about 38% of that of the previous method. The proposed face tracking method turned out to have a trade-off relationship between the execution time and the exactness. In all the cases except a special one, the tracking error ratio is as low as about 1%. Therefore, we expect the proposed face detection and tracking methods can be used individually or in combined for many applications that need fast execution and exact detection or tracking.