• Title/Summary/Keyword: illumination invariance

Search Result 10, Processing Time 0.023 seconds

Illumination Invariant Face Tracking on Smart Phones using Skin Locus based CAMSHIFT

  • Bui, Hoang Nam;Kim, SooHyung;Na, In Seop
    • Smart Media Journal
    • /
    • v.2 no.4
    • /
    • pp.9-19
    • /
    • 2013
  • This paper gives a review on three illumination issues of face tracking on smart phones: dark scenes, sudden lighting change and backlit effect. First, we propose a fast and robust face tracking method utilizing continuous adaptive mean shift algorithm (CAMSHIFT) and CbCr skin locus. Initially, the skin locus obtained from training video data. After that, a modified CAMSHIFT version based on the skin locus is accordingly provided. Second, we suggest an enhancement method to increase the chance of detecting faces, an important initialization step for face tracking, under dark illumination. The proposed method works comparably with traditional CAMSHIFT or particle filter, and outperforms these methods when dealing with our public video data with the three illumination issues mentioned above.

  • PDF

Anomaly Sewing Pattern Detection for AIoT System using Deep Learning and Decision Tree

  • Nguyen Quoc Toan;Seongwon Cho
    • Smart Media Journal
    • /
    • v.13 no.2
    • /
    • pp.85-94
    • /
    • 2024
  • Artificial Intelligence of Things (AIoT), which combines AI and the Internet of Things (IoT), has recently gained popularity. Deep neural networks (DNNs) have achieved great success in many applications. Deploying complex AI models on embedded boards, nevertheless, may be challenging due to computational limitations or intelligent model complexity. This paper focuses on an AIoT-based system for smart sewing automation using edge devices. Our technique included developing a detection model and a decision tree for a sufficient testing scenario. YOLOv5 set the stage for our defective sewing stitches detection model, to detect anomalies and classify the sewing patterns. According to the experimental testing, the proposed approach achieved a perfect score with accuracy and F1score of 1.0, False Positive Rate (FPR), False Negative Rate (FNR) of 0, and a speed of 0.07 seconds with file size 2.43MB.

An Illumination Invariant Traffic Sign Recognition in the Driving Environment for Intelligence Vehicles (지능형 자동차를 위한 조명 변화에 강인한 도로표지판 검출 및 인식)

  • Lee, Taewoo;Lim, Kwangyong;Bae, Guntae;Byun, Hyeran;Choi, Yeongwoo
    • Journal of KIISE
    • /
    • v.42 no.2
    • /
    • pp.203-212
    • /
    • 2015
  • This paper proposes a traffic sign recognition method in real road environments. The video stream in driving environments has two different characteristics compared to a general object video stream. First, the number of traffic sign types is limited and their shapes are mostly simple. Second, the camera cannot take clear pictures in the road scenes since there are many illumination changes and weather conditions are continuously changing. In this paper, we improve a modified census transform(MCT) to extract features effectively from the road scenes that have many illumination changes. The extracted features are collected by histograms and are transformed by the dense descriptors into very high dimensional vectors. Then, the high dimensional descriptors are encoded into a low dimensional feature vector by Fisher-vector coding and Gaussian Mixture Model. The proposed method shows illumination invariant detection and recognition, and the performance is sufficient to detect and recognize traffic signs in real-time with high accuracy.

Robust-to-rotation Iris Recognition Using Local Gradient Orientation Histogram (국부적 그래디언트 방향 히스토그램을 이용한 회전에 강인한 홍채 인식)

  • Choi, Chang-Soo;Jun, Byoung-Min
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.3C
    • /
    • pp.268-273
    • /
    • 2009
  • Iris recognition is a biometric technology which can identify a person using the iris pattern. It is important for the iris recognition system to extract the feature which is invariant to changes in iris patterns. Those changes can be occurred by the influence of lights, changes in the size of the pupil, and head tilting. In this paper, we propose a novel method based on local gradient orientation histogram which is robust to variations in illumination and rotations of iris patterns. The proposed method enables high-speed feature extraction and feature comparison because it requires no additional processing to obtain the rotation invariance, and shows comparable performance to the well-known previous methods.

Automatic Edge Detection Method for Mobile Robot Application (이동로봇을 위한 영상의 자동 엣지 검출 방법)

  • Kim Dongsu;Kweon Inso;Lee Wangheon
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.11 no.5
    • /
    • pp.423-428
    • /
    • 2005
  • This paper proposes a new edge detection method using a $3{\times}3$ ideal binary pattern and lookup table (LUT) for the mobile robot localization without any parameter adjustments. We take the mean of the pixels within the $3{\times}3$ block as a threshold by which the pixels are divided into two groups. The edge magnitude and orientation are calculated by taking the difference of average intensities of the two groups and by searching directional code in the LUT, respectively. And also the input image is not only partitioned into multiple groups according to their intensity similarities by the histogram, but also the threshold of each group is determined by fuzzy reasoning automatically. Finally, the edges are determined through non-maximum suppression using edge confidence measure and edge linking. Applying this edge detection method to the mobile robot localization using projective invariance of the cross ratio. we demonstrate the robustness of the proposed method to the illumination changes in a corridor environment.

Viewpoint Invariant Person Re-Identification for Global Multi-Object Tracking with Non-Overlapping Cameras

  • Gwak, Jeonghwan;Park, Geunpyo;Jeon, Moongu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.4
    • /
    • pp.2075-2092
    • /
    • 2017
  • Person re-identification is to match pedestrians observed from non-overlapping camera views. It has important applications in video surveillance such as person retrieval, person tracking, and activity analysis. However, it is a very challenging problem due to illumination, pose and viewpoint variations between non-overlapping camera views. In this work, we propose a viewpoint invariant method for matching pedestrian images using orientation of pedestrian. First, the proposed method divides a pedestrian image into patches and assigns angle to a patch using the orientation of the pedestrian under the assumption that a person body has the cylindrical shape. The difference between angles are then used to compute the similarity between patches. We applied the proposed method to real-time global multi-object tracking across multiple disjoint cameras with non-overlapping field of views. Re-identification algorithm makes global trajectories by connecting local trajectories obtained by different local trackers. The effectiveness of the viewpoint invariant method for person re-identification was validated on the VIPeR dataset. In addition, we demonstrated the effectiveness of the proposed approach for the inter-camera multiple object tracking on the MCT dataset with ground truth data for local tracking.

Vehicle License Plate Detection in Road Images (도로주행 영상에서의 차량 번호판 검출)

  • Lim, Kwangyong;Byun, Hyeran;Choi, Yeongwoo
    • Journal of KIISE
    • /
    • v.43 no.2
    • /
    • pp.186-195
    • /
    • 2016
  • This paper proposes a vehicle license plate detection method in real road environments using 8 bit-MCT features and a landmark-based Adaboost method. The proposed method allows identification of the potential license plate region, and generates a saliency map that presents the license plate's location probability based on the Adaboost classification score. The candidate regions whose scores are higher than the given threshold are chosen from the saliency map. Each candidate region is adjusted by the local image variance and verified by the SVM and the histograms of the 8bit-MCT features. The proposed method achieves a detection accuracy of 85% from various road images in Korea and Europe.

Baggage Recognition in Occluded Environment using Boosting Technique

  • Khanam, Tahmina;Deb, Kaushik
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.11
    • /
    • pp.5436-5458
    • /
    • 2017
  • Automatic Video Surveillance System (AVSS) has become important to computer vision researchers as crime has increased in the twenty-first century. As a new branch of AVSS, baggage detection has a wide area of security applications. Some of them are, detecting baggage in baggage restricted super shop, detecting unclaimed baggage in public space etc. However, in this paper, a detection & classification framework of baggage is proposed. Initially, background subtraction is performed instead of sliding window approach to speed up the system and HSI model is used to deal with different illumination conditions. Then, a model is introduced to overcome shadow effect. Then, occlusion of objects is detected using proposed mirroring algorithm to track individual objects. Extraction of rotational signal descriptor (SP-RSD-HOG) with support plane from Region of Interest (ROI) add rotation invariance nature in HOG. Finally, dynamic human body parameter setting approach enables the system to detect & classify single or multiple pieces of carried baggage even if some portions of human are absent. In baggage detection, a strong classifier is generated by boosting similarity measure based multi layer Support Vector Machine (SVM)s into HOG based SVM. This boosting technique has been used to deal with various texture patterns of baggage. Experimental results have discovered the system satisfactorily accurate and faster comparative to other alternatives.

Comparisons of Color Spaces for Shadow Elimination (그림자 제거를 위한 색상 공간의 비교)

  • Lee, Gwang-Gook;Uzair, Muhammad;Yoon, Ja-Young;Kim, Jae-Jun;Kim, Whoi-Yul
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.5
    • /
    • pp.610-622
    • /
    • 2008
  • Moving object segmentation is an essential technique for various video surveillance applications. The result of moving object segmentation often contains shadow regions caused by the color difference of shadow pixels. Hence, moving object segmentation is usually followed by a shadow elimination process to remove the false detection results. The common assumption adopted in previous works is that, under the illumination variation, the value of chromaticity components are preserved while the value of intensity component is changed. Hence, color transforms which separates luminance component and chromaticity component are usually utilized to remove shadow pixels. In this paper, various color spaces (YCbCr, HSI, normalized rgb, Yxy, Lab, c1c2c3) are examined to find the most appropriate color space for shadow elimination. So far, there have been some research efforts to compare the influence of various color spaces for shadow elimination. However, previous efforts are somewhat insufficient to compare the color distortions under illumination change in diverse color spaces, since they used a specific shadow elimination scheme or different thresholds for different color spaces. In this paper, to relieve the limitations of previous works, (1) the amount of gradients in shadow boundaries drawn to uniform colored regions are examined only for chromaticity components to compare the color distortion under illumination change and (2) the accuracy of background subtraction are analyzed via RoC curves to compare different color spaces without the problem of threshold level selection. Through experiments on real video sequences, YCbCr and normalized rgb color spaces showed good results for shadow elimination among various color spaces used for the experiments.

  • PDF

Recognition of Events by Human Motion for Context-aware Computing (상황인식 컴퓨팅을 위한 사람 움직임 이벤트 인식)

  • Cui, Yao-Huan;Shin, Seong-Yoon;Lee, Chang-Woo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.4
    • /
    • pp.47-57
    • /
    • 2009
  • Event detection and recognition is an active and challenging topic recent in Computer Vision. This paper describes a new method for recognizing events caused by human motion from video sequences in an office environment. The proposed approach analyzes human motions using Motion History Image (MHI) sequences, and is invariant to body shapes. types or colors of clothes and positions of target objects. The proposed method has two advantages; one is thant the proposed method is less sensitive to illumination changes comparing with the method using color information of objects of interest, and the other is scale invariance comparing with the method using a prior knowledge like appearances or shapes of objects of interest. Combined with edge detection, geometrical characteristics of the human shape in the MHI sequences are considered as the features. An advantage of the proposed method is that the event detection framework is easy to extend by inserting the descriptions of events. In addition, the proposed method is the core technology for event detection systems based on context-aware computing as well as surveillance systems based on computer vision techniques.