• Title/Summary/Keyword: feature histogram

Search Result 377, Processing Time 0.029 seconds

Effective Point Dataset Removal for High-Speed 3D Scanning Processes (고속 3D 스캐닝 프로세스를 위한 효과적인 점데이터 제거)

  • Lim, Sukhyun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.11
    • /
    • pp.1660-1665
    • /
    • 2022
  • Recently, many industries are using three dimensional scanning technology. As the performance of the 3D scanner gradually improves, a sampling step to reduce a point data or a remove step to remove a part determined to be noise are generally performed in post processing. However, total point data by long time scanning cannot be processed at once in spite of performing such those additional processes. In general, a method using a multi threaded environment is widely used, but as the scanning process work time increases, the processing performance gradually decreases due to various environmental conditions and accumulated operations. This paper proposes a method to initially remove point data judged to be unnecessary by calculating accumulated fast point feature histogram values from coming point data of the 3D scanner in real time. The entire 3D scanning process can be reduced using this approach.

Object Detection and Classification Using Extended Descriptors for Video Surveillance Applications (비디오 감시 응용에서 확장된 기술자를 이용한 물체 검출과 분류)

  • Islam, Mohammad Khairul;Jahan, Farah;Min, Jae-Hong;Baek, Joong-Hwan
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.4
    • /
    • pp.12-20
    • /
    • 2011
  • In this paper, we propose an efficient object detection and classification algorithm for video surveillance applications. Previous researches mainly concentrated either on object detection or classification using particular type of feature e.g., Scale Invariant Feature Transform (SIFT) or Speeded Up Robust Feature (SURF) etc. In this paper we propose an algorithm that mutually performs object detection and classification. We combinedly use heterogeneous types of features such as texture and color distribution from local patches to increase object detection and classification rates. We perform object detection using spatial clustering on interest points, and use Bag of Words model and Naive Bayes classifier respectively for image representation and classification. Experimental results show that our combined feature is better than the individual local descriptor in object classification rate.

A Phase-related Feature Extraction Method for Robust Speaker Verification (열악한 환경에 강인한 화자인증을 위한 위상 기반 특징 추출 기법)

  • Kwon, Chul-Hong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.14 no.3
    • /
    • pp.613-620
    • /
    • 2010
  • Additive noise and channel distortion strongly degrade the performance of speaker verification systems, as it introduces distortion of the features of speech. This distortion causes a mismatch between the training and recognition conditions such that acoustic models trained with clean speech do not model noisy and channel distorted speech accurately. This paper presents a phase-related feature extraction method in order to improve the robustness of the speaker verification systems. The instantaneous frequency is computed from the phase of speech signals and features from the histogram of the instantaneous frequency are obtained. Experimental results show that the proposed technique offers significant improvements over the standard techniques in both clean and adverse testing environments.

Face Feature Extraction for Face Recognition (얼굴 인식을 위한 얼굴 특징점 추출)

  • Yang, Ryong;Chae, Duk-Jae;Lee, Sang-Bum
    • Journal of the Korea Computer Industry Society
    • /
    • v.3 no.12
    • /
    • pp.1765-1774
    • /
    • 2002
  • A face recognition is currently the field which many research have been processed actively. But many problems must be solved the previous problem. First, We must recognize the face of the object taking a location various lighting change and change of the camera into account. In this paper, we proposed that new method to fund feature within fast and correct computation time after scanning PC camera and ID card picture. It converted RGB color space to YUV. A face skin color extracts which equalize a histogram of Y ingredient without the luminance. After, the method use V' ingredient which transformes V ingredient of YUV and then find the face feature. The reult of the experiment shows getting correct input face image from ID Card picture and PC camera.

  • PDF

A study on Robust Feature Image for Texture Classification and Detection (텍스쳐 분류 및 검출을 위한 강인한 특징이미지에 관한 연구)

  • Kim, Young-Sub;Ahn, Jong-Young;Kim, Sang-Bum;Hur, Kang-In
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.10 no.5
    • /
    • pp.133-138
    • /
    • 2010
  • In this paper, we make up a feature image including spatial properties and statistical properties on image, and format covariance matrices using region variance magnitudes. By using it to texture classification, this paper puts a proposal for tough texture classification way to illumination, noise and rotation. Also we offer a way to minimalize performance time of texture classification using integral image expressing middle image for fast calculation of region sum. To estimate performance evaluation of proposed way, this paper use a Brodatz texture image, and so conduct a noise addition and histogram specification and create rotation image. And then we conduct an experiment and get better performance over 96%.

Content-Based Image Retrieval Using Visual Features and Fuzzy Integral (시각 특징과 퍼지 적분을 이용한 내용기반 영상 검색)

  • Song Young-Jun;Kim Nam;Kim Mi-Hye;Kim Dong-Woo
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.5
    • /
    • pp.20-28
    • /
    • 2006
  • This paper proposes visual-feature extraction for each band in wavelet domain with both spatial frequency features and multi resolution features, and the combination of visual features using fuzzy integral. In addition, it uses color feature expression method taking advantage of the frequency of the same color after color quantization for reducing quantization error, a disadvantage of the existing color histogram intersection method. Also, it is found that the final similarity can be represented in a linear combination of the respective factors(Homogram, color, energy) when each factor is independent one another. With respect to the combination patterns the fuzzy measurement is defined and the fuzzy integral is taken. Experiments are peformed on a database containing 1,000 color images. The proposed method gives better performance than the conventional method in both objective and subjective performance evaluation.

  • PDF

Object Tracking with Sparse Representation based on HOG and LBP Features

  • Boragule, Abhijeet;Yeo, JungYeon;Lee, GueeSang
    • International Journal of Contents
    • /
    • v.11 no.3
    • /
    • pp.47-53
    • /
    • 2015
  • Visual object tracking is a fundamental problem in the field of computer vision, as it needs a proper model to account for drastic appearance changes that are caused by shape, textural, and illumination variations. In this paper, we propose a feature-based visual-object-tracking method with a sparse representation. Generally, most appearance-based models use the gray-scale pixel values of the input image, but this might be insufficient for a description of the target object under a variety of conditions. To obtain the proper information regarding the target object, the following combination of features has been exploited as a corresponding representation: First, the features of the target templates are extracted by using the HOG (histogram of gradient) and LBPs (local binary patterns); secondly, a feature-based sparsity is attained by solving the minimization problems, whereby the target object is represented by the selection of the minimum reconstruction error. The strengths of both features are exploited to enhance the overall performance of the tracker; furthermore, the proposed method is integrated with the particle-filter framework and achieves a promising result in terms of challenging tracking videos.

Feature Extraction Method of 2D-DCT for Facial Expression Recognition (얼굴 표정인식을 위한 2D-DCT 특징추출 방법)

  • Kim, Dong-Ju;Lee, Sang-Heon;Sohn, Myoung-Kyu
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.3
    • /
    • pp.135-138
    • /
    • 2014
  • This paper devices a facial expression recognition method robust to overfitting using 2D-DCT and EHMM algorithm. In particular, this paper achieves enhanced recognition performance by setting up a large window size for 2D-DCT feature extraction and extracting the observation vectors of EHMM. The experimental results on the CK facial expression database and the JAFFE facial expression database showed that the facial expression recognition accuracy was improved according as window size is large. Also, the proposed method revealed the recognition accuracy of 87.79% and showed enhanced recognition performance ranging from 46.01% to 50.05% in comparison to previous approaches based on histogram feature, when CK database is employed for training and JAFFE database is used to test the recognition accuracy.

Enhanced CT-image for Covid-19 classification using ResNet 50

  • Lobna M. Abouelmagd;Manal soubhy Ali Elbelkasy
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.1
    • /
    • pp.119-126
    • /
    • 2024
  • Disease caused by the coronavirus (COVID-19) is sweeping the globe. There are numerous methods for identifying this disease using a chest imaging. Computerized Tomography (CT) chest scans are used in this study to detect COVID-19 disease using a pretrain Convolutional Neural Network (CNN) ResNet50. This model is based on image dataset taken from two hospitals and used to identify Covid-19 illnesses. The pre-train CNN (ResNet50) architecture was used for feature extraction, and then fully connected layers were used for classification, yielding 97%, 96%, 96%, 96% for accuracy, precision, recall, and F1-score, respectively. When combining the feature extraction techniques with the Back Propagation Neural Network (BPNN), it produced accuracy, precision, recall, and F1-scores of 92.5%, 83%, 92%, and 87.3%. In our suggested approach, we use a preprocessing phase to improve accuracy. The image was enhanced using the Contrast Limited Adaptive Histogram Equalization (CLAHE) algorithm, which was followed by cropping the image before feature extraction with ResNet50. Finally, a fully connected layer was added for classification, with results of 99.1%, 98.7%, 99%, 98.8% in terms of accuracy, precision, recall, and F1-score.

A Study on Enhancing the Performance of Detecting Lip Feature Points for Facial Expression Recognition Based on AAM (AAM 기반 얼굴 표정 인식을 위한 입술 특징점 검출 성능 향상 연구)

  • Han, Eun-Jung;Kang, Byung-Jun;Park, Kang-Ryoung
    • The KIPS Transactions:PartB
    • /
    • v.16B no.4
    • /
    • pp.299-308
    • /
    • 2009
  • AAM(Active Appearance Model) is an algorithm to extract face feature points with statistical models of shape and texture information based on PCA(Principal Component Analysis). This method is widely used for face recognition, face modeling and expression recognition. However, the detection performance of AAM algorithm is sensitive to initial value and the AAM method has the problem that detection error is increased when an input image is quite different from training data. Especially, the algorithm shows high accuracy in case of closed lips but the detection error is increased in case of opened lips and deformed lips according to the facial expression of user. To solve these problems, we propose the improved AAM algorithm using lip feature points which is extracted based on a new lip detection algorithm. In this paper, we select a searching region based on the face feature points which are detected by AAM algorithm. And lip corner points are extracted by using Canny edge detection and histogram projection method in the selected searching region. Then, lip region is accurately detected by combining color and edge information of lip in the searching region which is adjusted based on the position of the detected lip corners. Based on that, the accuracy and processing speed of lip detection are improved. Experimental results showed that the RMS(Root Mean Square) error of the proposed method was reduced as much as 4.21 pixels compared to that only using AAM algorithm.