• Title/Summary/Keyword: 분류기 알고리즘

Search Result 599, Processing Time 0.025 seconds

Face Tracking and Recognition in Video with PCA-based Pose-Classification and (2D)2PCA recognition algorithm (비디오속의 얼굴추적 및 PCA기반 얼굴포즈분류와 (2D)2PCA를 이용한 얼굴인식)

  • Kim, Jin-Yul;Kim, Yong-Seok
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.23 no.5
    • /
    • pp.423-430
    • /
    • 2013
  • In typical face recognition systems, the frontal view of face is preferred to reduce the complexity of the recognition. Thus individuals may be required to stare into the camera, or the camera should be located so that the frontal images are acquired easily. However these constraints severely restrict the adoption of face recognition to wide applications. To alleviate this problem, in this paper, we address the problem of tracking and recognizing faces in video captured with no environmental control. The face tracker extracts a sequence of the angle/size normalized face images using IVT (Incremental Visual Tracking) algorithm that is known to be robust to changes in appearance. Since no constraints have been imposed between the face direction and the video camera, there will be various poses in face images. Thus the pose is identified using a PCA (Principal Component Analysis)-based pose classifier, and only the pose-matched face images are used to identify person against the pre-built face DB with 5-poses. For face recognition, PCA, (2D)PCA, and $(2D)^2PCA$ algorithms have been tested to compute the recognition rate and the execution time.

Welfare Interface using Multiple Facial Features Tracking (다중 얼굴 특징 추적을 이용한 복지형 인터페이스)

  • Ju, Jin-Sun;Shin, Yun-Hee;Kim, Eun-Yi
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.1
    • /
    • pp.75-83
    • /
    • 2008
  • We propose a welfare interface using multiple fecial features tracking, which can efficiently implement various mouse operations. The proposed system consist of five modules: face detection, eye detection, mouth detection, facial feature tracking, and mouse control. The facial region is first obtained using skin-color model and connected-component analysis(CCs). Thereafter the eye regions are localized using neutral network(NN)-based texture classifier that discriminates the facial region into eye class and non-eye class, and then mouth region is localized using edge detector. Once eye and mouth regions are localized they are continuously and correctly tracking by mean-shift algorithm and template matching, respectively. Based on the tracking results, mouse operations such as movement or click are implemented. To assess the validity of the proposed system, it was applied to the interface system for web browser and was tested on a group of 25 users. The results show that our system have the accuracy of 99% and process more than 21 frame/sec on PC for the $320{\times}240$ size input image, as such it can supply a user-friendly and convenient access to a computer in real-time operation.

Feature Extraction Algorithm for Distant Unmmaned Aerial Vehicle Detection (원거리 무인기 신호 식별을 위한 특징추출 알고리즘)

  • Kim, Juho;Lee, Kibae;Bae, Jinho;Lee, Chong Hyun
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.53 no.3
    • /
    • pp.114-123
    • /
    • 2016
  • The effective feature extraction method for unmanned aerial vehicle (UAV) detection is proposed and verified in this paper. The UAV engine sound is harmonic complex tone whose frequency ratio is integer and its variation is continuous in time. Using these characteristic, we propose the feature vector composed of a mean and standard deviation of difference value between fundamental frequency with 1st overtone as well as mean variation of their frequency. It was revealed by simulation that the suggested feature vector has excellent discrimination in target signal identification from various interfering signals including frequency variation with time. By comparing Fisher scores, three features based on frequency show outstanding discrimination of measured UAV signals with low signal to noise ratio (SNR). Detection performance with simulated interference signal is compared by MFCC by using ELM classifier and the suggested feature vector shows 37.6% of performance improvement As the SNR increases with time, the proposed feature can detect the target signal ahead of MFCC that needs 4.5 dB higher signal power to detect the target.

Intelligent I/O Subsystem for Future A/V Embedded Device (멀티미디어 기기를 위한 지능형 입출력 서브시스템)

  • Jang, Hyung-Kyu;Won, Yoo-Jip;Ryu, Jae-Min;Shim, Jun-Seok;Boldyrev, Serguei
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.33 no.1_2
    • /
    • pp.79-91
    • /
    • 2006
  • The intelligent disk can improve the overall performance of the I/O subsystem by processing the I/O operations in the disk side. At present time, however, realizing the intelligent disk seems to be impossible because of the limitation of the I/O subsystem and the lack of the backward compatibility with the traditional I/O interface scheme. In this paper, we proposed new model for the intelligent disk that dynamically optimizes the I/O subsystem using the information that is only related to the physical sector. In this way, the proposed model does not break the compatibility with the traditional I/O interface scheme. For these works, the boosting algorithm that upgrades a weak learner by repeating teaming is used. If the last learner classifies a recent I/O workload as the multimedia workload, the disk reads more sectors. Also, by embedding this functionality as a firmware or a embedded OS within the disk, the overall I/O subsystem can be operated more efficiently without the additional workload.

Steganalysis Using Histogram Characteristic and Statistical Moments of Wavelet Subbands (웨이블릿 부대역의 히스토그램 특성과 통계적 모멘트를 이용한 스테그분석)

  • Hyun, Seung-Hwa;Park, Tae-Hee;Kim, Young-In;Kim, Yoo-Shin;Eom, Il-Kyu
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.47 no.6
    • /
    • pp.57-65
    • /
    • 2010
  • In this paper, we present a universal steganalysis scheme. The proposed method extract features of two types. First feature set is extracted from histogram characteristic of the wavelet subbands. Second feature set is determined by statistical moments of wavelet characteristic functions. 3-level wavelet decomposition is performed for stego image and cover image using the Haar wavelet basis. We extract one features from 9 high frequency subbands of 12 subbands. The number of second features is 39. We use total 48 features for steganalysis. Multi layer perceptron(MLP) is applied as classifier to distinguish between cover images and stego images. To evaluate the proposed steganalysis method, we use the CorelDraw image database. We test the performance of our proposed steganalysis method over LSB method, spread spectrum data hiding method, blind spread spectrum data hiding method and F5 data hiding method. The proposed method outperforms the previous methods in sensitivity, specificity, error rate and area under ROC curve, etc.

Level 3 Type Land Use Land Cover (LULC) Characteristics Based on Phenological Phases of North Korea (생물계절 상 분석을 통한 Level 3 type 북한 토지피복 특성)

  • Yu, Jae-Shim;Park, Chong-Hwa;Lee, Seung-Ho
    • Korean Journal of Remote Sensing
    • /
    • v.27 no.4
    • /
    • pp.457-466
    • /
    • 2011
  • The objectives of this study are to produce level 3 type LULC map and analysis of phenological features of North Korea, ISODATA clustering of the 88scenes of MVC of MODIS NDVI in 2008 and 8scenes in 2009 was carried out. Analysis of phenological phases based mapping method was conducted, In level 2 type map, the confusion matrix was summarized and Kappa coefficient was calculated. Total of 27 typical habitat types that represent the dominant species or vegetation density that cover land surface of North Korea in 2008 were made. The total of 27 classes includes the 17 forest biotopes, 7 different croplands, 2 built up types and one water body. Dormancy phase of winter (${\sigma}^2$ = 0.348) and green up phase in spring (${\sigma}^2$ = 0.347) displays phenological dynamics when much vegetation growth changes take place. Overall accuracy is (851/955) 85.85% and Kappa coefficient is 0.84. Phenological phase based mapping method was possible to minimize classification error when analyzing the inaccessible land of North Korea.

Automatic Classification Technique of Offence Patterns using Neural Networks in Soccer Game (뉴럴네트워크를 이용한 축구경기 공격패턴 자동분류에 관한 연구)

  • Kim, Hyun-Sook;Yoon, Ho-Sub;Hwang, Chong-Sun;Yang, Young-Kyu
    • Annual Conference of KIPS
    • /
    • 2001.10a
    • /
    • pp.727-730
    • /
    • 2001
  • 멀티미디어 환경의 급속한 발전에 의해 영상처리 기술은 인간의 인체와 관련하여 얼굴인식, 제스처 인식에 관한 응용과 더불어 스포츠 관련분야로 깊숙히 정착하고 있다. 그러나 입력영상으로부터 움직이고 있는 선수들의 동작을 추출 및 추적하는 일은 컴퓨터비전 연구의 난 문제 중의 하나로 알려져 있다. 이러한 축구경기의 TV 중계에 있어서 하이라이트 장면의 자동추출(자동색인)은 그 경기의 가장 집약적인 표현이며, 축구경기 전체를 한 눈에 파악할 수 있도록 해주는 요약(summary)이자 intensive actions이고 경기의 진수이다. 따라서 축구경기와 같이 비교적 기 시간(대체로 1시간 30분) 동안 다수의 선수(양 팀 합해서 22명)들이 서로 복잡하게 뒤얽히면서 진행하는 경기의 하이라이트 장면을 효과적으로 포착하여 표현해 줄 수 있다면 TV를 통해서 경기를 관람하는 시청자들에게는 경기의 진행상황을 한 눈에 효과적으로 파악할 수 있게 해주어 흥미진진한 경기관람을 할 수 있게 해주고, 경기의 진행자들(감독, 코치, 선수 등)에게는 고차원적이고 과학적인 정보를 효과적으로 제공함으로써 한층 진보된 경기기법을 개발하고 과학적인 경기전략을 세울 수 있게 해준다. 본 논문은 이상과 같이 팀 스포츠(Team Spots)의 일종인 축구경기 하이라이트 장면의 자동색인을 위해 뉴럴네트워크 기법을 이용하여 그룹 포메이션(Group Formation) 중의 공격패턴 자동분류 기법을 개발하고 이를 검증하였다. 본 연구에서는 축구경기장 내의 빈번하게 변화하는 장면들을 자동으로 분할하여 대표 프레임을 선정하고, 대표 프레임 상에서 선수들의 위치정보와 공의 위치정보 등을 기초로 하여 경기 중에 이루어지는 선수들의 그룹 포메이션을 추적하여 그룹행동(group behavior)을 분석하고, 뉴럴네트워크의 BP(Back-Propagation) 알고리즘을 사용하여 축구경기 공격패턴을 자동으로 인식 및 분류함으로써 축구경기 하이라이트 장면의 자동추출을 위한 기반을 마련하였다. 본 연구의 실험에는 '98 프랑스 월드컵 축구경기의 다양한 공격패턴에 대한 비디오 영상에서 각각 좌측공격 60개, 우측공격 74개, 중앙공격 72개, 코너킥 39개, 프리킥 52개의 총 297개의 데이터를 추출하여 사용하였다. 실험과는 좌측공격 91.7%, 우측공격 100%, 중앙공격 87.5%, 코너킥 97.4%, 프리킥 75%로서 매우 양호한 인식율을 보였다.

  • PDF

Estimation of Significant Wave Heights from X-Band Radar Based on ANN Using CNN Rainfall Classifier (CNN 강우여부 분류기를 적용한 ANN 기반 X-Band 레이다 유의파고 보정)

  • Kim, Heeyeon;Ahn, Kyungmo;Oh, Chanyeong
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.33 no.3
    • /
    • pp.101-109
    • /
    • 2021
  • Wave observations using a marine X-band radar are conducted by analyzing the backscattered radar signal from sea surfaces. Wave parameters are extracted using Modulation Transfer Function obtained from 3D wave number and frequency spectra which are calculated by 3D FFT of time series of sea surface images (42 images per minute). The accuracy of estimation of the significant wave height is, therefore, critically dependent on the quality of radar images. Wave observations during Typhoon Maysak and Haishen in the summer of 2020 show large errors in the estimation of the significant wave heights. It is because of the deteriorated radar images due to raindrops falling on the sea surface. This paper presents the algorithm developed to increase the accuracy of wave heights estimation from radar images by adopting convolution neural network(CNN) which automatically classify radar images into rain and non-rain cases. Then, an algorithm for deriving the Hs is proposed by creating different ANN models and selectively applying them according to the rain or non-rain cases. The developed algorithm applied to heavy rain cases during typhoons and showed critically improved results.

Forensic Decision of Median Filtering by Pixel Value's Gradients of Digital Image (디지털 영상의 픽셀값 경사도에 의한 미디언 필터링 포렌식 판정)

  • RHEE, Kang Hyeon
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.6
    • /
    • pp.79-84
    • /
    • 2015
  • In a distribution of digital image, there is a serious problem that is a distribution of the altered image by a forger. For the problem solution, this paper proposes a median filtering (MF) image forensic decision algorithm using a feature vector according to the pixel value's gradients. In the proposed algorithm, AR (Autoregressive) coefficients are computed from pixel value' gradients of original image then 1th~6th order coefficients to be six feature vector. And the reconstructed image is produced by the solution of Poisson's equation with the gradients. From the difference image between original and its reconstructed image, four feature vector (Average value, Max. value and the coordinate i,j of Max. value) is extracted. Subsequently, Two kinds of the feature vector combined to 10 Dim. feature vector that is used in the learning of a SVM (Support Vector Machine) classification for MF (Median Filtering) detector of the altered image. On the proposed algorithm of the median filtering detection, compare to MFR (Median Filter Residual) scheme that had the same 10 Dim. feature vectors, the performance is excellent at Unaltered, Averaging filtering ($3{\times}3$) and JPEG (QF=90) images, and less at Gaussian filtering ($3{\times}3$) image. However, in the measured performances of all items, AUC (Area Under Curve) by the sensitivity and 1-specificity is approached to 1. Thus, it is confirmed that the grade evaluation of the proposed algorithm is 'Excellent (A)'.

Selective Word Embedding for Sentence Classification by Considering Information Gain and Word Similarity (문장 분류를 위한 정보 이득 및 유사도에 따른 단어 제거와 선택적 단어 임베딩 방안)

  • Lee, Min Seok;Yang, Seok Woo;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.105-122
    • /
    • 2019
  • Dimensionality reduction is one of the methods to handle big data in text mining. For dimensionality reduction, we should consider the density of data, which has a significant influence on the performance of sentence classification. It requires lots of computations for data of higher dimensions. Eventually, it can cause lots of computational cost and overfitting in the model. Thus, the dimension reduction process is necessary to improve the performance of the model. Diverse methods have been proposed from only lessening the noise of data like misspelling or informal text to including semantic and syntactic information. On top of it, the expression and selection of the text features have impacts on the performance of the classifier for sentence classification, which is one of the fields of Natural Language Processing. The common goal of dimension reduction is to find latent space that is representative of raw data from observation space. Existing methods utilize various algorithms for dimensionality reduction, such as feature extraction and feature selection. In addition to these algorithms, word embeddings, learning low-dimensional vector space representations of words, that can capture semantic and syntactic information from data are also utilized. For improving performance, recent studies have suggested methods that the word dictionary is modified according to the positive and negative score of pre-defined words. The basic idea of this study is that similar words have similar vector representations. Once the feature selection algorithm selects the words that are not important, we thought the words that are similar to the selected words also have no impacts on sentence classification. This study proposes two ways to achieve more accurate classification that conduct selective word elimination under specific regulations and construct word embedding based on Word2Vec embedding. To select words having low importance from the text, we use information gain algorithm to measure the importance and cosine similarity to search for similar words. First, we eliminate words that have comparatively low information gain values from the raw text and form word embedding. Second, we select words additionally that are similar to the words that have a low level of information gain values and make word embedding. In the end, these filtered text and word embedding apply to the deep learning models; Convolutional Neural Network and Attention-Based Bidirectional LSTM. This study uses customer reviews on Kindle in Amazon.com, IMDB, and Yelp as datasets, and classify each data using the deep learning models. The reviews got more than five helpful votes, and the ratio of helpful votes was over 70% classified as helpful reviews. Also, Yelp only shows the number of helpful votes. We extracted 100,000 reviews which got more than five helpful votes using a random sampling method among 750,000 reviews. The minimal preprocessing was executed to each dataset, such as removing numbers and special characters from text data. To evaluate the proposed methods, we compared the performances of Word2Vec and GloVe word embeddings, which used all the words. We showed that one of the proposed methods is better than the embeddings with all the words. By removing unimportant words, we can get better performance. However, if we removed too many words, it showed that the performance was lowered. For future research, it is required to consider diverse ways of preprocessing and the in-depth analysis for the co-occurrence of words to measure similarity values among words. Also, we only applied the proposed method with Word2Vec. Other embedding methods such as GloVe, fastText, ELMo can be applied with the proposed methods, and it is possible to identify the possible combinations between word embedding methods and elimination methods.