• Title/Summary/Keyword: Feature image

Search Result 3,599, Processing Time 0.032 seconds

Image Retrieval Using Color Correlogram from a Segmented Image (분할된 영상에서의 칼라 코렐로그램을 이용한 영상검색)

  • 안명석;조석제
    • Proceedings of the Korea Institute of Convergence Signal Processing
    • /
    • 2000.12a
    • /
    • pp.153-156
    • /
    • 2000
  • Recently, there has been studied on feature extraction method for efficient content-based image retrieval. Especially, Many researchers have been studying on extracting feature from color Information, because of its advantages. This paper proposes a feature and its extraction method based on color correlogram that is extracted from color information in an image. the proposed method is computed from the image segmented into two parts; the complex part and the plain part. Our experiments show that the performance of the proposed method is better as compared with that of the original color correlogram method.

  • PDF

Feature-based Image Analysis for Object Recognition on Satellite Photograph (인공위성 영상의 객체인식을 위한 영상 특징 분석)

  • Lee, Seok-Jun;Jung, Soon-Ki
    • Journal of the HCI Society of Korea
    • /
    • v.2 no.2
    • /
    • pp.35-43
    • /
    • 2007
  • This paper presents a system for image matching and recognition based on image feature detection and description techniques from artificial satellite photographs. We propose some kind of parameters from the varied environmental elements happen by image handling process. The essential point of this experiment is analyzes that affects match rate and recognition accuracy when to change of state of each parameter. The proposed system is basically inspired by Lowe's SIFT(Scale-Invariant Transform Feature) algorithm. The descriptors extracted from local affine invariant regions are saved into database, which are defined by k-means performed on the 128-dimensional descriptor vectors on an artificial satellite photographs from Google earth. And then, a label is attached to each cluster of the feature database and acts as guidance for an appeared building's information in the scene from camera. This experiment shows the various parameters and compares the affected results by changing parameters for the process of image matching and recognition. Finally, the implementation and the experimental results for several requests are shown.

  • PDF

Emotion Recognition and Expression Method using Bi-Modal Sensor Fusion Algorithm (다중 센서 융합 알고리즘을 이용한 감정인식 및 표현기법)

  • Joo, Jong-Tae;Jang, In-Hun;Yang, Hyun-Chang;Sim, Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.13 no.8
    • /
    • pp.754-759
    • /
    • 2007
  • In this paper, we proposed the Bi-Modal Sensor Fusion Algorithm which is the emotional recognition method that be able to classify 4 emotions (Happy, Sad, Angry, Surprise) by using facial image and speech signal together. We extract the feature vectors from speech signal using acoustic feature without language feature and classify emotional pattern using Neural-Network. We also make the feature selection of mouth, eyes and eyebrows from facial image. and extracted feature vectors that apply to Principal Component Analysis(PCA) remakes low dimension feature vector. So we proposed method to fused into result value of emotion recognition by using facial image and speech.

Full face recognition using the feature extracted gy shape analyzing and the back-propagation algorithm (형태분석에 의한 특징 추출과 BP알고리즘을 이용한 정면 얼굴 인식)

  • 최동선;이주신
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.33B no.10
    • /
    • pp.63-71
    • /
    • 1996
  • This paper proposes a method which analyzes facial shape and extracts positions of eyes regardless of the tilt and the size of input iamge. With the extracted feature parameters of facial element by the method, full human faces are recognized by a neural network which BP algorithm is applied on. Input image is changed into binary codes, and then labelled. Area, circumference, and circular degree of the labelled binary image are obtained by using chain code and defined as feature parameters of face image. We first extract two eyes from the similarity and distance of feature parameter of each facial element, and then input face image is corrected by standardizing on two extracted eyes. After a mask is genrated line historgram is applied to finding the feature points of facial elements. Distances and angles between the feature points are used as parameters to recognize full face. To show the validity learning algorithm. We confirmed that the proposed algorithm shows 100% recognition rate on both learned and non-learned data for 20 persons.

  • PDF

Animal Fur Recognition Algorithm Based on Feature Fusion Network

  • Liu, Peng;Lei, Tao;Xiang, Qian;Wang, Zexuan;Wang, Jiwei
    • Journal of Multimedia Information System
    • /
    • v.9 no.1
    • /
    • pp.1-10
    • /
    • 2022
  • China is a big country in animal fur industry. The total production and consumption of fur are increasing year by year. However, the recognition of fur in the fur production process still mainly relies on the visual identification of skilled workers, and the stability and consistency of products cannot be guaranteed. In response to this problem, this paper proposes a feature fusion-based animal fur recognition network on the basis of typical convolutional neural network structure, relying on rapidly developing deep learning techniques. This network superimposes texture feature - the most prominent feature of fur image - into the channel dimension of input image. The output feature map of the first layer convolution is inverted to obtain the inverted feature map and concat it into the original output feature map, then Leaky ReLU is used for activation, which makes full use of the texture information of fur image and the inverted feature information. Experimental results show that the algorithm improves the recognition accuracy by 9.08% on Fur_Recognition dataset and 6.41% on CIFAR-10 dataset. The algorithm in this paper can change the current situation that fur recognition relies on manual visual method to classify, and can lay foundation for improving the efficiency of fur production technology.

Edge Feature Extract CBIRS for Car Retrieval : CBIRS/EFI (차량 검색을 위한 측면 에지 특징 추출 내용기반 검색 : CBIRS/EFI)

  • Koo, Gun-Seo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.11
    • /
    • pp.75-82
    • /
    • 2010
  • The paper proposed CBIRS/EFI with contents based search technique using edge feature information of the object from image information of the object which is uncertain. In order to search specially efficiently case of partial image information of the object, we used the search technique which extracts outline information and color information in feature information of object. In order to experiment this, we extracted side edge feature information of the vehicle for feature information of the object after capture the car image of the underground garage. This is the system which applies a contents base search by the result which analyzes the image which extracts a feature, an original image to search and a last similar measurement result. This system compared in FE-CBIRS systems which are an existing feature extraction contents base image retrieval system and the function which improves the accuracy and an effectiveness of search rate was complemented. The performance appraisal of CBIRS/EFI systems applied edge extraction feature information and color information of the cars. And we compared a color feature search time, a shape characteristic search time and a search rate from the process which searches area feature information. We extracted the case 91.84% of car edge feature extraction rate. And a average search time of CBIRS/EFI is showing a difference of average 0.4-0.9 seconds than FE-CBIRS from vehicle. color search time, shape characteristic search time and similar search time. So, it was proven with the fact that is excellent.

Genetic lesion matching algorithm using medical image (의료영상 이미지를 이용한 유전병변 정합 알고리즘)

  • Cho, Young-bok;Woo, Sung-Hee;Lee, Sang-Ho;Han, Chang-Su
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.5
    • /
    • pp.960-966
    • /
    • 2017
  • In this paper, we proposed an algorithm that can extract lesion by inputting a medical image. Feature points are extracted using SIFT algorithm to extract genetic training of medical image. To increase the intensity of the feature points, the input image and that raining image are matched using vector similarity and the lesion is extracted. The vector similarity match can quickly lead to lesions. Since the direction vector is generated from the local feature point pair, the direction itself only shows the local feature, but it has the advantage of comparing the similarity between the other vectors existing between the two images and expanding to the global feature. The experimental results show that the lesion matching error rate is 1.02% and the processing speed is improved by about 40% compared to the case of not using the feature point intensity information.

AUTOMATIC SELECTION AND ADJUSTMENT OF FEATURES FOR IMAGE CLASSIFICATION

  • Saiki, Kenji;Nagao, Tomoharu
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.525-528
    • /
    • 2009
  • Recently, image classification has been an important task in various fields. Generally, the performance of image classification is not good without the adjustment of image features. Therefore, it is desired that the way of automatic feature extraction. In this paper, we propose an image classification method which adjusts image features automatically. We assume that texture features are useful in image classification tasks because natural images are composed of several types of texture. Thus, the classification accuracy rate is improved by using distribution of texture features. We obtain texture features by calculating image features from a current considering pixel and its neighborhood pixels. And we calculate image features from distribution of textures feature. Those image features are adjusted to image classification tasks using Genetic Algorithm. We apply proposed method to classifying images into "head" or "non-head" and "male" or "female".

  • PDF

Content-Based Image Retrieval Algorithm Using HAQ Algorithm and Moment-Based Feature (HAQ 알고리즘과 Moment 기반 특징을 이용한 내용 기반 영상 검색 알고리즘)

  • 김대일;강대성
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.4
    • /
    • pp.113-120
    • /
    • 2004
  • In this paper, we propose an efficient feature extraction and image retrieval algorithm for content-based retrieval method. First, we extract the object using Gaussian edge detector for input image which is key frames of MPEG video and extract the object features that are location feature, distributed dimension feature and invariant moments feature. Next, we extract the characteristic color feature using the proposed HAQ(Histogram Analysis md Quantization) algorithm. Finally, we implement an retrieval of four features in sequence with the proposed matching method for query image which is a shot frame except the key frames of MPEG video. The purpose of this paper is to propose the novel content-based image retrieval algerian which retrieves the key frame in the shot boundary of MPEG video belonging to the scene requested by user. The experimental results show an efficient retrieval for 836 sample images in 10 music videos using the proposed algorithm.

Contend Base Image Retrieval using Color Feature of Central Region and Optimized Comparing Bin (중앙 영역의 컬러 특징과 최적화된 빈 수를 이용한 내용기 반 영상검색)

  • Ryu, Eun-Ju;Song, Young-Jun;Park, Won-Bae;Ahn, Jae-Hyeong
    • The KIPS Transactions:PartB
    • /
    • v.11B no.5
    • /
    • pp.581-586
    • /
    • 2004
  • In this paper, we proposed a content-based image retrieval using a color feature for central region and its optimized comparing bin method. Human's visual characteristic is influenced by existent of central object. So we supposed that object is centrally located in image and then we extract color feature at central region. When the background of image is simple, the retrieval result can be bad affected by major color of background. Our method overcome this drawback as a result of the human visual characteristic. After we transform Image into HSV color space, we extract color feature from the quantized image with 16 level. The experimental results showed that the method using the eight high rank bin is better than using the 16 bin The case which extracts the feature with image's central region was superior compare with the case which extracts the feature with the whole image about 5%.