• Title/Summary/Keyword: color histogram similarity

Search Result 78, Processing Time 0.023 seconds

A Design and Implementation of a Content_Based Image Retrieval System using Color Space and Keywords (칼라공간과 키워드를 이용한 내용기반 화상검색 시스템 설계 및 구현)

  • Kim, Cheol-Ueon;Choi, Ki-Ho
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.6
    • /
    • pp.1418-1432
    • /
    • 1997
  • Most general content_based image retrieval techniques use color and texture as retrieval indices. In color techniques, color histogram and color pair based color retrieval techniques suffer from a lack of spatial information and text. And This paper describes the design and implementation of content_based image retrieval system using color space and keywords. The preprocessor for image retrieval has used the coordinate system of the existing HSI(Hue, Saturation, Intensity) and preformed to split One image into chromatic region and achromatic region respectively, It is necessary to normalize the size of image for 200*N or N*200 and to convert true colors into 256 color. Two color histograms for background and object are used in order to decide on color selection in the color space. Spatial information is obtained using a maximum entropy discretization. It is possible to choose the class, color, shape, location and size of image by using keyword. An input color is limited by 15 kinds keyword of chromatic and achromatic colors of the Korea Industrial Standards. Image retrieval method is used as the key of retrieval properties in the similarity. The weight values of color space ${\alpha}(%)and\;keyword\;{\beta}(%)$ can be chosen by the user in inputting the query words, controlling the values according to the properties of image_contents. The result of retrieval in the test using extracted feature such as color space and keyword to the query image are lower that those of weight value. In the case of weight value, the average of te measuring parameters shows approximate Precision(0.858), Recall(0.936), RT(1), MT(0). The above results have proved higher retrieval effects than the content_based image retrieval by using color space of keywords.

  • PDF

Color Image Encryption using MLCA and Transformation of Coordinates (MLCA와 좌표변환을 이용한 컬러 영상의 암호화)

  • Yun, Jae-Sik;Nam, Tae-Hee;Cho, Sung-Jin;Kim, Seok-Tae
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.14 no.6
    • /
    • pp.1469-1475
    • /
    • 2010
  • This paper presents a problem of existing encryption methods using pseudo-random numbers based on MLCA or complemented MLCA and proposes a method to resolve this problem. The existing encryption methods have a problem which the edge of original image appear on encrypted image because the image have color similarity of adjacent pixels. In this proposed method, we transform the value and spatial coordinates of all pixels by using pseudo-random numbers based on MLCA. This method can resolve the problem of existing methods and improve the level of encryption by encrypting pixel coordinates and pixel values of original image. The effectiveness of the proposed method is proved by conducting histogram and key space analysis.

Segmentation and Recognition of Traffic Signs using Shape Information and Edge Image in Real Image (실영상에서 형태 정보와 에지 영상을 이용한 교통 표지판 영역 추출과 인식)

  • Kwak, Hyun-Wook;Oh,Jun-Taek;Kim, Wook-Hyun
    • The KIPS Transactions:PartB
    • /
    • v.11B no.2
    • /
    • pp.149-158
    • /
    • 2004
  • This study proposes a method for segmentation and recognition of traffic signs using shape information and edge image in real image. It first segments traffic sign candidate regions by connected component algorithm from binary images, obtained by utilizing the RGB color ratio of each pixel in the image, and then extracts actual traffic signs based on their symmetries on X- and Y-axes. Histogram equalization is performed for unsegmented candidate regions caused by low contrast in the image. In the recognition stage, it utilizes shape information including projection profiles on X- and Y-axes, moment, and the number of crossings and distance which concentric circular patterns and 8-directional rays from region center intersects with edges of traffic signs. It finally performs recognition by measuring similarity with the templates in the database. It will be shown from several experimental results that the system is robust to environmental factors, such as light and weather condition.

Person Identification based on Clothing Feature (의상 특징 기반의 동일인 식별)

  • Choi, Yoo-Joo;Park, Sun-Mi;Cho, We-Duke;Kim, Ku-Jin
    • Journal of the Korea Computer Graphics Society
    • /
    • v.16 no.1
    • /
    • pp.1-7
    • /
    • 2010
  • With the widespread use of vision-based surveillance systems, the capability for person identification is now an essential component. However, the CCTV cameras used in surveillance systems tend to produce relatively low-resolution images, making it difficult to use face recognition techniques for person identification. Therefore, an algorithm is proposed for person identification in CCTV camera images based on the clothing. Whenever a person is authenticated at the main entrance of a building, the clothing feature of that person is extracted and added to the database. Using a given image, the clothing area is detected using background subtraction and skin color detection techniques. The clothing feature vector is then composed of textural and color features of the clothing region, where the textural feature is extracted based on a local edge histogram, while the color feature is extracted using octree-based quantization of a color map. When given a query image, the person can then be identified by finding the most similar clothing feature from the database, where the Euclidean distance is used as the similarity measure. Experimental results show an 80% success rate for person identification with the proposed algorithm, and only a 43% success rate when using face recognition.

A Study on Features Analysis for Retrieving Image Containing Personal Information on the Web (인터넷상에서 개인식별정보가 포함된 영상 검색을 위한 특징정보 분석에 관한 연구)

  • Kim, Jong-Bae
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.48 no.3
    • /
    • pp.91-101
    • /
    • 2011
  • Internet is becoming increasingly popular due to the rapid development of information and communication technology. There has been a convenient social activities such as the mutual exchange of information, e-commerce, internet banking, etc. through cyberspace on a computer. However, by using the convenience of the internet, the personal IDs(identity card, driving license, passport, student ID, etc.) represented by the electronic media are exposed on the internet frequently. Therefore, this study propose a feature extraction method to analyze the characteristics of image files containing personal information and a image retrieval method to find the images using the extracted features. The proposed method selects the feature information from color, texture, and shape of the images, and the images as searched by similarity analysis between feature information. The result which it experiments from the image which it acquires from the web-based image DB and correct image retrieval rate is 89%, the computing time per frame is 0.17 seconds. The proposed method can be efficiently apply a system to search the image files containing personal information and to determine the criteria of exposure of personal information.

Spatiotemporal Saliency-Based Video Summarization on a Smartphone (스마트폰에서의 시공간적 중요도 기반의 비디오 요약)

  • Lee, Won Beom;Williem, Williem;Park, In Kyu
    • Journal of Broadcast Engineering
    • /
    • v.18 no.2
    • /
    • pp.185-195
    • /
    • 2013
  • In this paper, we propose a video summarization technique on a smartphone, based on spatiotemporal saliency. The proposed technique detects scene changes by computing the difference of the color histogram, which is robust to camera and object motion. Then the similarity between adjacent frames, face region, and frame saliency are computed to analyze the spatiotemporal saliency in a video clip. Over-segmented hierarchical tree is created using scene changes and is updated iteratively using mergence and maintenance energies computed during the analysis procedure. In the updated hierarchical tree, segmented frames are extracted by applying a greedy algorithm on the node with high saliency when it satisfies the reduction ratio and the minimum interval requested by the user. Experimental result shows that the proposed method summaries a 2 minute-length video in about 10 seconds on a commercial smartphone. The summarization quality is superior to the commercial video editing software, Muvee.

Coated Tongue Region Extraction using the Fluorescence Response of the Tongue Coating by Ultraviolet Light Source (설태의 자외선 형광 반응을 이용한 설태 영역 추출)

  • Choi, Chang-Yur;Lee, Woo-Beom;Hong, You-Sik;Nam, Dong-Hyun;Lee, Sang-Suk
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.12 no.4
    • /
    • pp.181-188
    • /
    • 2012
  • An effective extraction method for extracting a coated tongue is proposed in this paper, which is used as the diagnostic criteria in the tongue diagnosis. Proposed method uses the fluorescence response characteristics of the coated tongue that is occurred by using the ultraviolet light. Specially, this method can solved the previous problems including the issue in the limits of the diagnosis environment and in the objectivity of the diagnosis results. In our method, original tongue image is acquired by using the ultraviolet light, and binarization is performed by thresholding a valley-points in the histogram that corresponds to the color difference of tongue body and tongue coating. Final view image is presented to the oriental doctor, after applying the canny-edge algorithm to the binary image, and edge image is added to the original image. In order to evaluate the performance of the our proposed method, after building a various tongue image, we compared the true region of coated tongue by the oriental doctor's hand with the extracted region by the our method. As a result, the proposed method showed the average 87.87% extraction ratio. The shape of the extracted coated tongue region showed also significantly higher similarity.

Emotion-based Video Scene Retrieval using Interactive Genetic Algorithm (대화형 유전자 알고리즘을 이용한 감성기반 비디오 장면 검색)

  • Yoo Hun-Woo;Cho Sung-Bae
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.10 no.6
    • /
    • pp.514-528
    • /
    • 2004
  • An emotion-based video scene retrieval algorithm is proposed in this paper. First, abrupt/gradual shot boundaries are detected in the video clip representing a specific story Then, five video features such as 'average color histogram' 'average brightness', 'average edge histogram', 'average shot duration', and 'gradual change rate' are extracted from each of the videos and mapping between these features and the emotional space that user has in mind is achieved by an interactive genetic algorithm. Once the proposed algorithm has selected videos that contain the corresponding emotion from initial population of videos, feature vectors from the selected videos are regarded as chromosomes and a genetic crossover is applied over them. Next, new chromosomes after crossover and feature vectors in the database videos are compared based on the similarity function to obtain the most similar videos as solutions of the next generation. By iterating above procedures, new population of videos that user has in mind are retrieved. In order to show the validity of the proposed method, six example categories such as 'action', 'excitement', 'suspense', 'quietness', 'relaxation', 'happiness' are used as emotions for experiments. Over 300 commercial videos, retrieval results show 70% effectiveness in average.