• Title/Summary/Keyword: Image Features

Search Result 3,365, Processing Time 0.039 seconds

An Extended Concept-based Image Retrieval System : E-COIRS (확장된 개념 기반 이미지 검색 시스템)

  • Kim, Yong-Il;Yang, Jae-Dong;Yang, Hyoung-Jeong
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.3
    • /
    • pp.303-317
    • /
    • 2002
  • In this paper, we design and implement E-COIRS enabling users to query with concepts and image features used for further refining the concepts. For example, E-COIRS supports the query "retrieve images containing black home appliance to north of reception set. "The query includes two types of concepts: IS-A and composite. "home appliance"is an IS-A concept, and "reception set" is a composite concept. For evaluating such a query. E-COIRS includes three important components: a visual image indexer, thesauri and a query processor. Each pair of objects in an mage captured by the visual image indexer is converted into a triple. The triple consists of the two object identifiers (oids) and their spatial relationship. All the features of an object is referenced by its old. A composite concept is detected by the triple thesaurus and IS-A concept is recolonized by the fuzzy term thesaurus. The query processor obtains an image set by matching each triple in a user with an inverted file and CS-Tree. To support efficient storage use and fast retrieval on high-dimensional feature vectors, E-COIRS uses Cell-based Signature tree(CS-Tree). E-COIRS is a more advanced content-based image retrieval system than other systems which support only concepts or image features.

An Image Retrieving Scheme Using Salient Features and Annotation Watermarking

  • Wang, Jenq-Haur;Liu, Chuan-Ming;Syu, Jhih-Siang;Chen, Yen-Lin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.1
    • /
    • pp.213-231
    • /
    • 2014
  • Existing image search systems allow users to search images by keywords, or by example images through content-based image retrieval (CBIR). On the other hand, users might learn more relevant textual information about an image from its text captions or surrounding contexts within documents or Web pages. Without such contexts, it's difficult to extract semantic description directly from the image content. In this paper, we propose an annotation watermarking system for users to embed text descriptions, and retrieve more relevant textual information from similar images. First, tags associated with an image are converted by two-dimensional code and embedded into the image by discrete wavelet transform (DWT). Next, for images without annotations, similar images can be obtained by CBIR techniques and embedded annotations can be extracted. Specifically, we use global features such as color ratios and dominant sub-image colors for preliminary filtering. Then, local features such as Scale-Invariant Feature Transform (SIFT) descriptors are extracted for similarity matching. This design can achieve good effectiveness with reasonable processing time in practical systems. Our experimental results showed good accuracy in retrieving similar images and extracting relevant tags from similar images.

Identification of Transformed Image Using the Composition of Features

  • Yang, Won-Keun;Cho, A-Young;Cho, Ik-Hwan;Oh, Weon-Geun;Jeong, Dong-Seok
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.6
    • /
    • pp.764-776
    • /
    • 2008
  • Image identification is the process of checking whether the query image is the transformed version of the specific original image or not. In this paper, image identification method based on feature composition is proposed. Used features include color distance, texture information and average pixel intensity. We extract color characteristics using color distance and texture information by Modified Generalized Symmetry Transform as well as average intensity of each pixel as features. Individual feature is quantized adaptively to be used as bins of histogram. The histogram is normalized according to data type and it is used as the signature in comparing the query image with database images. In matching part, Manhattan distance is used for measuring distance between two signatures. To evaluate the performance of the proposed method, independent test and accuracy test are achieved. In independent test, 60,433 images are used to evaluate the ability of discrimination between different images. And 4,002 original images and its 29 transformed versions are used in accuracy test, which evaluate the ability that the proposed algorithm can find the original image correctly when some transforms was applied in original image. Experiment results show that the proposed identification method has good performance in accuracy test. And the proposed method is very useful in real environment because of its high accuracy and fast matching capacity.

  • PDF

A Classifier for Textured Images Based on Matrix Feature (행렬 속성을 이용하는 질감 영상 분별기)

  • 김준철;이준환
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.3
    • /
    • pp.91-102
    • /
    • 1994
  • For the analysis of textured image, it requires large storage space and computation time to calculate the matrix features such as SGLDM(Spatial Gray Level Dependence Matrix). NGLDM(Neighboring Gray Level Dependence Matrix). NSGLDM(Neighboring Spatial Gray Level Dependence Matrix) and GLRLM(Gray Level Run Length Matrix). In spite of a large amount of information that each matrix contains, a set of several correlated scalar features calculated from the matrix is not sufficient to approximate it. In this paper, we propose a new classifier for textured images based on these matrices in which the projected vectors of each matrix on the meaningful directions are used as features. In the proposed method, an unknown image is classified to the class of a known image that gives the maximum similarity between the projected model vector from the known image and the vector from the unknown image. In the experiment to classify images of agricultural products, the proposed method shows good performance as much as 85-95% of correct classification ratio.

  • PDF

An Extracting and Indexing Schema of Compressed Medical Images (축소변환된 의료 이미지의 질감 특징 추출과 인덱싱)

  • 위희정;엄기현
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2000.04a
    • /
    • pp.328-331
    • /
    • 2000
  • In this paper , we propose a texture feature extraction method of reduce the massive computational time on extracting texture, features of large sized medical such as MRI, CT-scan , and an index structure, called GLTFT, to speed up the retrieval performance. For these, the original image is transformed into a compressed image by Wavelet transform , and textural features such as contrast, energy, entropy, and homogeneity of the compressed image is extracted by using GLCM(Gray Level Co-occurrence Metrix) . The proposed index structure is organized by using the textural features. The processing in compressed domain can give the solution of storage space and the reduction of computational time of feature extracting . And , by GLTFT index structure, image retrieval performance can be expected to be improved by reducing the retrieval range . Our experiment on 270 MRIs as image database shows that shows that such expectation can be got.

  • PDF

Impulse Noise Detection Using Self-Organizing Neural Network and Its Application to Selective Median Filtering (Self-Organizing Neural Network를 이용한 임펄스 노이즈 검출과 선택적 미디언 필터 적용)

  • Lee Chong Ho;Dong Sung Soo;Wee Jae Woo;Song Seung Min
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.54 no.3
    • /
    • pp.166-173
    • /
    • 2005
  • Preserving image features, edges and details in the process of impulsive noise filtering is an important problem. To avoid image blurring, only corrupted pixels must be filtered. In this paper, we propose an effective impulse noise detection method using Self-Organizing Neural Network(SONN) which applies median filter selectively for removing random-valued impulse noises while preserving image features, edges and details. Using a $3\times3$ window, we obtain useful local features with which impulse noise patterns are classified. SONN is trained with sample image patterns and each pixel pattern is classified by its local information in the image. The results of the experiments with various images which are the noise range of $5-15\%$ show that our method performs better than other methods which use multiple threshold values for impulse noise detection.

Texture-based PCA for Analyzing Document Image (텍스처 정보 기반의 PCA를 이용한 문서 영상의 분석)

  • Kim, Bo-Ram;Kim, Wook-Hyun
    • Proceedings of the IEEK Conference
    • /
    • 2006.06a
    • /
    • pp.283-284
    • /
    • 2006
  • In this paper, we propose a novel segmentation and classification method using texture features for the document image. First, we extract the local entropy and then segment the document image to separate the background and the foreground using the Otsu's method. Finally, we classify the segmented regions into each component using PCA(principle component analysis) algorithm based on the texture features that are extracted from the co-occurrence matrix for the entropy image. The entropy-based segmentation is robust to not only noise and the change of light, but also skew and rotation. Texture features are not restricted from any form of the document image and have a superior discrimination for each component. In addition, PCA algorithm used for the classifier can classify the components more robustly than neural network.

  • PDF

Visual Location Recognition Using Time-Series Streetview Database (시계열 스트리트뷰 데이터베이스를 이용한 시각적 위치 인식 알고리즘)

  • Park, Chun-Su;Choeh, Joon-Yeon
    • Journal of the Semiconductor & Display Technology
    • /
    • v.18 no.4
    • /
    • pp.57-61
    • /
    • 2019
  • Nowadays, portable digital cameras such as smart phone cameras are being popularly used for entertainment and visual information recording. Given a database of geo-tagged images, a visual location recognition system can determine the place depicted in a query photo. One of the most common visual location recognition approaches is the bag-of-words method where local image features are clustered into visual words. In this paper, we propose a new bag-of-words-based visual location recognition algorithm using time-series streetview database. The proposed algorithm selects only a small subset of image features which will be used in image retrieval process. By reducing the number of features to be used, the proposed algorithm can reduce the memory requirement of the image database and accelerate the retrieval process.

Using Radon Transform for Image Retrieval (영상 검색을 위한 Radon 변형의 이용)

  • Seo, Jeong-Man
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.6
    • /
    • pp.65-71
    • /
    • 2009
  • The basic features in the indexing and retrieval of the image is used color, shape, and texture in traditional image retrieval method. We do not use these features and offers a new way. For content-based video indexing and retrieval, visual features used to measure the similarity of the geometric method is presented. This method is called the Radon transform. Without separation, this method is calculated based on the geometric distribution of image. In the experiment has a very good search results.

Spam Image Detection Model based on Deep Learning for Improving Spam Filter

  • Seong-Guk Nam;Dong-Gun Lee;Yeong-Seok Seo
    • Journal of Information Processing Systems
    • /
    • v.19 no.3
    • /
    • pp.289-301
    • /
    • 2023
  • Due to the development and dissemination of modern technology, anyone can easily communicate using services such as social network service (SNS) through a personal computer (PC) or smartphone. The development of these technologies has caused many beneficial effects. At the same time, bad effects also occurred, one of which was the spam problem. Spam refers to unwanted or rejected information received by unspecified users. The continuous exposure of such information to service users creates inconvenience in the user's use of the service, and if filtering is not performed correctly, the quality of service deteriorates. Recently, spammers are creating more malicious spam by distorting the image of spam text so that optical character recognition (OCR)-based spam filters cannot easily detect it. Fortunately, the level of transformation of image spam circulated on social media is not serious yet. However, in the mail system, spammers (the person who sends spam) showed various modifications to the spam image for neutralizing OCR, and therefore, the same situation can happen with spam images on social media. Spammers have been shown to interfere with OCR reading through geometric transformations such as image distortion, noise addition, and blurring. Various techniques have been studied to filter image spam, but at the same time, methods of interfering with image spam identification using obfuscated images are also continuously developing. In this paper, we propose a deep learning-based spam image detection model to improve the existing OCR-based spam image detection performance and compensate for vulnerabilities. The proposed model extracts text features and image features from the image using four sub-models. First, the OCR-based text model extracts the text-related features, whether the image contains spam words, and the word embedding vector from the input image. Then, the convolution neural network-based image model extracts image obfuscation and image feature vectors from the input image. The extracted feature is determined whether it is a spam image by the final spam image classifier. As a result of evaluating the F1-score of the proposed model, the performance was about 14 points higher than the OCR-based spam image detection performance.