• Title/Summary/Keyword: Automatic multi-object extraction

Search Result 11, Processing Time 0.021 seconds

Automation of Snake for Extraction of Multi-Object Contours from a Natural Scene (자연배경에서 여러 객체 윤곽선의 추출을 위한 스네이크의 자동화)

  • 최재혁;서경석;김복만;최흥문
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.9 no.6
    • /
    • pp.712-717
    • /
    • 2003
  • A novel multi-snake is proposed for efficient extraction of multi-object contours from a natural scene. An NTGST(noise-tolerant generalized symmetry transform) is used as a context-free attention operator to detect and locate multiple objects from a complex background and then the snake points are automatically initialized nearby the contour of each detected object using symmetry map of the NTGST before multiple snakes are introduced. These procedures solve the knotty subjects of automatic snake initialization and simultaneous extraction of multi-object contours in conventional snake algorithms. Because the snake points are initialized nearby the actual contour of each object, as close as possible, contours with high convexity and/or concavity can be easily extracted. The experimental results show that the proposed method can efficiently extract multi-object contours from a noisy and complex background of natural scenes.

Enhanced Object Extraction Method Based on Multi-channel Saliency Map (Saliency Map 다중 채널을 기반으로 한 개선된 객체 추출 방법)

  • Choi, Young-jin;Cui, Run;Kim, Kwang-Rag;Kim, Hyoung Joong
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.53 no.2
    • /
    • pp.53-61
    • /
    • 2016
  • Extracting focused object with saliency map is still remaining as one of the most highly tasked research area around computer vision for it is hard to estimate. Through this paper, we propose enhanced object extraction method based on multi-channel saliency map which could be done automatically without machine learning. Proposed Method shows a higher accuracy than Itti method using SLIC, Euclidean, and LBP algorithm as for object extraction. Experiments result shows that our approach is possible to be used for automatic object extraction without any previous training procedure through focusing on the main object from the image instead of estimating the whole image from background to foreground.

A study on automatic wear debris recognition by using particle feature extraction (입자 유형별 형상추출에 의한 마모입자 자동인식에 관한 연구)

  • ;;;Grigoriev, A.Y.
    • Proceedings of the Korean Society of Tribologists and Lubrication Engineers Conference
    • /
    • 1998.04a
    • /
    • pp.314-320
    • /
    • 1998
  • Wear debris morphology is closely related to the wear mode and mechanism occured. Image recognition of wear debris is, therefore, a powerful tool in wear monitoring. But it has usually required expert's experience and the results could be too subjective. Development of automatic tools for wear debris recognition is needed to solve this problem. In this work, an algorithm for automatic wear debris recognition was suggested and implemented by PC base software. The presented method defined a characteristic 3-dimensional feature space where typical types of wear debris were separately located by the knowledge-based system and compared the similarity of object wear debris concerned. The 3-dimensional feature space was obtained from multiple feature vectors by using a multi-dimensional scaling technique. The results showed that the presented automatic wear debris recognition was satisfactory in many cases application.

  • PDF

A Study on Automatic wear Debris Recognition by using Particle Feature Extraction (입자 유형별 형상추출에 의한 마모입자 자동인식에 관한 연구)

  • ;;;A. Y. Grigoriev
    • Tribology and Lubricants
    • /
    • v.15 no.2
    • /
    • pp.206-211
    • /
    • 1999
  • Wear debris morphology is closely related to the wear mode and mechanism occured. Image recognition of wear debris is, therefore, a powerful tool in wear monitoring. But it has usually required expert's experience and the results could be too subjective. Development of automatic tools for wear debris recognition is needed to solve this problem. In this work, an algorithm for automatic wear debris recognition was suggested and implemented by PC base software. The presented method defined a characteristic 3-dimensional feature space where typical types of wear debris were separately located by the knowledge-based system and compared the similarity of object wear debris concerned. The 3-dimensional feature space was obtained from multiple feature vectors by using a multi-dimensional scaling technique. The results showed that the presented automatic wear debris recognition was satisfactory in many cases application.

Acceleration of Viewport Extraction for Multi-Object Tracking Results in 360-degree Video (360도 영상에서 다중 객체 추적 결과에 대한 뷰포트 추출 가속화)

  • Heesu Park;Seok Ho Baek;Seokwon Lee;Myeong-jin Lee
    • Journal of Advanced Navigation Technology
    • /
    • v.27 no.3
    • /
    • pp.306-313
    • /
    • 2023
  • Realistic and graphics-based virtual reality content is based on 360-degree videos, and viewport extraction through the viewer's intention or automatic recommendation function is essential. This paper designs a viewport extraction system based on multiple object tracking in 360-degree videos and proposes a parallel computing structure necessary for multiple viewport extraction. The viewport extraction process in 360-degree videos is parallelized by composing pixel-wise threads, through 3D spherical surface coordinate transformation from ERP coordinates and 2D coordinate transformation of 3D spherical surface coordinates within the viewport. The proposed structure evaluated the computation time for up to 30 viewport extraction processes in aerial 360-degree video sequences and confirmed up to 5240 times acceleration compared to the CPU-based computation time proportional to the number of viewports. When using high-speed I/O or memory buffers that can reduce ERP frame I/O time, viewport extraction time can be further accelerated by 7.82 times. The proposed parallelized viewport extraction structure can be applied to simultaneous multi-access services for 360-degree videos or virtual reality contents and video summarization services for individual users.

Multi-Level Segmentation of Infrared Images with Region of Interest Extraction

  • Yeom, Seokwon
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.16 no.4
    • /
    • pp.246-253
    • /
    • 2016
  • Infrared (IR) imaging has been researched for various applications such as surveillance. IR radiation has the capability to detect thermal characteristics of objects under low-light conditions. However, automatic segmentation for finding the object of interest would be challenging since the IR detector often provides the low spatial and contrast resolution image without color and texture information. Another hindrance is that the image can be degraded by noise and clutters. This paper proposes multi-level segmentation for extracting regions of interest (ROIs) and objects of interest (OOIs) in the IR scene. Each level of the multi-level segmentation is composed of a k-means clustering algorithm, an expectation-maximization (EM) algorithm, and a decision process. The k-means clustering initializes the parameters of the Gaussian mixture model (GMM), and the EM algorithm estimates those parameters iteratively. During the multi-level segmentation, the area extracted at one level becomes the input to the next level segmentation. Thus, the segmentation is consecutively performed narrowing the area to be processed. The foreground objects are individually extracted from the final ROI windows. In the experiments, the effectiveness of the proposed method is demonstrated using several IR images, in which human subjects are captured at a long distance. The average probability of error is shown to be lower than that obtained from other conventional methods such as Gonzalez, Otsu, k-means, and EM methods.

Multi-robot Mapping Using Omnidirectional-Vision SLAM Based on Fisheye Images

  • Choi, Yun-Won;Kwon, Kee-Koo;Lee, Soo-In;Choi, Jeong-Won;Lee, Suk-Gyu
    • ETRI Journal
    • /
    • v.36 no.6
    • /
    • pp.913-923
    • /
    • 2014
  • This paper proposes a global mapping algorithm for multiple robots from an omnidirectional-vision simultaneous localization and mapping (SLAM) approach based on an object extraction method using Lucas-Kanade optical flow motion detection and images obtained through fisheye lenses mounted on robots. The multi-robot mapping algorithm draws a global map by using map data obtained from all of the individual robots. Global mapping takes a long time to process because it exchanges map data from individual robots while searching all areas. An omnidirectional image sensor has many advantages for object detection and mapping because it can measure all information around a robot simultaneously. The process calculations of the correction algorithm are improved over existing methods by correcting only the object's feature points. The proposed algorithm has two steps: first, a local map is created based on an omnidirectional-vision SLAM approach for individual robots. Second, a global map is generated by merging individual maps from multiple robots. The reliability of the proposed mapping algorithm is verified through a comparison of maps based on the proposed algorithm and real maps.

Novel Intent based Dimension Reduction and Visual Features Semi-Supervised Learning for Automatic Visual Media Retrieval

  • kunisetti, Subramanyam;Ravichandran, Suban
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.230-240
    • /
    • 2022
  • Sharing of online videos via internet is an emerging and important concept in different types of applications like surveillance and video mobile search in different web related applications. So there is need to manage personalized web video retrieval system necessary to explore relevant videos and it helps to peoples who are searching for efficient video relates to specific big data content. To evaluate this process, attributes/features with reduction of dimensionality are computed from videos to explore discriminative aspects of scene in video based on shape, histogram, and texture, annotation of object, co-ordination, color and contour data. Dimensionality reduction is mainly depends on extraction of feature and selection of feature in multi labeled data retrieval from multimedia related data. Many of the researchers are implemented different techniques/approaches to reduce dimensionality based on visual features of video data. But all the techniques have disadvantages and advantages in reduction of dimensionality with advanced features in video retrieval. In this research, we present a Novel Intent based Dimension Reduction Semi-Supervised Learning Approach (NIDRSLA) that examine the reduction of dimensionality with explore exact and fast video retrieval based on different visual features. For dimensionality reduction, NIDRSLA learns the matrix of projection by increasing the dependence between enlarged data and projected space features. Proposed approach also addressed the aforementioned issue (i.e. Segmentation of video with frame selection using low level features and high level features) with efficient object annotation for video representation. Experiments performed on synthetic data set, it demonstrate the efficiency of proposed approach with traditional state-of-the-art video retrieval methodologies.

Multi-camera System Calibration with Built-in Relative Orientation Constraints (Part 2) Automation, Implementation, and Experimental Results

  • Lari, Zahra;Habib, Ayman;Mazaheri, Mehdi;Al-Durgham, Kaleel
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.32 no.3
    • /
    • pp.205-216
    • /
    • 2014
  • Multi-camera systems have been widely used as cost-effective tools for the collection of geospatial data for various applications. In order to fully achieve the potential accuracy of these systems for object space reconstruction, careful system calibration should be carried out prior to data collection. Since the structural integrity of the involved cameras' components and system mounting parameters cannot be guaranteed over time, multi-camera system should be frequently calibrated to confirm the stability of the estimated parameters. Therefore, automated techniques are needed to facilitate and speed up the system calibration procedure. The automation of the multi-camera system calibration approach, which was proposed in the first part of this paper, is contingent on the automated detection, localization, and identification of the object space signalized targets in the images. In this paper, the automation of the proposed camera calibration procedure through automatic target extraction and labelling approaches will be presented. The introduced automated system calibration procedure is then implemented for a newly-developed multi-camera system while considering the optimum configuration for the data collection. Experimental results from the implemented system calibration procedure are finally presented to verify the feasibility the proposed automated procedure. Qualitative and quantitative evaluation of the estimated system calibration parameters from two-calibration sessions is also presented to confirm the stability of the cameras' interior orientation and system mounting parameters.

A Thoracic Spine Segmentation Technique for Automatic Extraction of VHS and Cobb Angle from X-ray Images (X-ray 영상에서 VHS와 콥 각도 자동 추출을 위한 흉추 분할 기법)

  • Ye-Eun, Lee;Seung-Hwa, Han;Dong-Gyu, Lee;Ho-Joon, Kim
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.1
    • /
    • pp.51-58
    • /
    • 2023
  • In this paper, we propose an organ segmentation technique for the automatic extraction of medical diagnostic indicators from X-ray images. In order to calculate diagnostic indicators of heart disease and spinal disease such as VHS(vertebral heart scale) and Cobb angle, it is necessary to accurately segment the thoracic spine, carina, and heart in a chest X-ray image. A deep neural network model in which the high-resolution representation of the image for each layer and the structure converted into a low-resolution feature map are connected in parallel was adopted. This structure enables the relative position information in the image to be effectively reflected in the segmentation process. It is shown that learning performance can be improved by combining the OCR module, in which pixel information and object information are mutually interacted in a multi-step process, and the channel attention module, which allows each channel of the network to be reflected as different weight values. In addition, a method of augmenting learning data is presented in order to provide robust performance against changes in the position, shape, and size of the subject in the X-ray image. The effectiveness of the proposed theory was evaluated through an experiment using 145 human chest X-ray images and 118 animal X-ray images.