• Title/Summary/Keyword: Device Feature Extraction

Search Result 46, Processing Time 0.024 seconds

Accuracy Estimation of Electro-optical Camera (EOC) on KOMPSAT-1

  • Park, Woon-Yong;Hong, Sun-Houn;Song, Youn-Kyung
    • Korean Journal of Geomatics
    • /
    • v.2 no.1
    • /
    • pp.47-55
    • /
    • 2002
  • Remote sensing is the science and art of obtaining information about an object, area or phenomenon through the analysis of data acquired by a device that is not in contact with the object, area, or phenomenon under investigation./sup 1)/ EOC (Electro -Optical Camera) sensor loaded on the KOMPSAT-1 (Korea Multi- Purpose Satellite-1) performs the earth remote sensing operation. EOC can get high-resolution images of ground distance 6.6m during photographing; it is possible to get a tilt image by tilting satellite body up to 45 degrees at maximum. Accordingly, the device developed in this study enables to obtain images by photographing one pair of tilt image for the same point from two different planes. KOMPSAT-1 aims to obtain a Korean map with a scale of 1:25,000 with high resolution. The KOMPSAT-1 developed automated feature extraction system based on stereo satellite image. It overcomes the limitations of sensor and difficulties associated with preprocessing quite effectively. In case of using 6, 7 and 9 ground control points, which are evenly spread in image, with 95% of reliability for horizontal and vertical position, 3-dimensional positioning was available with accuracy of 6.0752m and 9.8274m. Therefore, less than l0m of design accuracy in KOMPSAT-1 was achieved. Also the ground position error of ortho-image, with reliability of 95%, is 17.568m. And elevation error showing 36.82m was enhanced. The reason why elevation accuracy was not good compared with the positioning accuracy used stereo image was analyzed as a problem of image matching system. Ortho-image system is advantageous if accurate altitude and production of digital elevation model are desired. The Korean map drawn on a scale of 1: 25,000 by using the new technique of KOMPSAT-1 EOC image adopted in the present study produces accurate result compared to existing mapping techniques involving high costs with less efficiency.

  • PDF

Mobile Camera-Based Positioning Method by Applying Landmark Corner Extraction (랜드마크 코너 추출을 적용한 모바일 카메라 기반 위치결정 기법)

  • Yoo Jin Lee;Wansang Yoon;Sooahm Rhee
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.6_1
    • /
    • pp.1309-1320
    • /
    • 2023
  • The technological development and popularization of mobile devices have developed so that users can check their location anywhere and use the Internet. However, in the case of indoors, the Internet can be used smoothly, but the global positioning system (GPS) function is difficult to use. There is an increasing need to provide real-time location information in shaded areas where GPS is not received, such as department stores, museums, conference halls, schools, and tunnels, which are indoor public places. Accordingly, research on the recent indoor positioning technology based on light detection and ranging (LiDAR) equipment is increasing to build a landmark database. Focusing on the accessibility of building a landmark database, this study attempted to develop a technique for estimating the user's location by using a single image taken of a landmark based on a mobile device and the landmark database information constructed in advance. First, a landmark database was constructed. In order to estimate the user's location only with the mobile image photographing the landmark, it is essential to detect the landmark from the mobile image, and to acquire the ground coordinates of the points with fixed characteristics from the detected landmark. In the second step, by applying the bag of words (BoW) image search technology, the landmark photographed by the mobile image among the landmark database was searched up to a similar 4th place. In the third step, one of the four candidate landmarks searched through the scale invariant feature transform (SIFT) feature point extraction technique and Homography random sample consensus(RANSAC) was selected, and at this time, filtering was performed once more based on the number of matching points through threshold setting. In the fourth step, the landmark image was projected onto the mobile image through the Homography matrix between the corresponding landmark and the mobile image to detect the area of the landmark and the corner. Finally, the user's location was estimated through the location estimation technique. As a result of analyzing the performance of the technology, the landmark search performance was measured to be about 86%. As a result of comparing the location estimation result with the user's actual ground coordinate, it was confirmed that it had a horizontal location accuracy of about 0.56 m, and it was confirmed that the user's location could be estimated with a mobile image by constructing a landmark database without separate expensive equipment.

Study on Structure Visual Inspection Technology using Drones and Image Analysis Techniques (드론과 이미지 분석기법을 활용한 구조물 외관점검 기술 연구)

  • Kim, Jong-Woo;Jung, Young-Woo;Rhim, Hong-Chul
    • Journal of the Korea Institute of Building Construction
    • /
    • v.17 no.6
    • /
    • pp.545-557
    • /
    • 2017
  • The study is about the efficient alternative to concrete surface in the field of visual inspection technology for deteriorated infrastructure. By combining industrial drones and deep learning based image analysis techniques with traditional visual inspection and research, we tried to reduce manpowers, time requirements and costs, and to overcome the height and dome structures. On board device mounted on drones is consisting of a high resolution camera for detecting cracks of more than 0.3 mm, a lidar sensor and a embeded image processor module. It was mounted on an industrial drones, took sample images of damage from the site specimen through automatic flight navigation. In addition, the damege parts of the site specimen was used to measure not only the width and length of cracks but white rust also, and tried up compare them with the final image analysis detected results. Using the image analysis techniques, the damages of 54ea sample images were analyzed by the segmentation - feature extraction - decision making process, and extracted the analysis parameters using supervised mode of the deep learning platform. The image analysis of newly added non-supervised 60ea image samples was performed based on the extracted parameters. The result presented in 90.5 % of the damage detection rate.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.

A Study on Stroke Extraction for Handwritten Korean Character Recognition (필기체 한글 문자 인식을 위한 획 추출에 관한 연구)

  • Choi, Young-Kyoo;Rhee, Sang-Burm
    • The KIPS Transactions:PartB
    • /
    • v.9B no.3
    • /
    • pp.375-382
    • /
    • 2002
  • Handwritten character recognition is classified into on-line handwritten character recognition and off-line handwritten character recognition. On-line handwritten character recognition has made a remarkable outcome compared to off-line hacdwritten character recognition. This method can acquire the dynamic written information such as the writing order and the position of a stroke by means of pen-based electronic input device such as a tablet board. On the contrary, Any dynamic information can not be acquired in off-line handwritten character recognition since there are extreme overlapping between consonants and vowels, and heavily noisy images between strokes, which change the recognition performance with the result of the preprocessing. This paper proposes a method that effectively extracts the stroke including dynamic information of characters for off-line Korean handwritten character recognition. First of all, this method makes improvement and binarization of input handwritten character image as preprocessing procedure using watershed algorithm. The next procedure is extraction of skeleton by using the transformed Lu and Wang's thinning: algorithm, and segment pixel array is extracted by abstracting the feature point of the characters. Then, the vectorization is executed with a maximum permission error method. In the case that a few strokes are bound in a segment, a segment pixel array is divided with two or more segment vectors. In order to reconstruct the extracted segment vector with a complete stroke, the directional component of the vector is mortified by using right-hand writing coordinate system. With combination of segment vectors which are adjacent and can be combined, the reconstruction of complete stroke is made out which is suitable for character recognition. As experimentation, it is verified that the proposed method is suitable for handwritten Korean character recognition.

Using Skeleton Vector Information and RNN Learning Behavior Recognition Algorithm (스켈레톤 벡터 정보와 RNN 학습을 이용한 행동인식 알고리즘)

  • Kim, Mi-Kyung;Cha, Eui-Young
    • Journal of Broadcast Engineering
    • /
    • v.23 no.5
    • /
    • pp.598-605
    • /
    • 2018
  • Behavior awareness is a technology that recognizes human behavior through data and can be used in applications such as risk behavior through video surveillance systems. Conventional behavior recognition algorithms have been performed using the 2D camera image device or multi-mode sensor or multi-view or 3D equipment. When two-dimensional data was used, the recognition rate was low in the behavior recognition of the three-dimensional space, and other methods were difficult due to the complicated equipment configuration and the expensive additional equipment. In this paper, we propose a method of recognizing human behavior using only CCTV images without additional equipment using only RGB and depth information. First, the skeleton extraction algorithm is applied to extract points of joints and body parts. We apply the equations to transform the vector including the displacement vector and the relational vector, and study the continuous vector data through the RNN model. As a result of applying the learned model to various data sets and confirming the accuracy of the behavior recognition, the performance similar to that of the existing algorithm using the 3D information can be verified only by the 2D information.

Face Recognition Under Ubiquitous Environments (유비쿼터스 환경을 이용한 얼굴인식)

  • Go, Hyoun-Joo;Kim, Hyung-Bae;Yang, Dong-Hwa;Park, Jang-Hwan;Chun, Myung-Geun
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.4
    • /
    • pp.431-437
    • /
    • 2004
  • This paper propose a facial recognition method based on an ubiquitous computing that is one of next generation intelligence technology fields. The facial images are acquired by a mobile device so-called cellular phone camera. We consider a mobile security using facial feature extraction and recognition process. Facial recognition is performed by the PCA and fuzzy LDA algorithm. Applying the discrete wavelet based on multi-resolution analysis, we compress the image data for mobile system environment. Euclidean metric is applied to measure the similarity among acquired features and then obtain the recognition rate. Finally we use the mobile equipment to show the efficiency of method. From various experiments, we find that our proposed method shows better results, even though the resolution of mobile camera is lower than conventional camera.

Activity Recognition of Workers and Passengers onboard Ships Using Multimodal Sensors in a Smartphone (선박 탑승자를 위한 다중 센서 기반의 스마트폰을 이용한 활동 인식 시스템)

  • Piyare, Rajeev Kumar;Lee, Seong Ro
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.39C no.9
    • /
    • pp.811-819
    • /
    • 2014
  • Activity recognition is a key component in identifying the context of a user for providing services based on the application such as medical, entertainment and tactical scenarios. Instead of applying numerous sensor devices, as observed in many previous investigations, we are proposing the use of smartphone with its built-in multimodal sensors as an unobtrusive sensor device for recognition of six physical daily activities. As an improvement to previous works, accelerometer, gyroscope and magnetometer data are fused to recognize activities more reliably. The evaluation indicates that the IBK classifier using window size of 2s with 50% overlapping yields the highest accuracy (i.e., up to 99.33%). To achieve this peak accuracy, simple time-domain and frequency-domain features were extracted from raw sensor data of the smartphone.

A Survey on Deep Learning based Face Recognition for User Authentication (사용자 인증을 위한 딥러닝 기반 얼굴인식 기술 동향)

  • Mun, Hyung-Jin;Kim, Gea-Hee
    • Journal of Industrial Convergence
    • /
    • v.17 no.3
    • /
    • pp.23-29
    • /
    • 2019
  • Object recognition distinguish objects which are different from each other. But Face recognition distinguishes Identity of Faces with Similar Patterns. Feature extraction algorithm such as LBP, HOG, Gabor is being replaced with Deep Learning. As the technology that identify individual face with machine learning using Deep Learning Technology is developing, The Face Recognition Technology is being used in various field. In particular, the technology can provide individual and detailed service by being used in various offline environments requiring user identification, such as Smart Mirror. Face Recognition Technology can be developed as the technology that authenticate user easily by device like Smart Mirror and provide service authenticated user. In this paper, we present investigation about Face Recognition among various techniques for user authentication and analysis of Python source case of Face recognition and possibility of various service using Face Recognition Technology.

Simultaneous monitoring of motion ECG of two subjects using Bluetooth Piconet and baseline drift

  • Dave, Tejal;Pandya, Utpal
    • Biomedical Engineering Letters
    • /
    • v.8 no.4
    • /
    • pp.365-371
    • /
    • 2018
  • Uninterrupted monitoring of multiple subjects is required for mass causality events, in hospital environment or for sports by medical technicians or physicians. Movement of subjects under monitoring requires such system to be wireless, sometimes demands multiple transmitters and a receiver as a base station and monitored parameter must not be corrupted by any noise before further diagnosis. A Bluetooth Piconet network is visualized, where each subject carries a Bluetooth transmitter module that acquires vital sign continuously and relays to Bluetooth enabled device where, further signal processing is done. In this paper, a wireless network is realized to capture ECG of two subjects performing different activities like cycling, jogging, staircase climbing at 100 Hz frequency using prototyped Bluetooth module. The paper demonstrates removal of baseline drift using Fast Fourier Transform and Inverse Fast Fourier Transform and removal of high frequency noise using moving average and S-Golay algorithm. Experimental results highlight the efficacy of the proposed work to monitor any vital sign parameters of multiple subjects simultaneously. The importance of removing baseline drift before high frequency noise removal is shown using experimental results. It is possible to use Bluetooth Piconet frame work to capture ECG simultaneously for more than two subjects. For the applications where there will be larger body movement, baseline drift removal is a major concern and hence along with wireless transmission issues, baseline drift removal before high frequency noise removal is necessary for further feature extraction.