• Title/Summary/Keyword: Image preprocessing

Search Result 699, Processing Time 0.033 seconds

Improvement of Underground Cavity and Structure Detection Performance Through Machine Learning-based Diffraction Separation of GPR Data (기계학습 기반 회절파 분리 적용을 통한 GPR 탐사 자료의 도로 하부 공동 및 구조물 탐지 성능 향상)

  • Sooyoon Kim;Joongmoo Byun
    • Geophysics and Geophysical Exploration
    • /
    • v.26 no.4
    • /
    • pp.171-184
    • /
    • 2023
  • Machine learning (ML)-based cavity detection using a large amount of survey data obtained from vehicle-mounted ground penetrating radar (GPR) has been actively studied to identify underground cavities. However, only simple image processing techniques have been used for preprocessing the ML input, and many conventional seismic and GPR data processing techniques, which have been used for decades, have not been fully exploited. In this study, based on the idea that a cavity can be identified using diffraction, we applied ML-based diffraction separation to GPR data to increase the accuracy of cavity detection using the YOLO v5 model. The original ML-based seismic diffraction separation technique was modified, and the separated diffraction image was used as the input to train the cavity detection model. The performance of the proposed method was verified using public GPR data released by the Seoul Metropolitan Government. Underground cavities and objects were more accurately detected using separated diffraction images. In the future, the proposed method can be useful in various fields in which GPR surveys are used.

Estimation of Rice Heading Date of Paddy Rice from Slanted and Top-view Images Using Deep Learning Classification Model (딥 러닝 분류 모델을 이용한 직하방과 경사각 영상 기반의 벼 출수기 판별)

  • Hyeok-jin Bak;Wan-Gyu Sang;Sungyul Chang;Dongwon Kwon;Woo-jin Im;Ji-hyeon Lee;Nam-jin Chung;Jung-Il Cho
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.25 no.4
    • /
    • pp.337-345
    • /
    • 2023
  • Estimating the rice heading date is one of the most crucial agricultural tasks related to productivity. However, due to abnormal climates around the world, it is becoming increasingly challenging to estimate the rice heading date. Therefore, a more objective classification method for estimating the rice heading date is needed than the existing methods. This study, we aimed to classify the rice heading stage from various images using a CNN classification model. We collected top-view images taken from a drone and a phenotyping tower, as well as slanted-view images captured with a RGB camera. The collected images underwent preprocessing to prepare them as input data for the CNN model. The CNN architectures employed were ResNet50, InceptionV3, and VGG19, which are commonly used in image classification models. The accuracy of the models all showed an accuracy of 0.98 or higher regardless of each architecture and type of image. We also used Grad-CAM to visually check which features of the image the model looked at and classified. Then verified our model accurately measure the rice heading date in paddy fields. The rice heading date was estimated to be approximately one day apart on average in the four paddy fields. This method suggests that the water head can be estimated automatically and quantitatively when estimating the rice heading date from various paddy field monitoring images.

A Comparative Study on the Effective Deep Learning for Fingerprint Recognition with Scar and Wrinkle (상처와 주름이 있는 지문 판별에 효율적인 심층 학습 비교연구)

  • Kim, JunSeob;Rim, BeanBonyka;Sung, Nak-Jun;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.21 no.4
    • /
    • pp.17-23
    • /
    • 2020
  • Biometric information indicating measurement items related to human characteristics has attracted great attention as security technology with high reliability since there is no fear of theft or loss. Among these biometric information, fingerprints are mainly used in fields such as identity verification and identification. If there is a problem such as a wound, wrinkle, or moisture that is difficult to authenticate to the fingerprint image when identifying the identity, the fingerprint expert can identify the problem with the fingerprint directly through the preprocessing step, and apply the image processing algorithm appropriate to the problem. Solve the problem. In this case, by implementing artificial intelligence software that distinguishes fingerprint images with cuts and wrinkles on the fingerprint, it is easy to check whether there are cuts or wrinkles, and by selecting an appropriate algorithm, the fingerprint image can be easily improved. In this study, we developed a total of 17,080 fingerprint databases by acquiring all finger prints of 1,010 students from the Royal University of Cambodia, 600 Sokoto open data sets, and 98 Korean students. In order to determine if there are any injuries or wrinkles in the built database, criteria were established, and the data were validated by experts. The training and test datasets consisted of Cambodian data and Sokoto data, and the ratio was set to 8: 2. The data of 98 Korean students were set up as a validation data set. Using the constructed data set, five CNN-based architectures such as Classic CNN, AlexNet, VGG-16, Resnet50, and Yolo v3 were implemented. A study was conducted to find the model that performed best on the readings. Among the five architectures, ResNet50 showed the best performance with 81.51%.

Robust Eye Localization using Multi-Scale Gabor Feature Vectors (다중 해상도 가버 특징 벡터를 이용한 강인한 눈 검출)

  • Kim, Sang-Hoon;Jung, Sou-Hwan;Cho, Seong-Won;Chung, Sun-Tae
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.45 no.1
    • /
    • pp.25-36
    • /
    • 2008
  • Eye localization means localization of the center of the pupils, and is necessary for face recognition and related applications. Most of eye localization methods reported so far still need to be improved about robustness as well as precision for successful applications. In this paper, we propose a robust eye localization method using multi-scale Gabor feature vectors without big computational burden. The eye localization method using Gabor feature vectors is already employed in fuck as EBGM, but the method employed in EBGM is known not to be robust with respect to initial values, illumination, and pose, and may need extensive search range for achieving the required performance, which may cause big computational burden. The proposed method utilizes multi-scale approach. The proposed method first tries to localize eyes in the lower resolution face image by utilizing Gabor Jet similarity between Gabor feature vector at an estimated initial eye coordinates and the Gabor feature vectors in the eye model of the corresponding scale. Then the method localizes eyes in the next scale resolution face image in the same way but with initial eye points estimated from the eye coordinates localized in the lower resolution images. After repeating this process in the same way recursively, the proposed method funally localizes eyes in the original resolution face image. Also, the proposed method provides an effective illumination normalization to make the proposed multi-scale approach more robust to illumination, and additionally applies the illumination normalization technique in the preprocessing stage of the multi-scale approach so that the proposed method enhances the eye detection success rate. Experiment results verify that the proposed eye localization method improves the precision rate without causing big computational overhead compared to other eye localization methods reported in the previous researches and is robust to the variation of post: and illumination.

Multiresolution 4- 8 Tile Hierarchy Construction for Realtime Visualization of Planetary Scale Geological Information (행성 규모 지리 정보의 실시간 시각화를 위한 다계층 4-8 타일 구조의 구축)

  • Jin, Jong-Wook;Wohn, Kwang-Yun
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.9 no.4
    • /
    • pp.12-21
    • /
    • 2006
  • Recently, Very large and high resolution geological data from aerial or satellite imagery are available. Many researches and applications require to do realtime visualization of interest geological area or entire planet. Important operation of wide-spreaded terrain realtime visualization technique is the appropriate model resolution selection from pre-processed multi-resolution model hierarchy depend upon participant's view. For embodying such realtime rendering system with large geometric data, Preprocessing multi-resolution hierarchy from large scale geological information of interest area is required. In this research, recent Cubic multiresolution 4-8 tile hierarchy is selected for global planetary applications. Based upon the tile hierarchy, It constructs the selective terminal level tile mesh for original geological information area and starts to sample individual generated tiles for terminal level tiles. It completes the hierarchy by constructing intermediate tiles with low pass filtering in bottom-up direction. This research embodies series of efficient cubic 4-8 tile hierarchy construction mechanism with out-of-core storage. The planetary scale Mars' geographical altitude data and image data were selected for the experiment.

  • PDF

Stereo Matching For Satellite Images using The Classified Terrain Information (지형식별정보를 이용한 입체위성영상매칭)

  • Bang, Soo-Nam;Cho, Bong-Whan
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.4 no.1 s.6
    • /
    • pp.93-102
    • /
    • 1996
  • For an atomatic generation of DEM(Digital Elevation Model) by computer, it is a time-consumed work to determine adquate matches from stereo images. Correlation and evenly distributed area-based method is generally used for matching operation. In this paper, we propose a new approach that computes matches efficiantly by changing the size of mask window and search area according to the given terrain information. For image segmentation, at first edge-preserving smoothing filter is used for preprocessing, and then region growing algorithm is applied for the filterd images. The segmented regions are classifed into mountain, plain and water area by using MRF(Markov Random Filed) model. Maching is composed of predicting parallex and fine matching. Predicted parallex determines the location of search area in fine matching stage. The size of search area and mask window is determined by terrain information for each pixel. The execution time of matching is reduced by lessening the size of search area in the case of plain and water. For the experiments, four images which are covered $10km{\times}10km(1024{\times}1024\;pixel)$ of Taejeon-Kumsan in each are studied. The result of this study shows that the computing time of the proposed method using terrain information for matching operation can be reduced from 25% to 35%.

  • PDF

Directional Feature Extraction of Handwritten Numerals using Local min/max Operations (Local min/max 연산을 이용한 필기체 숫자의 방향특징 추출)

  • Jung, Soon-Won;Park, Joong-Jo
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.10 no.1
    • /
    • pp.7-12
    • /
    • 2009
  • In this paper, we propose a directional feature extraction method for off-line handwritten numerals by using the morphological operations. Direction features are obtained from four directional line images, each of which contains horizontal, vertical, right-diagonal and left-diagonal lines in entire numeral lines. Conventional method for extracting directional features uses Kirsch masks which generate edge-shaped double line images for each direction, whereas our method uses directional erosion operations and generate single line images for each direction. To apply these directional erosion operations to the numeral image, preprocessing steps such as thinning and dilation are required, but resultant directional lines are more similar to numeral lines themselves. Our four [$4{\times}4$] directional features of a numeral are obtained from four directional line images through a zoning method. For obtaining the higher recognition rates of the handwrittern numerals, we use the multiple feature which is comprised of our proposed feature and the conventional features of a kirsch directional feature and a concavity feature. For recognition test with given features, we use a multi-layer perceptron neural network classifier which is trained with the back propagation algorithm. Through the experiments with the CENPARMI numeral database of Concordia University, we have achieved a recognition rate of 98.35%.

  • PDF

Effective Wavefield Separation of Reflected P- and PS-Waves in Multicomponent Seismic Data by Using Rotation Transform with Stacking (다성분 탄성파탐사자료에서 회전 변환과 중합을 이용한 효과적인 P파 반사파와 PS파 반사파의 분리)

  • Jeong, Soocheol;Byun, Joongmoo;Seol, Soon Jee
    • Geophysics and Geophysical Exploration
    • /
    • v.16 no.1
    • /
    • pp.6-17
    • /
    • 2013
  • Multicomponent seismic data including both P- and PS-waves have advantages in discriminating the type of pore fluid, characterizing the lithologic attributes and producing the high resolution image. However, multicomponent seismic data recorded at the vertical and horizontal component receivers contain both P- and PS-waves which have different features, simultaneously. Therefore, the wavefield separation of P- and PS-waves as a preprocessing is inevitable in order to use the multicomponent seismic data successfully. In this study, we analyzed the previous study of the wavefield separation method suggested by Jeong and Byun in 2011, where the approximated reflection angle calculated only from one refernce depth is used in rotation transform, and showed its limitation for seismic data containing various reflected events from the multi-layered structure. In order to overcome its limitation, we suggested a new effective wavefield separation method of P- and PS-waves. In new method, we calculate the reflection angles with various reference depths and apply rotation transforms to the data with those reflection angles. Then we stack all results to obtain the final separated data. To verify our new method, we applied it to the synthetic data sets from a multi-layered model, a fault model, and the Marmousi-2 model. The results showed that the proposed method separated successfully P- and PS-reflection events from the multicomponent data from mild dipping layered model as long as the dip is not too steep.

Current Status of Hyperspectral Data Processing Techniques for Monitoring Coastal Waters (연안해역 모니터링을 위한 초분광영상 처리기법 현황)

  • Kim, Sun-Hwa;Yang, Chan-Su
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.18 no.1
    • /
    • pp.48-63
    • /
    • 2015
  • In this study, we introduce various hyperspectral data processing techniques for the monitoring of shallow and coastal waters to enlarge the application range and to improve the accuracy of the end results in Korea. Unlike land, more accurate atmospheric correction is needed in coastal region showing relatively low reflectance in visible wavelengths. Sun-glint which occurs due to a geometry of sun-sea surface-sensor is another issue for the data processing in the ocean application of hyperspectal imagery. After the preprocessing of the hyperspectral data, a semi-analytical algorithm based on a radiative transfer model and a spectral library can be used for bathymetry mapping in coastal area, type classification and status monitoring of benthos or substrate classification. In general, semi-analytical algorithms using spectral information obtained from hyperspectral imagey shows higher accuracy than an empirical method using multispectral data. The water depth and quality are constraint factors in the ocean application of optical data. Although a radiative transfer model suggests the theoretical limit of about 25m in depth for bathymetry and bottom classification, hyperspectral data have been used practically at depths of up to 10 m in shallow and coastal waters. It means we have to focus on the maximum depth of water and water quality conditions that affect the coastal applicability of hyperspectral data, and to define the spectral library of coastal waters to classify the types of benthos and substrates.

Fat Client-Based Abstraction Model of Unstructured Data for Context-Aware Service in Edge Computing Environment (에지 컴퓨팅 환경에서의 상황인지 서비스를 위한 팻 클라이언트 기반 비정형 데이터 추상화 방법)

  • Kim, Do Hyung;Mun, Jong Hyeok;Park, Yoo Sang;Choi, Jong Sun;Choi, Jae Young
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.10 no.3
    • /
    • pp.59-70
    • /
    • 2021
  • With the recent advancements in the Internet of Things, context-aware system that provides customized services become important to consider. The existing context-aware systems analyze data generated around the user and abstract the context information that expresses the state of situations. However, these datasets is mostly unstructured and have difficulty in processing with simple approaches. Therefore, providing context-aware services using the datasets should be managed in simplified method. One of examples that should be considered as the unstructured datasets is a deep learning application. Processes in deep learning applications have a strong coupling in a way of abstracting dataset from the acquisition to analysis phases, it has less flexible when the target analysis model or applications are modified in functional scalability. Therefore, an abstraction model that separates the phases and process the unstructured dataset for analysis is proposed. The proposed abstraction utilizes a description name Analysis Model Description Language(AMDL) to deploy the analysis phases by each fat client is a specifically designed instance for resource-oriented tasks in edge computing environments how to handle different analysis applications and its factors using the AMDL and Fat client profiles. The experiment shows functional scalability through examples of AMDL and Fat client profiles targeting a vehicle image recognition model for vehicle access control notification service, and conducts process-by-process monitoring for collection-preprocessing-analysis of unstructured data.