• Title/Summary/Keyword: Grid-Based Algorithm

Search Result 678, Processing Time 0.024 seconds

Development of Change Detection Technique Using Time Seriate Remotely Sensed Satellite Images with User Friendly GIS Interface (사용자 중심적 GIS 인터페이스를 이용한 시계열적 원격탐사 영상의 변화탐지 기법의 개발)

  • 양인태;한성만;윤희천;김흥규
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.22 no.2
    • /
    • pp.151-159
    • /
    • 2004
  • The diversity, expansion of human activity and rapid urbanization make modem society to faced with problems like damage of nature and drain of natural resources. Under these circumstances rapid and accurate change detection techniques, which can detect wide range utilization changes, are needed for efficient management and utilization plan of national territory. In this study to perform change detection from remote sensing images, space analysis technique contained in Geographic Information System is applied. And from this technique, the software. that can execute new change detection algorithm, query, inquiry and analysis, is produced. This software is on the basis of graphic user interface and has many functions such as format conversion, grid calculation, statistical processing, display and reference. In this study, simultaneously change detection for multi-temporal satellite images can be performed and integrated one change image about four different periods was produced. Further more software user can acquire land cover change information for an specific area through querying and questioning about yearly changes. Finally making of every application module for change detection into one window based visual basic program, can be produced user convenience and automatic performances.

Design and Implementation of Ethereum-based Future Power Trading System (이더리움 기반의 선물(Future) 전력 거래 시스템 설계)

  • Youm, Sungkwan;Lee, Heekwon;Shin, Kwang-Seong
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.584-585
    • /
    • 2021
  • As the production of new and renewable energy such as solar and wind power has diversified, microgrid systems that can simultaneously produce and consume have been introduced. In general, a decrease in electricity prices through solar power is expected in summer, so producer protection is required. In this paper, we propose a transparent and safe gift power transaction system between users using blockchain in a microgrid environment. A futures is simply a contract in which the buyer is obligated to buy electricity or the seller is obliged to sell electricity at a fixed price and a predetermined futures price. This system proposes a futures trading algorithm that searches for futures prices and concludes power transactions with automated operations without user intervention by using a smart contract, a reliable executable code within the blockchain network. If a power producer thinks that the price during the peak production period is likely to decrease during production planning, it sells futures first in the futures market and buys back futures during the peak production period to make a profit in the spot market. losses can be compensated. In addition, if there is a risk that the price of electricity will rise when a sales contract is concluded, a broker can compensate for a loss in the spot market by first buying futures in the futures market and liquidating futures when the sales contract is fulfilled.

  • PDF

Predicting Crime Risky Area Using Machine Learning (머신러닝기반 범죄발생 위험지역 예측)

  • HEO, Sun-Young;KIM, Ju-Young;MOON, Tae-Heon
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.21 no.4
    • /
    • pp.64-80
    • /
    • 2018
  • In Korea, citizens can only know general information about crime. Thus it is difficult to know how much they are exposed to crime. If the police can predict the crime risky area, it will be possible to cope with the crime efficiently even though insufficient police and enforcement resources. However, there is no prediction system in Korea and the related researches are very much poor. From these backgrounds, the final goal of this study is to develop an automated crime prediction system. However, for the first step, we build a big data set which consists of local real crime information and urban physical or non-physical data. Then, we developed a crime prediction model through machine learning method. Finally, we assumed several possible scenarios and calculated the probability of crime and visualized the results in a map so as to increase the people's understanding. Among the factors affecting the crime occurrence revealed in previous and case studies, data was processed in the form of a big data for machine learning: real crime information, weather information (temperature, rainfall, wind speed, humidity, sunshine, insolation, snowfall, cloud cover) and local information (average building coverage, average floor area ratio, average building height, number of buildings, average appraised land value, average area of residential building, average number of ground floor). Among the supervised machine learning algorithms, the decision tree model, the random forest model, and the SVM model, which are known to be powerful and accurate in various fields were utilized to construct crime prevention model. As a result, decision tree model with the lowest RMSE was selected as an optimal prediction model. Based on this model, several scenarios were set for theft and violence cases which are the most frequent in the case city J, and the probability of crime was estimated by $250{\times}250m$ grid. As a result, we could find that the high crime risky area is occurring in three patterns in case city J. The probability of crime was divided into three classes and visualized in map by $250{\times}250m$ grid. Finally, we could develop a crime prediction model using machine learning algorithm and visualized the crime risky areas in a map which can recalculate the model and visualize the result simultaneously as time and urban conditions change.

Development of Sentiment Analysis Model for the hot topic detection of online stock forums (온라인 주식 포럼의 핫토픽 탐지를 위한 감성분석 모형의 개발)

  • Hong, Taeho;Lee, Taewon;Li, Jingjing
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.1
    • /
    • pp.187-204
    • /
    • 2016
  • Document classification based on emotional polarity has become a welcomed emerging task owing to the great explosion of data on the Web. In the big data age, there are too many information sources to refer to when making decisions. For example, when considering travel to a city, a person may search reviews from a search engine such as Google or social networking services (SNSs) such as blogs, Twitter, and Facebook. The emotional polarity of positive and negative reviews helps a user decide on whether or not to make a trip. Sentiment analysis of customer reviews has become an important research topic as datamining technology is widely accepted for text mining of the Web. Sentiment analysis has been used to classify documents through machine learning techniques, such as the decision tree, neural networks, and support vector machines (SVMs). is used to determine the attitude, position, and sensibility of people who write articles about various topics that are published on the Web. Regardless of the polarity of customer reviews, emotional reviews are very helpful materials for analyzing the opinions of customers through their reviews. Sentiment analysis helps with understanding what customers really want instantly through the help of automated text mining techniques. Sensitivity analysis utilizes text mining techniques on text on the Web to extract subjective information in the text for text analysis. Sensitivity analysis is utilized to determine the attitudes or positions of the person who wrote the article and presented their opinion about a particular topic. In this study, we developed a model that selects a hot topic from user posts at China's online stock forum by using the k-means algorithm and self-organizing map (SOM). In addition, we developed a detecting model to predict a hot topic by using machine learning techniques such as logit, the decision tree, and SVM. We employed sensitivity analysis to develop our model for the selection and detection of hot topics from China's online stock forum. The sensitivity analysis calculates a sentimental value from a document based on contrast and classification according to the polarity sentimental dictionary (positive or negative). The online stock forum was an attractive site because of its information about stock investment. Users post numerous texts about stock movement by analyzing the market according to government policy announcements, market reports, reports from research institutes on the economy, and even rumors. We divided the online forum's topics into 21 categories to utilize sentiment analysis. One hundred forty-four topics were selected among 21 categories at online forums about stock. The posts were crawled to build a positive and negative text database. We ultimately obtained 21,141 posts on 88 topics by preprocessing the text from March 2013 to February 2015. The interest index was defined to select the hot topics, and the k-means algorithm and SOM presented equivalent results with this data. We developed a decision tree model to detect hot topics with three algorithms: CHAID, CART, and C4.5. The results of CHAID were subpar compared to the others. We also employed SVM to detect the hot topics from negative data. The SVM models were trained with the radial basis function (RBF) kernel function by a grid search to detect the hot topics. The detection of hot topics by using sentiment analysis provides the latest trends and hot topics in the stock forum for investors so that they no longer need to search the vast amounts of information on the Web. Our proposed model is also helpful to rapidly determine customers' signals or attitudes towards government policy and firms' products and services.

A Simulation-Based Investigation of an Advanced Traveler Information System with V2V in Urban Network (시뮬레이션기법을 통한 차량 간 통신을 이용한 첨단교통정보시스템의 효과 분석 (도시 도로망을 중심으로))

  • Kim, Hoe-Kyoung
    • Journal of Korean Society of Transportation
    • /
    • v.29 no.5
    • /
    • pp.121-138
    • /
    • 2011
  • More affordable and available cutting-edge technologies (e.g., wireless vehicle communication) are regarded as a possible alternative to the fixed infrastructure-based traffic information system requiring the expensive infrastructure investments and mostly implemented in the uninterrupted freeway network with limited spatial system expansion. This paper develops an advanced decentralized traveler information System (ATIS) using vehicle-to-vehicle (V2V) communication system whose performance (drivers' travel time savings) are enhanced by three complementary functions (autonomous automatic incident detection algorithm, reliable sample size function, and driver behavior model) and evaluates it in the typical $6{\times}6$ urban grid network with non-recurrent traffic state (traffic incident) with the varying key parameters (traffic flow, communication radio range, and penetration ratio), employing the off-the-shelf microscopic simulation model (VISSIM) under the ideal vehicle communication environment. Simulation outputs indicate that as the three key parameters are increased more participating vehicles are involved for traffic data propagation in the less communication groups at the faster data dissemination speed. Also, participating vehicles saved their travel time by dynamically updating the up-to-date traffic states and searching for the new route. Focusing on the travel time difference of (instant) re-routing vehicles, lower traffic flow cases saved more time than higher traffic flow ones. This is because a relatively small number of vehicles in 300vph case re-route during the most system-efficient time period (the early time of the traffic incident) but more vehicles in 514vph case re-route during less system-efficient time period, even after the incident is resolved. Also, normally re-routings on the network-entering links saved more travel time than any other places inside the network except the case where the direct effect of traffic incident triggers vehicle re-routings during the effective incident time period and the location and direction of the incident link determines the spatial distribution of re-routing vehicles.

Evaluating efficiency of Coaxial MLC VMAT plan for spine SBRT (Spine SBRT 치료시 Coaxial MLC VMAT plan의 유용성 평가)

  • Son, Sang Jun;Mun, Jun Ki;Kim, Dae Ho;Yoo, Suk Hyun
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.26 no.2
    • /
    • pp.313-320
    • /
    • 2014
  • Purpose : The purpose of the study is to evaluate the efficiency of Coaxial MLC VMAT plan (Using $273^{\circ}$ and $350^{\circ}$ collimator angle) That the leaf motion direction aligned with axis of OAR (Organ at risk, It means spinal cord or cauda equine in this study.) compare to Universal MLC VMAT plan (using $30^{\circ}$ and $330^{\circ}$ collimator angle) for spine SBRT. Materials and Methods : The 10 cases of spine SBRT that treated with VMAT planned by Coaxial MLC and Varian TBX were enrolled. Those cases were planned by Eclipse (Ver. 10.0.42, Varian, USA), PRO3 (Progressive Resolution Optimizer 10.0.28) and AAA (Anisotropic Analytic Algorithm Ver. 10.0.28) with coplanar $360^{\circ}$ arcs and 10MV FFF (Flattening filter free). Each arc has $273^{\circ}$ and $350^{\circ}$ collimator angle, respectively. The Universal MLC VMAT plans are based on existing treatment plans. Those plans have the same parameters of existing treatment plans but collimator angle. To minimize the dose difference that shows up randomly on optimizing, all plans were optimized and calculated twice respectively. The calculation grid is 0.2 cm and all plans were normalized to the target V100%=90%. The indexes of evaluation are V10Gy, D0.03cc, Dmean of OAR (Organ at risk, It means spinal cord or cauda equine in this study.), H.I (Homogeneity index) of the target and total MU. All Coaxial VMAT plans were verified by gamma test with Mapcheck2 (Sun Nuclear Co., USA), Mapphan (Sun Nuclear Co., USA) and SNC patient (Sun Nuclear Co., USA Ver 6.1.2.18513). Results : The difference between the coaxial and the universal VMAT plans are follow. The coaxial VMAT plan is better in the V10Gy of OAR, Up to 4.1%, at least 0.4%, the average difference was 1.9% and In the D0.03cc of OAR, Up to 83.6 cGy, at least 2.2 cGy, the average difference was 33.3 cGy. In Dmean, Up to 34.8 cGy, at least -13.0 cGy, the average difference was 9.6 cGy that say the coaxial VMAT plans are better except few cases. H.I difference Up to 0.04, at least 0.01, the average difference was 0.02 and the difference of average total MU is 74.1 MU. The coaxial MLC VMAT plan is average 74.1 MU lesser then another. All IMRT verification gamma test results for the coaxial MLC VMAT plan passed over 90.0% at 1mm / 2%. Conclusion : Coaxial MLC VMAT treatment plan appeared to be favorable in most cases than the Universal MLC VMAT treatment planning. It is efficient in lowering the dose of the OAR V10Gy especially. As a result, the Coaxial MLC VMAT plan could be better than the Universal MLC VMAT plan in same MU.

Steel Plate Faults Diagnosis with S-MTS (S-MTS를 이용한 강판의 표면 결함 진단)

  • Kim, Joon-Young;Cha, Jae-Min;Shin, Junguk;Yeom, Choongsub
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.47-67
    • /
    • 2017
  • Steel plate faults is one of important factors to affect the quality and price of the steel plates. So far many steelmakers generally have used visual inspection method that could be based on an inspector's intuition or experience. Specifically, the inspector checks the steel plate faults by looking the surface of the steel plates. However, the accuracy of this method is critically low that it can cause errors above 30% in judgment. Therefore, accurate steel plate faults diagnosis system has been continuously required in the industry. In order to meet the needs, this study proposed a new steel plate faults diagnosis system using Simultaneous MTS (S-MTS), which is an advanced Mahalanobis Taguchi System (MTS) algorithm, to classify various surface defects of the steel plates. MTS has generally been used to solve binary classification problems in various fields, but MTS was not used for multiclass classification due to its low accuracy. The reason is that only one mahalanobis space is established in the MTS. In contrast, S-MTS is suitable for multi-class classification. That is, S-MTS establishes individual mahalanobis space for each class. 'Simultaneous' implies comparing mahalanobis distances at the same time. The proposed steel plate faults diagnosis system was developed in four main stages. In the first stage, after various reference groups and related variables are defined, data of the steel plate faults is collected and used to establish the individual mahalanobis space per the reference groups and construct the full measurement scale. In the second stage, the mahalanobis distances of test groups is calculated based on the established mahalanobis spaces of the reference groups. Then, appropriateness of the spaces is verified by examining the separability of the mahalanobis diatances. In the third stage, orthogonal arrays and Signal-to-Noise (SN) ratio of dynamic type are applied for variable optimization. Also, Overall SN ratio gain is derived from the SN ratio and SN ratio gain. If the derived overall SN ratio gain is negative, it means that the variable should be removed. However, the variable with the positive gain may be considered as worth keeping. Finally, in the fourth stage, the measurement scale that is composed of selected useful variables is reconstructed. Next, an experimental test should be implemented to verify the ability of multi-class classification and thus the accuracy of the classification is acquired. If the accuracy is acceptable, this diagnosis system can be used for future applications. Also, this study compared the accuracy of the proposed steel plate faults diagnosis system with that of other popular classification algorithms including Decision Tree, Multi Perception Neural Network (MLPNN), Logistic Regression (LR), Support Vector Machine (SVM), Tree Bagger Random Forest, Grid Search (GS), Genetic Algorithm (GA) and Particle Swarm Optimization (PSO). The steel plates faults dataset used in the study is taken from the University of California at Irvine (UCI) machine learning repository. As a result, the proposed steel plate faults diagnosis system based on S-MTS shows 90.79% of classification accuracy. The accuracy of the proposed diagnosis system is 6-27% higher than MLPNN, LR, GS, GA and PSO. Based on the fact that the accuracy of commercial systems is only about 75-80%, it means that the proposed system has enough classification performance to be applied in the industry. In addition, the proposed system can reduce the number of measurement sensors that are installed in the fields because of variable optimization process. These results show that the proposed system not only can have a good ability on the steel plate faults diagnosis but also reduce operation and maintenance cost. For our future work, it will be applied in the fields to validate actual effectiveness of the proposed system and plan to improve the accuracy based on the results.

The Relationship Analysis between the Epicenter and Lineaments in the Odaesan Area using Satellite Images and Shaded Relief Maps (위성영상과 음영기복도를 이용한 오대산 지역 진앙의 위치와 선구조선의 관계 분석)

  • CHA, Sung-Eun;CHI, Kwang-Hoon;JO, Hyun-Woo;KIM, Eun-Ji;LEE, Woo-Kyun
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.19 no.3
    • /
    • pp.61-74
    • /
    • 2016
  • The purpose of this paper is to analyze the relationship between the location of the epicenter of a medium-sized earthquake(magnitude 4.8) that occurred on January 20, 2007 in the Odaesan area with lineament features using a shaded relief map(1/25,000 scale) and satellite images from LANDSAT-8 and KOMPSAT-2. Previous studies have analyzed lineament features in tectonic settings primarily by examining two-dimensional satellite images and shaded relief maps. These methods, however, limit the application of the visual interpretation of relief features long considered as the major component of lineament extraction. To overcome some existing limitations of two-dimensional images, this study examined three-dimensional images, produced from a Digital Elevation Model and drainage network map, for lineament extraction. This approach reduces mapping errors introduced by visual interpretation. In addition, spline interpolation was conducted to produce density maps of lineament frequency, intersection, and length required to estimate the density of lineament at the epicenter of the earthquake. An algorithm was developed to compute the Value of the Relative Density(VRD) representing the relative density of lineament from the map. The VRD is the lineament density of each map grid divided by the maximum density value from the map. As such, it is a quantified value that indicates the concentration level of the lineament density across the area impacted by the earthquake. Using this algorithm, the VRD calculated at the earthquake epicenter using the lineament's frequency, intersection, and length density maps ranged from approximately 0.60(min) to 0.90(max). However, because there were differences in mapped images such as those for solar altitude and azimuth, the mean of VRD was used rather than those categorized by the images. The results show that the average frequency of VRD was approximately 0.85, which was 21% higher than the intersection and length of VRD, demonstrating the close relationship that exists between lineament and the epicenter. Therefore, it is concluded that the density map analysis described in this study, based on lineament extraction, is valid and can be used as a primary data analysis tool for earthquake research in the future.