• Title/Summary/Keyword: Heuristics for $A^*$ algorithm

Search Result 204, Processing Time 0.019 seconds

Efficient Construction of Euclidean Steiner Minimum Tree Using Combination of Delaunay Triangulation and Minimum Spanning Tree (들로네 삼각망과 최소신장트리를 결합한 효율적인 유클리드 스타이너 최소트리 생성)

  • Kim, Inbum
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.1
    • /
    • pp.57-64
    • /
    • 2014
  • As Steiner minimum tree building belongs to NP-Complete problem domain, heuristics for the problem ask for immense amount execution time and computations in numerous inputs. In this paper, we propose an efficient mechanism of euclidean Steiner minimum tree construction for numerous inputs using combination of Delaunay triangulation and Prim's minimum spanning tree algorithm. Trees built by proposed mechanism are compared respectively with the Prim's minimum spanning tree and minimums spanning tree based Steiner minimum tree. For 30,000 input nodes, Steiner minimum tree by proposed mechanism shows about 2.1% tree length less and 138.2% execution time more than minimum spanning tree, and does about 0.013% tree length less and 18.9% execution time less than minimum spanning tree based Steiner minimum tree in experimental results. Therefore the proposed mechanism can work moderately well to many useful applications where execution time is not critical but reduction of tree length is a key factor.

A case study on the mathematical problem solving performance of simultaneous equations for the students from a remedial course (특별보충과정 학생들의 문제해결수행에 대한 사례연구)

  • Ko, Sang-Sook;Lee, Sang-Hui
    • Journal of the Korean School Mathematics Society
    • /
    • v.9 no.1
    • /
    • pp.105-120
    • /
    • 2006
  • The Seventh Curriculum makes sure that those students who don't have a proper understanding of contents required at a certain stage take a remedial course. But a trend contrary to the intention is formed since there is no systematic education for such a course and thus more students get to fall into the group of low achievement. In particular, solving a simultaneous equation in a rote way without understanding influences negatively students' achievement. Schoenfeld introduced the basic elements of one's own mathematical problem solving process and behavior, referred to Polya's. Employing Schoenfeld's strategy, this study aimed to induce students' active participation in math classes, as well as to focus on a mathematical problem solving process during the study. Two students were selected from a remedial course at 00 Middle School and administered with a qualitative case study method over 17 lessons, each of which lasted for 30 minutes. In the beginning, they used such knowledge as facts and definitions a lot. There was a tendency of their resorting to intuitive knowledge more when they lacked basic knowledge or met with a difficult question. As the lessons were given, however, they improved their ability to implement algorithm procedures and used more familiar ones with the developed common procedures in the area of resources.

  • PDF

Optimizing Similarity Threshold and Coverage of CBR (사례기반추론의 유사 임계치 및 커버리지 최적화)

  • Ahn, Hyunchul
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.8
    • /
    • pp.535-542
    • /
    • 2013
  • Since case-based reasoning(CBR) has many advantages, it has been used for supporting decision making in various areas including medical checkup, production planning, customer classification, and so on. However, there are several factors to be set by heuristics when designing effective CBR systems. Among these factors, this study addresses the issue of selecting appropriate neighbors in case retrieval step. As the criterion for selecting appropriate neighbors, conventional studies have used the preset number of neighbors to combine(i.e. k of k-nearest neighbor), or the relative portion of the maximum similarity. However, this study proposes to use the absolute similarity threshold varying from 0 to 1, as the criterion for selecting appropriate neighbors to combine. In this case, too small similarity threshold value may make the model rarely produce the solution. To avoid this, we propose to adopt the coverage, which implies the ratio of the cases in which solutions are produced over the total number of the training cases, and to set it as the constraint when optimizing the similarity threshold. To validate the usefulness of the proposed model, we applied it to a real-world target marketing case of an online shopping mall in Korea. As a result, we found that the proposed model might significantly improve the performance of CBR.

Optimization of Support Vector Machines for Financial Forecasting (재무예측을 위한 Support Vector Machine의 최적화)

  • Kim, Kyoung-Jae;Ahn, Hyun-Chul
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.241-254
    • /
    • 2011
  • Financial time-series forecasting is one of the most important issues because it is essential for the risk management of financial institutions. Therefore, researchers have tried to forecast financial time-series using various data mining techniques such as regression, artificial neural networks, decision trees, k-nearest neighbor etc. Recently, support vector machines (SVMs) are popularly applied to this research area because they have advantages that they don't require huge training data and have low possibility of overfitting. However, a user must determine several design factors by heuristics in order to use SVM. For example, the selection of appropriate kernel function and its parameters and proper feature subset selection are major design factors of SVM. Other than these factors, the proper selection of instance subset may also improve the forecasting performance of SVM by eliminating irrelevant and distorting training instances. Nonetheless, there have been few studies that have applied instance selection to SVM, especially in the domain of stock market prediction. Instance selection tries to choose proper instance subsets from original training data. It may be considered as a method of knowledge refinement and it maintains the instance-base. This study proposes the novel instance selection algorithm for SVMs. The proposed technique in this study uses genetic algorithm (GA) to optimize instance selection process with parameter optimization simultaneously. We call the model as ISVM (SVM with Instance selection) in this study. Experiments on stock market data are implemented using ISVM. In this study, the GA searches for optimal or near-optimal values of kernel parameters and relevant instances for SVMs. This study needs two sets of parameters in chromosomes in GA setting : The codes for kernel parameters and for instance selection. For the controlling parameters of the GA search, the population size is set at 50 organisms and the value of the crossover rate is set at 0.7 while the mutation rate is 0.1. As the stopping condition, 50 generations are permitted. The application data used in this study consists of technical indicators and the direction of change in the daily Korea stock price index (KOSPI). The total number of samples is 2218 trading days. We separate the whole data into three subsets as training, test, hold-out data set. The number of data in each subset is 1056, 581, 581 respectively. This study compares ISVM to several comparative models including logistic regression (logit), backpropagation neural networks (ANN), nearest neighbor (1-NN), conventional SVM (SVM) and SVM with the optimized parameters (PSVM). In especial, PSVM uses optimized kernel parameters by the genetic algorithm. The experimental results show that ISVM outperforms 1-NN by 15.32%, ANN by 6.89%, Logit and SVM by 5.34%, and PSVM by 4.82% for the holdout data. For ISVM, only 556 data from 1056 original training data are used to produce the result. In addition, the two-sample test for proportions is used to examine whether ISVM significantly outperforms other comparative models. The results indicate that ISVM outperforms ANN and 1-NN at the 1% statistical significance level. In addition, ISVM performs better than Logit, SVM and PSVM at the 5% statistical significance level.