• Title/Summary/Keyword: Search algorithms

Search Result 1,333, Processing Time 0.028 seconds

To increase positioning accuracy in an IPS environment WPS Search Algorithm Design (IPS 환경에서 위치 정확도를 높이기 위한 WPS 탐색 알고리즘 설계)

  • Lee, Hyoun-Sup;Shim, Man-Taek;Jang, Si-Woong
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.75-76
    • /
    • 2022
  • WPS is a system that finds the location of a current moving object through information of a wireless AP. The current location is determined by utilizing the outdoor and indoor AP signal strength characteristics. Even if the map is configured for the first service, the map information of the DB built by the variability of the AP signal cannot be 100% reliable in the case of outdoor. Therefore, various algorithms are needed to increase the accuracy of WPS. This paper proposes a system to increase positioning accuracy by applying various methods to determine the current position of a moving object.

  • PDF

A hybrid algorithm for classifying rock joints based on improved artificial bee colony and fuzzy C-means clustering algorithm

  • Ji, Duofa;Lei, Weidong;Chen, Wenqin
    • Geomechanics and Engineering
    • /
    • v.31 no.4
    • /
    • pp.353-364
    • /
    • 2022
  • This study presents a hybrid algorithm for classifying the rock joints, where the improved artificial bee colony (IABC) and the fuzzy C-means (FCM) clustering algorithms are incorporated to take advantage of the artificial bee colony (ABC) algorithm by tuning the FCM clustering algorithm to obtain the more reasonable and stable result. A coefficient is proposed to reduce the amount of blind random searches and speed up convergence, thus achieving the goals of optimizing and improving the ABC algorithm. The results from the IABC algorithm are used as initial parameters in FCM to avoid falling to the local optimum in the local search, thus obtaining stable classifying results. Two validity indices are adopted to verify the rationality and practicability of the IABC-FCM algorithm in classifying the rock joints, and the optimal amount of joint sets is obtained based on the two validity indices. Two illustrative examples, i.e., the simulated rock joints data and the field-survey rock joints data, are used in the verification to check the feasibility and practicability in rock engineering for the proposed algorithm. The results show that the IABC-FCM algorithm could be applicable in classifying the rock joint sets.

Slime mold and four other nature-inspired optimization algorithms in analyzing the concrete compressive strength

  • Yinghao Zhao;Hossein Moayedi;Loke Kok Foong;Quynh T. Thi
    • Smart Structures and Systems
    • /
    • v.33 no.1
    • /
    • pp.65-91
    • /
    • 2024
  • The use of five optimization techniques for the prediction of a strength-based concrete mixture's best-fit model is examined in this work. Five optimization techniques are utilized for this purpose: Slime Mold Algorithm (SMA), Black Hole Algorithm (BHA), Multi-Verse Optimizer (MVO), Vortex Search (VS), and Whale Optimization Algorithm (WOA). MATLAB employs a hybrid learning strategy to train an artificial neural network that combines least square estimation with backpropagation. Thus, 72 samples are utilized as training datasets and 31 as testing datasets, totaling 103. The multi-layer perceptron (MLP) is used to analyze all data, and results are verified by comparison. For training datasets in the best-fit models of SMA-MLP, BHA-MLP, MVO-MLP, VS-MLP, and WOA-MLP, the statistical indices of coefficient of determination (R2) in training phase are 0.9603, 0.9679, 0.9827, 0.9841 and 0.9770, and in testing phase are 0.9567, 0.9552, 0.9594, 0.9888 and 0.9695 respectively. In addition, the best-fit structures for training for SMA, BHA, MVO, VS, and WOA (all combined with multilayer perceptron, MLP) are achieved when the term population size was modified to 450, 500, 250, 150, and 500, respectively. Among all the suggested options, VS could offer a stronger prediction network for training MLP.

Demand Response Based Optimal Microgrid Scheduling Problem Using A Multi-swarm Sine Cosine Algorithm

  • Chenye Qiu;Huixing Fang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.8
    • /
    • pp.2157-2177
    • /
    • 2024
  • Demand response (DR) refers to the customers' active reaction with respect to the changes of market pricing or incentive policies. DR plays an important role in improving network reliability, minimizing operational cost and increasing end users' benefits. Hence, the integration of DR in the microgrid (MG) management is gaining increasing popularity nowadays. This paper proposes a day-ahead MG scheduling framework in conjunction with DR and investigates the impact of DR in optimizing load profile and reducing overall power generation costs. A linear responsive model considering time of use (TOU) price and incentive is developed to model the active reaction of customers' consumption behaviors. Thereafter, a novel multi-swarm sine cosine algorithm (MSCA) is proposed to optimize the total power generation costs in the framework. In the proposed MSCA, several sub-swarms search for better solutions simultaneously which is beneficial for improving the population diversity. A cooperative learning scheme is developed to realize knowledge dissemination in the population and a competitive substitution strategy is proposed to prevent local optima stagnation. The simulation results obtained by the proposed MSCA are compared with other meta-heuristic algorithms to show its effectiveness in reducing overall generation costs. The outcomes with and without DR suggest that the DR program can effectively reduce the total generation costs and improve the stability of the MG network.

Study on failure mode prediction of reinforced concrete columns based on class imbalanced dataset

  • Mingyi Cai;Guangjun Sun;Bo Chen
    • Earthquakes and Structures
    • /
    • v.27 no.3
    • /
    • pp.177-189
    • /
    • 2024
  • Accurately predicting the failure modes of reinforced concrete (RC) columns is essential for structural design and assessment. In this study, the challenges of imbalanced datasets and complex feature selection in machine learning (ML) methods were addressed through an optimized ML approach. By combining feature selection and oversampling techniques, the prediction of seismic failure modes in rectangular RC columns was improved. Two feature selection methods were used to identify six input parameters. To tackle class imbalance, the Borderline-SMOTE1 algorithm was employed, enhancing the learning capabilities of the models for minority classes. Eight ML algorithms were trained and fine-tuned using k-fold shuffle split cross-validation and grid search. The results showed that the artificial neural network model achieved 96.77% accuracy, while k-nearest neighbor, support vector machine, and random forest models each achieved 95.16% accuracy. The balanced dataset led to significant improvements, particularly in predicting the flexure-shear failure mode, with accuracy increasing by 6%, recall by 8%, and F1 scores by 7%. The use of the Borderline-SMOTE1 algorithm significantly improved the recognition of samples at failure mode boundaries, enhancing the classification performance of models like k-nearest neighbor and decision tree, which are highly sensitive to data distribution and decision boundaries. This method effectively addressed class imbalance and selected relevant features without requiring complex simulations like traditional methods, proving applicable for discerning failure modes in various concrete members under seismic action.

Korean Word Sense Disambiguation using Dictionary and Corpus (사전과 말뭉치를 이용한 한국어 단어 중의성 해소)

  • Jeong, Hanjo;Park, Byeonghwa
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.1-13
    • /
    • 2015
  • As opinion mining in big data applications has been highlighted, a lot of research on unstructured data has made. Lots of social media on the Internet generate unstructured or semi-structured data every second and they are often made by natural or human languages we use in daily life. Many words in human languages have multiple meanings or senses. In this result, it is very difficult for computers to extract useful information from these datasets. Traditional web search engines are usually based on keyword search, resulting in incorrect search results which are far from users' intentions. Even though a lot of progress in enhancing the performance of search engines has made over the last years in order to provide users with appropriate results, there is still so much to improve it. Word sense disambiguation can play a very important role in dealing with natural language processing and is considered as one of the most difficult problems in this area. Major approaches to word sense disambiguation can be classified as knowledge-base, supervised corpus-based, and unsupervised corpus-based approaches. This paper presents a method which automatically generates a corpus for word sense disambiguation by taking advantage of examples in existing dictionaries and avoids expensive sense tagging processes. It experiments the effectiveness of the method based on Naïve Bayes Model, which is one of supervised learning algorithms, by using Korean standard unabridged dictionary and Sejong Corpus. Korean standard unabridged dictionary has approximately 57,000 sentences. Sejong Corpus has about 790,000 sentences tagged with part-of-speech and senses all together. For the experiment of this study, Korean standard unabridged dictionary and Sejong Corpus were experimented as a combination and separate entities using cross validation. Only nouns, target subjects in word sense disambiguation, were selected. 93,522 word senses among 265,655 nouns and 56,914 sentences from related proverbs and examples were additionally combined in the corpus. Sejong Corpus was easily merged with Korean standard unabridged dictionary because Sejong Corpus was tagged based on sense indices defined by Korean standard unabridged dictionary. Sense vectors were formed after the merged corpus was created. Terms used in creating sense vectors were added in the named entity dictionary of Korean morphological analyzer. By using the extended named entity dictionary, term vectors were extracted from the input sentences and then term vectors for the sentences were created. Given the extracted term vector and the sense vector model made during the pre-processing stage, the sense-tagged terms were determined by the vector space model based word sense disambiguation. In addition, this study shows the effectiveness of merged corpus from examples in Korean standard unabridged dictionary and Sejong Corpus. The experiment shows the better results in precision and recall are found with the merged corpus. This study suggests it can practically enhance the performance of internet search engines and help us to understand more accurate meaning of a sentence in natural language processing pertinent to search engines, opinion mining, and text mining. Naïve Bayes classifier used in this study represents a supervised learning algorithm and uses Bayes theorem. Naïve Bayes classifier has an assumption that all senses are independent. Even though the assumption of Naïve Bayes classifier is not realistic and ignores the correlation between attributes, Naïve Bayes classifier is widely used because of its simplicity and in practice it is known to be very effective in many applications such as text classification and medical diagnosis. However, further research need to be carried out to consider all possible combinations and/or partial combinations of all senses in a sentence. Also, the effectiveness of word sense disambiguation may be improved if rhetorical structures or morphological dependencies between words are analyzed through syntactic analysis.

Analysis and Evaluation of Frequent Pattern Mining Technique based on Landmark Window (랜드마크 윈도우 기반의 빈발 패턴 마이닝 기법의 분석 및 성능평가)

  • Pyun, Gwangbum;Yun, Unil
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.101-107
    • /
    • 2014
  • With the development of online service, recent forms of databases have been changed from static database structures to dynamic stream database structures. Previous data mining techniques have been used as tools of decision making such as establishment of marketing strategies and DNA analyses. However, the capability to analyze real-time data more quickly is necessary in the recent interesting areas such as sensor network, robotics, and artificial intelligence. Landmark window-based frequent pattern mining, one of the stream mining approaches, performs mining operations with respect to parts of databases or each transaction of them, instead of all the data. In this paper, we analyze and evaluate the techniques of the well-known landmark window-based frequent pattern mining algorithms, called Lossy counting and hMiner. When Lossy counting mines frequent patterns from a set of new transactions, it performs union operations between the previous and current mining results. hMiner, which is a state-of-the-art algorithm based on the landmark window model, conducts mining operations whenever a new transaction occurs. Since hMiner extracts frequent patterns as soon as a new transaction is entered, we can obtain the latest mining results reflecting real-time information. For this reason, such algorithms are also called online mining approaches. We evaluate and compare the performance of the primitive algorithm, Lossy counting and the latest one, hMiner. As the criteria of our performance analysis, we first consider algorithms' total runtime and average processing time per transaction. In addition, to compare the efficiency of storage structures between them, their maximum memory usage is also evaluated. Lastly, we show how stably the two algorithms conduct their mining works with respect to the databases that feature gradually increasing items. With respect to the evaluation results of mining time and transaction processing, hMiner has higher speed than that of Lossy counting. Since hMiner stores candidate frequent patterns in a hash method, it can directly access candidate frequent patterns. Meanwhile, Lossy counting stores them in a lattice manner; thus, it has to search for multiple nodes in order to access the candidate frequent patterns. On the other hand, hMiner shows worse performance than that of Lossy counting in terms of maximum memory usage. hMiner should have all of the information for candidate frequent patterns to store them to hash's buckets, while Lossy counting stores them, reducing their information by using the lattice method. Since the storage of Lossy counting can share items concurrently included in multiple patterns, its memory usage is more efficient than that of hMiner. However, hMiner presents better efficiency than that of Lossy counting with respect to scalability evaluation due to the following reasons. If the number of items is increased, shared items are decreased in contrast; thereby, Lossy counting's memory efficiency is weakened. Furthermore, if the number of transactions becomes higher, its pruning effect becomes worse. From the experimental results, we can determine that the landmark window-based frequent pattern mining algorithms are suitable for real-time systems although they require a significant amount of memory. Hence, we need to improve their data structures more efficiently in order to utilize them additionally in resource-constrained environments such as WSN(Wireless sensor network).

Discrete Optimum Design of Sinusoidal Corrugated Web Girder (사인형 주름웨브보의 이산화 최적구조설계)

  • Shon, Su Deok;Yoo, Mi Na;Lee, Seung Jae
    • Journal of Korean Society of Steel Construction
    • /
    • v.24 no.6
    • /
    • pp.671-682
    • /
    • 2012
  • The use of sinusoidal corrugated web girder for the box-type girders and gable steel main frames has recently been increasing very much. The reasons are that the thin web of the girder affords a significant weight reduction compared with rolled beam and welded built-up girder, and that corrugation prevents the buckling failure of the web. Improvements of the automatic fabrication process makes mass production of the corrugated web and unit possible, and applications of this girder have been extended considerably. Thus, the research for the optimum design processer considering the production data is needed practically. For doing this research, we develope the discrete optimum structural design program in consideration of production list data for the research, and the program apply to the single girder under the uniform load and the concentrated load as numerical example. We consider objective function as minimum weight of the girder, and use slenderness ratio, stress of flanges and corrugated web, and the girder deflection as the constraint functions. And also the Genetic Algorithms is adopted to search the global minimum point by using the production list as a discrete design variable. Finally, to verify the optimality of the design, we conduct a comparison of the results of the discrete optimum design with those of the continuous one, and also analyze the characteristics of the optimum cross-section.

An Adaptive Priority-based Sequenced Route Query Processing Method in Road Networks (도로 네트워크 환경에서 적응적 우선순위 기반의 순차적 경로 처리 기법)

  • Ryu, Hyeongcheol;Jung, Sungwon
    • KIISE Transactions on Computing Practices
    • /
    • v.20 no.12
    • /
    • pp.652-657
    • /
    • 2014
  • Given a starting point, destination point and various Points Of Interest (POIs), which contain a full or partial order, for a user to visit we wish to create, a sequenced route from the starting point to the destination point that includes one member of each POI type in a particular order. This paper proposes a method for finding the approximate shortest route between the start point, destination point and one member of each POI type. There are currently two algorithms that perform this task but they both have weaknesses. One of the algorithms only considers the distance between the visited POI (or starting point) and POI to visit next. The other algorithm chooses candidate points near the straight-line distance between the start point and destination but does not consider the order of visits on the corresponding network path. This paper outlines an algorithm that chooses the candidate points that are nearer to the network path between the start point and destination using network search. The algorithm looks for routes using the candidate points and finds the approximate shortest route by assigning an adaptive priority to the route that visits more POIs in a short amount of time.

Application of Self-Adaptive Meta-Heuristic Optimization Algorithm for Muskingum Flood Routing (Muskingum 홍수추적을 위한 자가적응형 메타 휴리스틱 알고리즘의 적용)

  • Lee, Eui Hoon
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.7
    • /
    • pp.29-37
    • /
    • 2020
  • In the past, meta-heuristic optimization algorithms were developed to solve the problems caused by complex nonlinearities occurring in natural phenomena, and various studies have been conducted to examine the applicability of the developed algorithms. The self-adaptive vision correction algorithm (SAVCA) showed excellent performance in mathematics problems, but it did not apply to complex engineering problems. Therefore, it is necessary to review the application process of the SAVCA. The SAVCA, which was recently developed and showed excellent performance, was applied to the advanced Muskingum flood routing model (ANLMM-L) to examine the application and application process. First, initial solutions were generated by the SAVCA, and the fitness was then calculated by ANLMM-L. The new value selected by a local and global search was put into the SAVCA. A new solution was generated, and ANLMM-L was applied again to calculate the fitness. The final calculation was conducted by comparing and improving the results of the new solution and existing solutions. The sum of squares (SSQ) was used to calculate the error between the observed and calculated runoff, and the applied results were compared with the current models. SAVCA, which showed excellent performance in the Muskingum flood routing model, is expected to show excellent performance in a range of engineering problems.