• Title/Summary/Keyword: 알고리즘 시간효율성

Search Result 1,105, Processing Time 0.036 seconds

The Estimation Model of an Origin-Destination Matrix from Traffic Counts Using a Conjugate Gradient Method (Conjugate Gradient 기법을 이용한 관측교통량 기반 기종점 OD행렬 추정 모형 개발)

  • Lee, Heon-Ju;Lee, Seung-Jae
    • Journal of Korean Society of Transportation
    • /
    • v.22 no.1 s.72
    • /
    • pp.43-62
    • /
    • 2004
  • Conventionally the estimation method of the origin-destination Matrix has been developed by implementing the expansion of sampled data obtained from roadside interview and household travel survey. In the survey process, the bigger the sample size is, the higher the level of limitation, due to taking time for an error test for a cost and a time. Estimating the O-D matrix from observed traffic count data has been applied as methods of over-coming this limitation, and a gradient model is known as one of the most popular techniques. However, in case of the gradient model, although it may be capable of minimizing the error between the observed and estimated traffic volumes, a prior O-D matrix structure cannot maintained exactly. That is to say, unwanted changes may be occurred. For this reason, this study adopts a conjugate gradient algorithm to take into account two factors: estimation of the O-D matrix from the conjugate gradient algorithm while reflecting the prior O-D matrix structure maintained. This development of the O-D matrix estimation model is to minimize the error between observed and estimated traffic volumes. This study validates the model using the simple network, and then applies it to a large scale network. There are several findings through the tests. First, as the consequence of consistency, it is apparent that the upper level of this model plays a key role by the internal relationship with lower level. Secondly, as the respect of estimation precision, the estimation error is lied within the tolerance interval. Furthermore, the structure of the estimated O-D matrix has not changed too much, and even still has conserved some attributes.

An Energy Efficient Unequal Clustering Algorithm for Wireless Sensor Networks (무선 센서 네트워크에서의 에너지 효율적인 불균형 클러스터링 알고리즘)

  • Lee, Sung-Ju;Kim, Sung-Chun
    • The KIPS Transactions:PartC
    • /
    • v.16C no.6
    • /
    • pp.783-790
    • /
    • 2009
  • The necessity of wireless sensor networks is increasing in the recent years. So many researches are studied in wireless sensor networks. The clustering algorithm provides an effective way to prolong the lifetime of the wireless sensor networks. The one-hop routing of LEACH algorithm is an inefficient way in the energy consumption of cluster-head, because it transmits a data to the BS(Base Station) with one-hop. On the other hand, other clustering algorithms transmit data to the BS with multi-hop, because the multi-hop transmission is an effective way. But the multi-hop routing of other clustering algorithms which transmits data to BS with multi-hop have a data bottleneck state problem. The unequal clustering algorithm solved a data bottleneck state problem by increasing the routing path. Most of the unequal clustering algorithms partition the nodes into clusters of unequal size, and clusters closer to the BS have small-size the those farther away from the BS. However, the energy consumption of cluster-head in unequal clustering algorithm is more increased than other clustering algorithms. In the thesis, I propose an energy efficient unequal clustering algorithm which decreases the energy consumption of cluster-head and solves the data bottleneck state problem. The basic idea is divided a three part. First of all I provide that the election of appropriate cluster-head. Next, I offer that the decision of cluster-size which consider the distance from the BS, the energy state of node and the number of neighborhood node. Finally, I provide that the election of assistant node which the transmit function substituted for cluster-head. As a result, the energy consumption of cluster-head is minimized, and the energy consumption of total network is minimized.

Aerodynamic Design of EAV Propeller using a Multi-Level Design Optimization Framework (다단 최적 설계 프레임워크를 활용한 전기추진 항공기 프로펠러 공력 최적 설계)

  • Kwon, Hyung-Il;Yi, Seul-Gi;Choi, Seongim;Kim, Keunbae
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.41 no.3
    • /
    • pp.173-184
    • /
    • 2013
  • A multi-level design optimization framework for aerodynamic design of rotary wing such as propeller and helicopter rotor blades is presented in this study. Strategy of the proposed framework is to enhance aerodynamic performance by sequentially applying the planform and sectional design optimization. In the first level of a planform design, we used a genetic algorithm and blade element momentum theory (BEMT) based on two-dimensional aerodynamic database to find optimal planform variables. After an initial planform design, local flow conditions of blade sections are analyzed using high-fidelity CFD methods. During the next level, a sectional design optimization is conducted using two dimensional Navier-Stokes analysis and a gradient based optimization algorithm. When optimal airfoil shape is determined at the several spanwise locations, a planform design is performed again. Through this iterative design process, not only an optimal flow condition but also an optimal shape of an EAV propeller blade is obtained. To validate the optimized propeller-blade design, it is tested in wind-tunnel facility with different flow conditions. An efficiency, which is slightly less than the expected improvement of 7% predicted by our proposed design framework but is still satisfactory to enhance the aerodynamic performance of EAV system.

Gate-Level Conversion Methods between Boolean and Arithmetic Masks (불 마스크와 산술 마스크에 대한 게이트 레벨 변환기법)

  • Baek, Yoo-Jin
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.46 no.11
    • /
    • pp.8-15
    • /
    • 2009
  • Side-channel attacks including the differential power analysis attack are often more powerful than classical cryptanalysis and have to be seriously considered by cryptographic algorithm's implementers. Various countermeasures have been proposed against such attacks. In this paper, we deal with the masking method, which is known to be a very effective countermeasure against the differential power analysis attack and propose new gate-level conversion methods between Boolean and arithmetic masks. The new methods require only 6n-5 XOR and 2n-2 AND gates with 3n-2 gate delay for converting n-bit masks. The basic idea of the proposed methods is that the carry and the sum bits in the ripple adder are manipulated in a way that the adversary cannot detect the relation between these bits and the original raw data. Since the proposed methods use only bitwise operations, they are especially useful for DPA-securely implementing cryptographic algorithms in hardware which use both Boolean and arithmetic operations. For example, we applied them to securely implement the block encryption algorithm SEED in hardware and present its detailed implementation result.

A Seamline Extraction Technique Considering the Characteristic of NDVI for High Resolution Satellite Image Mosaics (고해상도 위성영상 모자이크를 위한 NDVI 특성을 이용한 접합선 추출 기법)

  • Kim, Jiyoung;Chae, Taebyeong;Byun, Younggi
    • Korean Journal of Remote Sensing
    • /
    • v.31 no.5
    • /
    • pp.395-408
    • /
    • 2015
  • High-resolution satellite image mosaics are becoming increasingly important in the field of remote sensing image analysis as an essential image processing to create a large image constructed from several smaller images. In this paper, we present an automatic seamline extraction technique and the procedure to generate a mosaic image by this technique. For more effective seamline extraction in the overlap region of adjacent images, an NDVI-based seamline extraction technique is developed, which takes advantage of the computational time and memory. The Normalized Difference Vegetation Index(NDVI) is an index of plant "greeness" or photosynthetic activity that is employed to extract the initial seamline. The NDVI can divide into manmade region and natural region. The cost image is obtained by the canny edge detector and the buffering technique is used to extract the ranging cost image. The seamline is extracted by applying the Dijkstra algorithm to a cost image generated through the labeling process of the extracted edge information. Histogram matching is also conducted to alleviate radiometric distortion between adjacent images acquired at different time. In the experimental results using the KOMPSAT-2/3 satellite imagery, it is confirmed that the proposed method greatly reduces the visual discontinuity caused by geometric difference of adjacent images and the computation time.

A Study on the RFID Biometrics System Based on Hippocampal Learning Algorithm Using NMF and LDA Mixture Feature Extraction (NMF와 LDA 혼합 특징추출을 이용한 해마 학습기반 RFID 생체 인증 시스템에 관한 연구)

  • Oh Sun-Moon;Kang Dae-Seong
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.4 s.310
    • /
    • pp.46-54
    • /
    • 2006
  • Recently, the important of a personal identification is increasing according to expansion using each on-line commercial transaction and personal ID-card. Although a personal ID-card embedded RFID(Radio Frequency Identification) tag is gradually increased, the way for a person's identification is deficiency. So we need automatic methods. Because RFID tag is vary small storage capacity of memory, it needs effective feature extraction method to store personal biometrics information. We need new recognition method to compare each feature. In this paper, we studied the face verification system using Hippocampal neuron modeling algorithm which can remodel the hippocampal neuron as a principle of a man's brain in engineering, then it can learn the feature vector of the face images very fast. and construct the optimized feature each image. The system is composed of two parts mainly. One is feature extraction using NMF(Non-negative Matrix Factorization) and LDA(Linear Discriminants Analysis) mixture algorithm and the other is hippocampal neuron modeling and recognition simulation experiments confirm the each recognition rate, that are face changes, pose changes and low-level quality image. The results of experiments, we can compare a feature extraction and learning method proposed in this paper of any other methods, and we can confirm that the proposed method is superior to the existing method.

Bitmap Indexes and Query Processing Strategies for Relational XML Twig Queries (관계형 XML 가지 패턴 질의를 위한 비트맵 인덱스와 질의 처리 기법)

  • Lee, Kyong-Ha;Moon, Bong-Ki;Lee, Kyu-Chul
    • Journal of KIISE:Databases
    • /
    • v.37 no.3
    • /
    • pp.146-164
    • /
    • 2010
  • Due to an increasing volume of XML data, it is considered prudent to store XML data on an industry-strength database system instead of relying on a domain specific application or a file system. For shredded XML data stored in relational tables, however, it may not be straightforward to apply existing algorithms for twig query processing, since most of the algorithms require XML data to be accessed in a form of streams of elements grouped by their tags and sorted in a particular order. In order to support XML query processing within the common framework of relational database systems, we first propose several bitmap indexes and their strategies for supporting holistic twig joining on XML data stored in relational tables. Since bitmap indexes are well supported in most of the commercial and open-source database systems, the proposed bitmapped indexes and twig query processing strategies can be incorporated into relational query processing framework with more ease. The proposed query processing strategies are efficient in terms of both time and space, because the compressed bitmap indexes stay compressed during data access. In addition, we propose a hybrid index which computes twig query solutions with only bit-vectors, without accessing labeled XML elements stored in the relational tables.

Self-optimizing feature selection algorithm for enhancing campaign effectiveness (캠페인 효과 제고를 위한 자기 최적화 변수 선택 알고리즘)

  • Seo, Jeoung-soo;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.173-198
    • /
    • 2020
  • For a long time, many studies have been conducted on predicting the success of campaigns for customers in academia, and prediction models applying various techniques are still being studied. Recently, as campaign channels have been expanded in various ways due to the rapid revitalization of online, various types of campaigns are being carried out by companies at a level that cannot be compared to the past. However, customers tend to perceive it as spam as the fatigue of campaigns due to duplicate exposure increases. Also, from a corporate standpoint, there is a problem that the effectiveness of the campaign itself is decreasing, such as increasing the cost of investing in the campaign, which leads to the low actual campaign success rate. Accordingly, various studies are ongoing to improve the effectiveness of the campaign in practice. This campaign system has the ultimate purpose to increase the success rate of various campaigns by collecting and analyzing various data related to customers and using them for campaigns. In particular, recent attempts to make various predictions related to the response of campaigns using machine learning have been made. It is very important to select appropriate features due to the various features of campaign data. If all of the input data are used in the process of classifying a large amount of data, it takes a lot of learning time as the classification class expands, so the minimum input data set must be extracted and used from the entire data. In addition, when a trained model is generated by using too many features, prediction accuracy may be degraded due to overfitting or correlation between features. Therefore, in order to improve accuracy, a feature selection technique that removes features close to noise should be applied, and feature selection is a necessary process in order to analyze a high-dimensional data set. Among the greedy algorithms, SFS (Sequential Forward Selection), SBS (Sequential Backward Selection), SFFS (Sequential Floating Forward Selection), etc. are widely used as traditional feature selection techniques. It is also true that if there are many risks and many features, there is a limitation in that the performance for classification prediction is poor and it takes a lot of learning time. Therefore, in this study, we propose an improved feature selection algorithm to enhance the effectiveness of the existing campaign. The purpose of this study is to improve the existing SFFS sequential method in the process of searching for feature subsets that are the basis for improving machine learning model performance using statistical characteristics of the data to be processed in the campaign system. Through this, features that have a lot of influence on performance are first derived, features that have a negative effect are removed, and then the sequential method is applied to increase the efficiency for search performance and to apply an improved algorithm to enable generalized prediction. Through this, it was confirmed that the proposed model showed better search and prediction performance than the traditional greed algorithm. Compared with the original data set, greed algorithm, genetic algorithm (GA), and recursive feature elimination (RFE), the campaign success prediction was higher. In addition, when performing campaign success prediction, the improved feature selection algorithm was found to be helpful in analyzing and interpreting the prediction results by providing the importance of the derived features. This is important features such as age, customer rating, and sales, which were previously known statistically. Unlike the previous campaign planners, features such as the combined product name, average 3-month data consumption rate, and the last 3-month wireless data usage were unexpectedly selected as important features for the campaign response, which they rarely used to select campaign targets. It was confirmed that base attributes can also be very important features depending on the type of campaign. Through this, it is possible to analyze and understand the important characteristics of each campaign type.

LFH: Low-Cost and Fast Handoff Scheme in Proxy Mobile IPv6 Networks with Multicasting Support (프록시 모바일 IPv6 네트워크에서 멀티캐스팅을 지원하는 저비용의 빠른 이동성관리 기법)

  • Kim, Eunhwa;Jeong, Jongpil
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.2 no.6
    • /
    • pp.265-278
    • /
    • 2013
  • With the recent advancements in various wireless communication technologies, the importance of mobile multicasting is coming to the fore, in an effort to use network resources more efficiently. In the past, when various mobile IP-based multicast techniques were proposed, the focus was put on the costs needed for network delivery for providing multicast services, as well as on minimizing the multicast handover delay. For techniques using MIPv6 (Mobile IPv6), a host-based mobility management protocol, however, it is fundamentally difficult to resolve the problems of handover delay and tunnel convergence. To resolve these problems, a network-based mobility management protocol called PMIPv6 (Proxy Mobile IPv6) was standardized. Although performance is improved in PMIPv6 over MIPv6, it still suffers from the problems of handover delay and tunnel convergence. In this paper, to overcome these limitations, a technique called LFH (Low-cost and Fast Handoff) is proposed for fast and low-cost mobility management with multicasting support in PMIPv6 networks. To reduce the interactions between the complex multicast routing protocol and the multicast messages, a simplified proxy method called MLD (Multicast Listener Discovery) is implemented and modified. Furthermore, a TCR (Tunnel Combination and Reconstruction) algorithm was used in the multicast handover procedure within the LMA (Local Mobility Anchor) domain, as well as in the multicast handover procedure between domains, in order to overcome the problem of tunnel convergence. As a result, it was found that LFH has reduced multicast delay compared to other types of multicast techniques, and that it requires lower costs as well.

The Role of Geographic Information System and Its Functional Intergration Strategy in the Conventional Transportation Planning Process (전통교통계획과정에 있어서 GIS의 역할 및 기능적 통합방안에 관한 연구)

  • Choi, Kee-Choo
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.1 no.1 s.1
    • /
    • pp.127-140
    • /
    • 1993
  • The purpose of this paper is to examine the possible benefits of combining transportation planning models with geographic information systems (GIS) in the hope that intergrating these systems can alleviate the inherent problems of transportation planning models such as user unfriendliness, labor intensiveness, and theoretical limitations. Specially, this paper focuses on the issue of incompatiblity between GIS and the conventional transportation planning models in dealing with network topologies. Resolving this conflict in topologies is a conerstone for eliminating the user-unfriendliness and labor-intensiveness issues. This paper presents the development of an algorithm that converts GIS topology into transportation network topology. The FORTRAN-based topology conversion algorithm generates transportation networks from the GIS cartographic file and establishes a communication charmel between the two systems.

  • PDF