• Title/Summary/Keyword: Interval Partition

Search Result 30, Processing Time 0.019 seconds

Numerical Formula and Verification of Web Robot for Collection Speedup of Web Documents

  • Kim Weon;Kim Young-Ki;Chin Yong-Ok
    • Journal of Internet Computing and Services
    • /
    • v.5 no.6
    • /
    • pp.1-10
    • /
    • 2004
  • A web robot is a software that has abilities of tracking and collecting web documents on the Internet(l), The performance scalability of recent web robots reached the limit CIS the number of web documents on the internet has increased sharply as the rapid growth of the Internet continues, Accordingly, it is strongly demanded to study on the performance scalability in searching and collecting documents on the web. 'Design of web robot based on Multi-Agent to speed up documents collection ' rather than 'Sequentially executing Web Robot based on the existing Fork-Join method' and the results of analysis on its performance scalability is presented in the thesis, For collection speedup, a Multi-Agent based web robot performs the independent process for inactive URL ('Dead-links' URL), which is caused by overloaded web documents, temporary network or web-server disturbance, after dividing them into each agent. The agents consist of four component; Loader, Extractor, Active URL Scanner and inactive URL Scanner. The thesis models a Multi-Agent based web robot based on 'Amdahl's Law' to speed up documents collection, introduces a numerical formula for collection speedup, and verifies its performance improvement by comparing data from the formula with data from experiments based on the formula. Moreover, 'Dynamic URL Partition algorithm' is introduced and realized to minimize the workload of the web server by maximizing a interval of the web server which can be a collection target.

  • PDF

Statistical Information-Based Hierarchical Fuzzy-Rough Classification Approach (통계적 정보기반 계층적 퍼지-러프 분류기법)

  • Son, Chang-S.;Seo, Suk-T.;Chung, Hwan-M.;Kwon, Soon-H.
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.17 no.6
    • /
    • pp.792-798
    • /
    • 2007
  • In this paper, we propose a hierarchical fuzzy-rough classification method based on statistical information for maximizing the performance of pattern classification and reducing the number of rules without learning approaches such as neural network, genetic algorithm. In the proposed method, statistical information is used for extracting the partition intervals of antecedent fuzzy sets at each layer on hierarchical fuzzy-rough classification systems and rough sets are used for minimizing the number of fuzzy if-then rules which are associated with the partition intervals extracted by statistical information. To show the effectiveness of the proposed method, we compared the classification results(e.g. the classification accuracy and the number of rules) of the proposed with those of the conventional methods on the Fisher's IRIS data. From the experimental results, we can confirm the fact that the proposed method considers only statistical information of the given data is similar to the classification performance of the conventional methods.

Atrial Fibrillation Pattern Analysis based on Symbolization and Information Entropy (부호화와 정보 엔트로피에 기반한 심방세동 (Atrial Fibrillation: AF) 패턴 분석)

  • Cho, Ik-Sung;Kwon, Hyeog-Soong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.5
    • /
    • pp.1047-1054
    • /
    • 2012
  • Atrial fibrillation (AF) is the most common arrhythmia encountered in clinical practice, and its risk increases with age. Conventionally, the way of detecting AF was the time·frequency domain analysis of RR variability. However, the detection of ECG signal is difficult because of the low amplitude of the P wave and the corruption by the noise. Also, the time·frequency domain analysis of RR variability has disadvantage to get the details of irregular RR interval rhythm. In this study, we describe an atrial fibrillation pattern analysis based on symbolization and information entropy. We transformed RR interval data into symbolic sequence through differential partition, analyzed RR interval pattern, quantified the complexity through Shannon entropy and detected atrial fibrillation. The detection algorithm was tested using the threshold between 10ms and 100ms on two databases, namely the MIT-BIH Atrial Fibrillation Database.

Distributed Genetic Algorithms for the TSP (분산 유전알고리즘의 TSP 적용)

  • 박유석
    • Journal of the Korea Safety Management & Science
    • /
    • v.3 no.3
    • /
    • pp.191-200
    • /
    • 2001
  • Parallel Genetic Algorithms partition the whole population into several sub-populations and search the optimal solution by exchanging the information each others periodically. Distributed Genetic Algorithm, one of Parallel Genetic Algorithms, divides a large population into several sub-populations and executes the traditional Genetic Algorithm on each sub-population independently. And periodically promising individuals selected from sub-populations are migrated by following the migration interval and migration rate to different sub-populations. In this paper, for the Travelling Salesman Problems, we analyze and compare with Distributed Genetic Algorithms using different Genetic Algorithms and using same Genetic Algorithms on each separated sub-population The simulation result shows that using different Genetic Algorithms obtains better results than using same Genetic Algorithms in Distributed Genetic Algorithms. This results look like the property of rapidly searching the approximated optima and keeping the variety of solution make interaction in different Genetic Algorithms.

  • PDF

Effects of Single Nucleotide Polymorphism Marker Density on Haplotype Block Partition

  • Kim, Sun Ah;Yoo, Yun Joo
    • Genomics & Informatics
    • /
    • v.14 no.4
    • /
    • pp.196-204
    • /
    • 2016
  • Many researchers have found that one of the most important characteristics of the structure of linkage disequilibrium is that the human genome can be divided into non-overlapping block partitions in which only a small number of haplotypes are observed. The location and distribution of haplotype blocks can be seen as a population property influenced by population genetic events such as selection, mutation, recombination and population structure. In this study, we investigate the effects of the density of markers relative to the full set of all polymorphisms in the region on the results of haplotype partitioning for five popular haplotype block partition methods: three methods in Haploview (confidence interval, four gamete test, and solid spine), MIG++ implemented in PLINK 1.9 and S-MIG++. We used several experimental datasets obtained by sampling subsets of single nucleotide polymorphism (SNP) markers of chromosome 22 region in the 1000 Genomes Project data and also the HapMap phase 3 data to compare the results of haplotype block partitions by five methods. With decreasing sampling ratio down to 20% of the original SNP markers, the total number of haplotype blocks decreases and the length of haplotype blocks increases for all algorithms. When we examined the marker-independence of the haplotype block locations constructed from the datasets of different density, the results using below 50% of the entire SNP markers were very different from the results using the entire SNP markers. We conclude that the haplotype block construction results should be used and interpreted carefully depending on the selection of markers and the purpose of the study.

Performance Analysis on Declustering High-Dimensional Data by GRID Partitioning (그리드 분할에 의한 다차원 데이터 디클러스터링 성능 분석)

  • Kim, Hak-Cheol;Kim, Tae-Wan;Li, Ki-Joune
    • The KIPS Transactions:PartD
    • /
    • v.11D no.5
    • /
    • pp.1011-1020
    • /
    • 2004
  • A lot of work has been done to improve the I/O performance of such a system that store and manage a massive amount of data by distributing them across multiple disks and access them in parallel. Most of the previous work has focused on an efficient mapping from a grid ceil, which is determined bY the interval number of each dimension, to a disk number on the assumption that each dimension is split into disjoint intervals such that entire data space is GRID-like partitioned. However, they have ignored the effects of a GRID partitioning scheme on declustering performance. In this paper, we enhance the performance of mapping function based declustering algorithms by applying a good GRID par-titioning method. For this, we propose an estimation model to count the number of grid cells intersected by a range query and apply a GRID partitioning scheme which minimizes query result size among the possible schemes. While it is common to do binary partition for high-dimensional data, we choose less number of dimensions than needed for binary partition and split several times along that dimensions so that we can reduce the number of grid cells touched by a query. Several experimental results show that the proposed estimation model gives accuracy within 0.5% error ratio regardless of query size and dimension. We can also improve the performance of declustering algorithm based on mapping function, called Kronecker Sequence, which has been known to be the best among the mapping functions for high-dimensional data, up to 23 times by applying an efficient GRID partitioning scheme.

An Effective Mitigation Method on the Signal-Integrity Effects by Splitting of a Return Current Plane (귀환 전류 평면의 분할에 기인하는 신호 무결성의 효과적인 대책 방법)

  • Jung, Ki-Bum;Jun, Chang-Han;Chung, Yeon-Choon
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.19 no.3
    • /
    • pp.366-375
    • /
    • 2008
  • Generally a return current plane(RCP) of high speed digital and analog part is partitioned. This is achieved in order to decrease the noise interference between subsystem in PCBs(Printed Circuit Boards). However, when the connected signal line exists between each sub system, this partition will cause unwanted effects. In a circuital point of view, RCP partition has a bad influence upon signal integrity. In a EMI(Electromagnetic Interference) point of view, the partition of the return current plane becomes a primary factor to increase the radiated emission. Component bridge(CB) is usecl for the way of maintaining signal integrity, still specific user's guide doesn't give sufficient principle. In a view point of signal integrity, design principle of multi-CB using method will be analyzed by measurement and simulation. And design principle of noise mitigation will be provided. Generally interval of CB is ${\lambda}/20$ ferrite bead. In this study. When multi-CB connection is applied, design principle of ferrite bead and chip resistor is proved by measurement and simulation. Multi-connected chip resistance$(0{\Omega})$ is proved to be more effective design method in the point of signal integrity.

Design of Multiple Model Fuzzy Predictors using Data Preprocessing and its Application (데이터 전처리를 이용한 다중 모델 퍼지 예측기의 설계 및 응용)

  • Bang, Young-Keun;Lee, Chul-Heui
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.58 no.1
    • /
    • pp.173-180
    • /
    • 2009
  • It is difficult to predict non-stationary or chaotic time series which includes the drift and/or the non-linearity as well as uncertainty. To solve it, we propose an effective prediction method which adopts data preprocessing and multiple model TS fuzzy predictors combined with model selection mechanism. In data preprocessing procedure, the candidates of the optimal difference interval are determined based on the correlation analysis, and corresponding difference data sets are generated in order to use them as predictor input instead of the original ones because the difference data can stabilize the statistical characteristics of those time series and better reveals their implicit properties. Then, TS fuzzy predictors are constructed for multiple model bank, where k-means clustering algorithm is used for fuzzy partition of input space, and the least squares method is applied to parameter identification of fuzzy rules. Among the predictors in the model bank, the one which best minimizes the performance index is selected, and it is used for prediction thereafter. Finally, the error compensation procedure based on correlation analysis is added to improve the prediction accuracy. Some computer simulations are performed to verify the effectiveness of the proposed method.

AN EVALUATION FORMULA FOR A GENERALIZED CONDITIONAL EXPECTATION WITH TRANSLATION THEOREMS OVER PATHS

  • Cho, Dong Hyun
    • Journal of the Korean Mathematical Society
    • /
    • v.57 no.2
    • /
    • pp.451-470
    • /
    • 2020
  • Let C[0, T] denote an analogue of Wiener space, the space of real-valued continuous functions on the interval [0, T]. For a partition 0 = t0 < t1 < ⋯ < tn < tn+1 = T of [0, T], define Xn : C[0, T] → ℝn+1 by Xn(x) = (x(t0), x(t1), …, x(tn)). In this paper we derive a simple evaluation formula for Radon-Nikodym derivatives similar to the conditional expectations of functions on C[0, T] with the conditioning function Xn which has a drift and does not contain the present position of paths. As applications of the formula with Xn, we evaluate the Radon-Nikodym derivatives of the functions ∫0T[x(t)]mdλ(t)(m∈ℕ) and [∫0Tx(t)dλ(t)]2 on C[0, T], where λ is a complex-valued Borel measure on [0, T]. Finally we derive two translation theorems for the Radon-Nikodym derivatives of the functions on C[0, T].

Fast Algorithm for 360-degree Videos Based on the Prediction of Cu Depth Range and Fast Mode Decision

  • Zhang, Mengmeng;Zhang, Jing;Liu, Zhi;Mao, Fuqi;Yue, Wen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.6
    • /
    • pp.3165-3181
    • /
    • 2019
  • Spherical videos, which are also called 360-degree videos, have become increasingly popular due to the rapid development of virtual reality technology. However, the large amount of data in such videos is a huge challenge for existing transmission system. To use the existing encode framework, it should be converted into a 2D image plane by using a specific projection format, e.g. the equi-rectangular projection (ERP) format. The existing high-efficiency video coding standard (HEVC) can effectively compress video content, but its enormous computational complexity makes the time spent on compressing high-frame-rate and high-resolution 360-degree videos disproportionate to the benefits of compression. Focusing on the ERP format characteristics of 360-degree videos, this work develops a fast decision algorithm for predicting the coding unit depth interval and adaptive mode decision for intra prediction mode. The algorithm makes full use of the video characteristics of the ERP format by dealing with pole and equatorial areas separately. It sets different reference blocks and determination conditions according to the degree of stretching, which can reduce the coding time while ensuring the quality. Compared with the original reference software HM-16.16, the proposed algorithm can reduce time consumption by 39.3% in the all-intra configuration, and the BD-rate increases by only 0.84%.