• Title/Summary/Keyword: Algorithm partition

Search Result 360, Processing Time 0.022 seconds

Atrial Fibrillation Pattern Analysis based on Symbolization and Information Entropy (부호화와 정보 엔트로피에 기반한 심방세동 (Atrial Fibrillation: AF) 패턴 분석)

  • Cho, Ik-Sung;Kwon, Hyeog-Soong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.5
    • /
    • pp.1047-1054
    • /
    • 2012
  • Atrial fibrillation (AF) is the most common arrhythmia encountered in clinical practice, and its risk increases with age. Conventionally, the way of detecting AF was the time·frequency domain analysis of RR variability. However, the detection of ECG signal is difficult because of the low amplitude of the P wave and the corruption by the noise. Also, the time·frequency domain analysis of RR variability has disadvantage to get the details of irregular RR interval rhythm. In this study, we describe an atrial fibrillation pattern analysis based on symbolization and information entropy. We transformed RR interval data into symbolic sequence through differential partition, analyzed RR interval pattern, quantified the complexity through Shannon entropy and detected atrial fibrillation. The detection algorithm was tested using the threshold between 10ms and 100ms on two databases, namely the MIT-BIH Atrial Fibrillation Database.

A Space Efficient Indexing Technique for DNA Sequences (공간 효율적인 DNA 시퀀스 인덱싱 방안)

  • Song, Hye-Ju;Park, Young-Ho;Loh, Woong-Kee
    • Journal of KIISE:Databases
    • /
    • v.36 no.6
    • /
    • pp.455-465
    • /
    • 2009
  • Suffix trees are widely used in similar sequence matching for DNA. They have several problems such as time consuming, large space usages of disks and memories and data skew, since DNA sequences are very large and do not fit in the main memory. Thus, in the paper, we present a space efficient indexing method called SENoM, allowing us to build trees without merging phases for the partitioned sub trees. The proposed method is constructed in two phases. In the first phase, we partition the suffixes of the input string based on a common variable-length prefix till the number of suffixes is smaller than a threshold. In the second phase, we construct a sub tree based on the disk using the suffix sets, and then write it to the disk. The proposed method, SENoM eliminates complex merging phases. We show experimentally that proposed method is effective as bellows. SENoM reduces the disk usage less than 35% and reduces the memory usage less than 20% compared with TRELLIS algorithm. SENoM is available to query efficiently using the prefix tree even when the length of query sequence is large.

A PAPR Reduction Technique by the Partial Transmit Reduction Sequences (부분 전송 감소열에 의한 첨두대 평균 전력비 저감 기법)

  • Han Tae-Young;Yoo Young-Dae;Choi Jung-Hun;Kwon Young-Soo;Kim Nam
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.17 no.6 s.109
    • /
    • pp.562-573
    • /
    • 2006
  • It is required to reduce the peak-to-average power ratio(PAPR) in an orthogonal frequency division multiplexing system or a multicarrier system. And it is needed to eliminate the transmission of the side information in the Partial Transmit Sequences. So, in this paper, a new technique is proposed, where the subcarriers used for the multiple signal representation are only utilized for the reduction of PAPR to eliminate the burden of transmitting the side information. That is, it is proposed by taking the modified minimization criteria of partial transmit sequences scheme instead of using the convex optimization or the fast algorithm of tone reservation(TR) technique As the result of simulation, the PAPR reduction capability of the proposed method is improved by 3.2 dB dB, 3.4 dB, 3.6 dB with M=2, 4, 8(M is the number of partition in the so-called partial transmit reduction sequences(PTRS)), when the iteration number of fast algorithm of TR is 10 and the data rate loss is 5 %. But it is degraded in the capability of PAPR reduction by 3.4 dB, 3.1 dB, 2.2 dB, comparing to the TR when the data rate loss is 20 %. Therefore, the proposed method is outperformed the TR technique with respect to the complexity and PAPR reduction capability when M=2.

An Energy Estimation-based Routing Protocol for Maximizing Network Lifetime in Wireless Sensor Networks (무선 센서네트워크에서 네트워크 수명을 최대화하기 위한 에너지 추정 기반의 라우팅 프로토콜)

  • Hong, Ran-Kyung;Kweon, Ki-Suk;Ghim, Ho-Jin;Yoon, Hyun-Soo
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.3
    • /
    • pp.281-285
    • /
    • 2008
  • Wireless sensor networks are closely related with the geometric environment in which they are deployed. We consider the probable case when a routing protocol runs on an environment with many complex obstacles like downtown surroundings. In addition, there are no unrealistic assumptions in order to increase practicality of the protocol. Our goal is to find a routing protocol for maximizing network lifetime by using only connectivity information in the complex sensor network environment. We propose a topology-based routing algorithm that accomplishes good performance in terms of network lifetime and routing complexity as measures. Our routing algorithm makes routing decision based on a weighted graph as topological abstraction of the complex network. The graph conduces to lifetime enhancement by giving alternative paths, distributing the skewed burden. An energy estimation method is used so as to maintain routing information without any additional cost. We show how our approach can be used to maximize network lifetime and by extensive simulation we prove that out approach gives good results in terms of both measures-network lifetime and routing complexity.

An Implementation of an Edge-based Algorithm for Separating and Intersecting Spherical Polygons (구 볼록 다각형 들의 분리 및 교차를 위한 간선 기반 알고리즘의 구현)

  • Ha, Jong-Seong;Cheon, Eun-Hong
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.28 no.9
    • /
    • pp.479-490
    • /
    • 2001
  • In this paper, we consider the method of partitioning a sphere into faces with a set of spherical convex polygons $\Gamma$=${P_1...P_n}$ for determining the maximum of minimum intersection. This problem is commonly related with five geometric problems that fin the densest hemisphere containing the maximum subset of $\Gamma$, a great circle separating $\Gamma$, a great circle bisecting $\Gamma$ and a great circle intersecting the minimum or maximum subset of $\Gamma$. In order to efficiently compute the minimum or maximum intersection of spherical polygons. we take the approach of edge-based partition, in which the ownerships of edges rather than faces are manipulated as the sphere is incrementally partitioned by each of the polygons. Finally, by gathering the unordered split edges with the maximum number of ownerships. we approximately obtain the centroids of the solution faces without constructing their boundaries. Our algorithm for finding the maximum intersection is analyzed to have an efficient time complexity O(nv) where n and v respectively, are the numbers of polygons and all vertices. Furthermore, it is practical from the view of implementation, since it computes numerical values. robustly and deals with all the degenerate cases, Using the similar approach, the boundary of a general intersection can be constructed in O(nv+LlogL) time, where : is the output-senstive number of solution edges.

  • PDF

Improved Method for Learning Context-Free Grammar using Tabular representation

  • Jung, Soon-Ho
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.2
    • /
    • pp.43-51
    • /
    • 2022
  • In this paper, we suggest the method to improve the existing method leaning context-free grammar(CFG) using tabular representation(TBL) as a chromosome of genetic algorithm in grammatical inference and show the more efficient experimental result. We have two improvements. The first is to improve the formula to reflect the learning evaluation of positive and negative examples at the same time for the fitness function. The second is to classify partitions corresponding to TBLs generated from positive learning examples according to the size of the learning string, proceed with the evolution process by class, and adjust the composition ratio according to the success rate to apply the learning method linked to survival in the next generation. These improvements provide better efficiency than the existing method by solving the complexity and difficulty in the crossover and generalization steps between several individuals according to the size of the learning examples. We experiment with the languages proposed in the existing method, and the results show a rather fast generation rate that takes fewer generations to complete learning with the same success rate than the existing method. In the future, this method can be tried for extended CYK, and furthermore, it suggests the possibility of being applied to more complex parsing tables.

Data-centric XAI-driven Data Imputation of Molecular Structure and QSAR Model for Toxicity Prediction of 3D Printing Chemicals (3D 프린팅 소재 화학물질의 독성 예측을 위한 Data-centric XAI 기반 분자 구조 Data Imputation과 QSAR 모델 개발)

  • ChanHyeok Jeong;SangYoun Kim;SungKu Heo;Shahzeb Tariq;MinHyeok Shin;ChangKyoo Yoo
    • Korean Chemical Engineering Research
    • /
    • v.61 no.4
    • /
    • pp.523-541
    • /
    • 2023
  • As accessibility to 3D printers increases, there is a growing frequency of exposure to chemicals associated with 3D printing. However, research on the toxicity and harmfulness of chemicals generated by 3D printing is insufficient, and the performance of toxicity prediction using in silico techniques is limited due to missing molecular structure data. In this study, quantitative structure-activity relationship (QSAR) model based on data-centric AI approach was developed to predict the toxicity of new 3D printing materials by imputing missing values in molecular descriptors. First, MissForest algorithm was utilized to impute missing values in molecular descriptors of hazardous 3D printing materials. Then, based on four different machine learning models (decision tree, random forest, XGBoost, SVM), a machine learning (ML)-based QSAR model was developed to predict the bioconcentration factor (Log BCF), octanol-air partition coefficient (Log Koa), and partition coefficient (Log P). Furthermore, the reliability of the data-centric QSAR model was validated through the Tree-SHAP (SHapley Additive exPlanations) method, which is one of explainable artificial intelligence (XAI) techniques. The proposed imputation method based on the MissForest enlarged approximately 2.5 times more molecular structure data compared to the existing data. Based on the imputed dataset of molecular descriptor, the developed data-centric QSAR model achieved approximately 73%, 76% and 92% of prediction performance for Log BCF, Log Koa, and Log P, respectively. Lastly, Tree-SHAP analysis demonstrated that the data-centric-based QSAR model achieved high prediction performance for toxicity information by identifying key molecular descriptors highly correlated with toxicity indices. Therefore, the proposed QSAR model based on the data-centric XAI approach can be extended to predict the toxicity of potential pollutants in emerging printing chemicals, chemical process, semiconductor or display process.

Hierarchical Organization of Embryo Data for Supporting Efficient Search (배아 데이터의 효율적 검색을 위한 계층적 구조화 방법)

  • Won, Jung-Im;Oh, Hyun-Kyo;Jang, Min-Hee;Kim, Sang-Wook
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.48 no.2
    • /
    • pp.16-27
    • /
    • 2011
  • Embryo is a very early stage of the development of multicellular organism such as animals and plants. It is an important research target for studying ontogeny because the fundamental body system of multicellular organism is determined during an embryo state. Researchers in the developmental biology have a large volume of embryo image databases for studying embryos and they frequently search for an embryo image efficiently from those databases. Thus, it is crucial to organize databases for their efficient search. Hierarchical clustering methods have been widely used for database organization. However, most of previous algorithms tend to produce a highly skewed tree as a result of clustering because they do not simultaneously consider both the size of a cluster and the number of objects within the cluster. The skewed tree requires much time to be traversed in users' search process. In this paper, we propose a method that effectively organizes a large volume of embryo image data in a balanced tree structure. We first represent embryo image data as a similarity-based graph. Next, we identify clusters by performing a graph partitioning algorithm repeatedly. We check constantly the size of a cluster and the number of objects, and partition clusters whose size is too large or whose number of objects is too high, which prevents clusters from growing too large or having too many objects. We show the superiority of the proposed method by extensive experiments. Moreover, we implement the visualization tool to help users quickly and easily navigate the embryo image database.

Parallelization Method of Slice-based video CODEC (슬라이스 기반 비디오 코덱 병렬화 기법)

  • Nam, Jung-Hak;Ji, Bong-Il;Jo, Hyun-Ho;Sim, Dong-Gyu;Cho, Dae-Sung
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.47 no.6
    • /
    • pp.48-56
    • /
    • 2010
  • Recently, we need to dramatically speed up real-time video encoding and decoding on mobile devices because complexity of video CODEC is significantly increasing along with the demand for multimedia service of high-quality and high-definition videos by users. A variety of research is conducted for parallelism of video processing using newly developed multi-core platforms. In this paper, we propose a method of parallelism based on slice partition of video compression CODEC. We propose a novel concept of a parallel slice for parallelism and propose a new coding order to be adequate to the parallel slice which keeps high coding efficiency. To minimize synchronization time of multiple parallel slices, we also propose a synchronization method to determinate whether the parallel slice could be independently decoded or not. Experimental results shows that we achieved 27.5% (40.7%) speed-up by parallelism with bit-rate increase of 3.4% (2.7%) for CIF sequences (720p sequences) by implementing the proposed algorithm on the H.264/AVC.

$\pi$/4 shift QPSK with Trellis-Code in Rayleigh Fading Channel (레일레이 페이딩 채널에서 Trellis 부호를 적용한 $\pi$/4 shift QPSK)

  • 김종일;이한섭;강창언
    • The Proceeding of the Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.3 no.2
    • /
    • pp.30-38
    • /
    • 1992
  • In this paper, in order to apply the $\pi$/4 shift QPSK to TCM, we propose the $\pi$/8 shift 8PSK modulation technique and the trellis-coded $\pi$/8 shift 8PSK performing signal set expansion and set partition by phase difference. In addition, the Viterbi decoder with branch metrics of the squared Euclidean distance of the first phase difference as well as the Lth phase difference is introduced in order to improve the bit error rate(BER) performance in differential detection of the trellis-coded $\pi$/8 shift 8 PSK. The proposed Viterbi decoder is conceptually the same as the sliding multiple de- tection by using the branch metric with first and Lth order phase difference. We investigate the performance of the uncoded .pi. /4 shift QPSK and the trellis-coded $\pi$/8 shift 8PSK with or without the Lth phase difference metric in an additive white Gaussian noise (AWGN) and Rayleigh fading channel using the Monte Carlo simulation. The study shows that the $\pi$/4 shift QPSK with the Trellis-code i. e. the trellis-coded $\pi$/8 shift 8PSK is an attractive scheme for power and bandlimited systems and especially, the Viterbi decoder with first and Lth phase difference metrics improves BER performance. Also, the next proposed algorithm can be used in the TC $\pi$/8 shift 8PSK as well as TC MDPSK.

  • PDF