• Title/Summary/Keyword: binary-tree

Search Result 298, Processing Time 0.028 seconds

A method for high-speed event processing in the real-time RFID middleware systems (실시간 RFID 미들웨어에서의 태그 데이터 고속 필터링 방법)

  • Park, Mi Sun;Kim, Yong Jin;Ryu, Minsoo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2009.11a
    • /
    • pp.435-436
    • /
    • 2009
  • RFID 시스템의 미들웨어는 태그에서 생성된 방대한 양의 데이터를 리더를 통해 전달받는다. RFID 미들웨어는 이러한 데이터를 정제하여 응용 애플리케이션에 전달하는 기능을 담당한다. 하지만 태그 데이터 정제 과정에서 발생되는 지연 시간은 RFID 미들웨의 응답성을 저하시킨다. 본 논문은 EPCglobal 의 RFID 미들웨어 표준인 ALE 에 의거하여 태그 데이터에 대한 다수의 필터링 조건들이 주어진 RFID 미들웨어 환경에서 실시간으로 수집되는 대용량의 태그에 대한 고속 필터링 엔진을 설계한다. 이를 위하여 Intermediate node 들이 key 값을 저장하는 Binary Search Tree 형태를 구성하여 태그를 필터링하는 방법을 제안한다. 결과로써 기존의 순차적인 RFID 데이터 필터링에 비해 고속의 필터링 성능을 보이며 특히 필터의 수가 증가할수록 필터링의 효율이 높아짐을 보인다.

Binary Tree Line Calculator (이진 트리 라인 계산기)

  • Ra, Gyungjun;Lee, Deokwoo
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • fall
    • /
    • pp.149-151
    • /
    • 2021
  • 최근 Kotlin, Swift 와 같은 목표 기능에대해 최적화된 다양한 고급 컴퓨터 언어들이 출범함에 따라 이에대한 중요도와 이해력이 필요하다. 그에따라 고급 컴퓨터 언어 제작의 프로토타입 제작에 의의를 두고, 전세계 사람들에게 통용되는 사칙연산에 기반한 수학기호 코드 언어를 해석하는 프로그램을 구현한다. 단순한 계산기라기보다는. 수학기호 문자를 컴파일하여 프로세스로 변환하는 컴파일러로 접근하고 특성파악, 구상, 구현 과정을거쳐 본 프로그램을 제작하고 고급 컴퓨터 언어에 대한 저자의 이해를 서술한다.

  • PDF

Horizon Run 5 Black Hole Populations and Pulsar Timing Array

  • Kim, Chunglee;Park, Hyo Sun;Kim, Juhan;Lommen, Andrea
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.46 no.2
    • /
    • pp.45.2-45.2
    • /
    • 2021
  • Merging of two supermassive black holes would generate gravitational waves that can be detected by the Pulsar Timing Array (PTA) in the nHz band. In order to assess the plausibility of GW detection with PTA and to develop the data analysis scheme, it is important to understand the underlying properties of black holes and black hole binaries. In this work, we present mass and redshift distributions of black hole mergers using the Horizon Run 5 (HR5) data and discuss their implications for GW detection. We find a general conjecture about the black hole merger tree is true with the Horizon Run 5. For example, a) relatively lighter black holes merge at higher redshifts and b) binary mergers do contribute to the formation of more massive black holes toward low redshifts. We also present our plan to use the black hole properties extracted from the HR5 data in order to generate simulated GW signals to be injected into actual PTA data analysis pipelines. Mass and distance obtained from the HR5 would be key ingredients to generate a more realistic PTA source data set.

  • PDF

Image Compression Using DCT Map FSVQ and Single - side Distribution Huffman Tree (DCT 맵 FSVQ와 단방향 분포 허프만 트리를 이용한 영상 압축)

  • Cho, Seong-Hwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.10
    • /
    • pp.2615-2628
    • /
    • 1997
  • In this paper, a new codebook design algorithm is proposed. It uses a DCT map based on two-dimensional discrete cosine of transform (2D DCT) and finite state vector quantizer (FSVQ) when the vector quantizer is designed for image transmission. We make the map by dividing input image according to edge quantity, then by the map, the significant features of training image are extracted by using the 2D DCT. A master codebook of FSVQ is generated by partitioning the training set using binary tree based on tree-structure. The state codebook is constructed from the master codebook, and then the index of input image is searched at not master codebook but state codebook. And, because the coding of index is important part for high speed digital transmission, it converts fixed length codes to variable length codes in terms of entropy coding rule. The huffman coding assigns transmission codes to codes of codebook. This paper proposes single-side growing huffman tree to speed up huffman code generation process of huffman tree. Compared with the pairwise nearest neighbor (PNN) and classified VQ (CVQ) algorithm, about Einstein and Bridge image, the new algorithm shows better picture quality with 2.04 dB and 2.48 dB differences as to PNN, 1.75 dB and 0.99 dB differences as to CVQ respectively.

  • PDF

Combined Image Retrieval System using Clustering and Condensation Method (클러스터링과 차원축약 기법을 통합한 영상 검색 시스템)

  • Lee Se-Han;Cho Jungwon;Choi Byung-Uk
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.43 no.1 s.307
    • /
    • pp.53-66
    • /
    • 2006
  • This paper proposes the combined image retrieval system that gives the same relevance as exhaustive search method while its performance can be considerably improved. This system is combined with two different retrieval methods and each gives the same results that full exhaustive search method does. Both of them are two-stage method. One uses condensation of feature vectors, and the other uses binary-tree clustering. These two methods extract the candidate images that always include correct answers at the first stage, and then filter out the incorrect images at the second stage. Inasmuch as these methods use equal algorithm, they can get the same result as full exhaustive search. The first method condenses the dimension of feature vectors, and it uses these condensed feature vectors to compute similarity of query and images in database. It can be found that there is an optimal condensation ratio which minimizes the overall retrieval time. The optimal ratio is applied to first stage of this method. Binary-tree clustering method, searching with recursive 2-means clustering, classifies each cluster dynamically with the same radius. For preserving relevance, its range of query has to be compensated at first stage. After candidate clusters were selected, final results are retrieved by computing similarities again at second stage. The proposed method is combined with above two methods. Because they are not dependent on each other, combined retrieval system can make a remarkable progress in performance.

A Hybrid Multi-Level Feature Selection Framework for prediction of Chronic Disease

  • G.S. Raghavendra;Shanthi Mahesh;M.V.P. Chandrasekhara Rao
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.12
    • /
    • pp.101-106
    • /
    • 2023
  • Chronic illnesses are among the most common serious problems affecting human health. Early diagnosis of chronic diseases can assist to avoid or mitigate their consequences, potentially decreasing mortality rates. Using machine learning algorithms to identify risk factors is an exciting strategy. The issue with existing feature selection approaches is that each method provides a distinct set of properties that affect model correctness, and present methods cannot perform well on huge multidimensional datasets. We would like to introduce a novel model that contains a feature selection approach that selects optimal characteristics from big multidimensional data sets to provide reliable predictions of chronic illnesses without sacrificing data uniqueness.[1] To ensure the success of our proposed model, we employed balanced classes by employing hybrid balanced class sampling methods on the original dataset, as well as methods for data pre-processing and data transformation, to provide credible data for the training model. We ran and assessed our model on datasets with binary and multivalued classifications. We have used multiple datasets (Parkinson, arrythmia, breast cancer, kidney, diabetes). Suitable features are selected by using the Hybrid feature model consists of Lassocv, decision tree, random forest, gradient boosting,Adaboost, stochastic gradient descent and done voting of attributes which are common output from these methods.Accuracy of original dataset before applying framework is recorded and evaluated against reduced data set of attributes accuracy. The results are shown separately to provide comparisons. Based on the result analysis, we can conclude that our proposed model produced the highest accuracy on multi valued class datasets than on binary class attributes.[1]

Calculation of Top Event Probability of Fault Tree using BDD (BDD를 이용한 사고수목 정상사상확률 계산)

  • Cho, Byeong Ho;Yum, Byeoungsoo;Kim, Sangahm
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.20 no.3
    • /
    • pp.654-662
    • /
    • 2016
  • As the number of gates and basic events in fault trees increases, it becomes difficult to calculate the exact probability of the top event. In order to overcome this difficulty the BDD methodology can be used to calculate the exact top event probability for small and medium size fault trees in short time. Fault trees are converted to BDD by using CUDD library functions and a failure path search algorithm is proposed to calculate the exact top event probability. The backward search algorithm is more efficient than the forward one in finding failure paths and in the calculation of the top event probability. This backward search algorithm can reduce searching time in the identification of disjoint failure paths from BDD and can be considered as an effective tool to find the cut sets and the minimal cut sets for the given fault trees.

A Fast Transmission of Mobile Agents Using Binomial Trees (바이노미얼 트리를 이용한 이동 에이전트의 빠른 전송)

  • Cho, Soo-Hyun;Kim, Young-Hak
    • The KIPS Transactions:PartA
    • /
    • v.9A no.3
    • /
    • pp.341-350
    • /
    • 2002
  • As network environments have been improved and the use of internet has been increased, mobile agent technologies are widely used in the fields of information retrieval, network management, electronic commerce, and parallel/distributed processing. Recently, a lot of researchers have studied the concepts of parallel/distributed processing based on mobile agents. SPMD is the parallel processing method which transmits a program to all the computers participated in parallel environment, and performs a work with different data. Therefore, to transmit fast a program to all the computers is one of important factors to reduce total execution time. In this paper, we consider the parallel environment consisting of mobile agents system, and propose a new method which transmits fast a mobile agent code to all the computers using binomial trees in order to efficiently perform the SPMD parallel processing. The proposed method is compared with another ones through experimental evaluation on the IBM's Aglets, and gets greatly better performance. Also this paper deals with fault tolerances which can be occurred in transmitting a mobile agent using binomial trees.

A Study on the Extraction of Road & Vehicles Using Image Processing Technique (영상처리 기술을 이용한 도로 및 차량 추출 기법에 관한 연구)

  • Ga, Chill-O;Byun, Young-Gi;Yu, Ki-Yun;Kim, Yong-Il
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.13 no.4 s.34
    • /
    • pp.3-9
    • /
    • 2005
  • The extraction of traffic information based on image processing is under broad research recently because the method based on image processing takes less cost and effort than the traditional method based on physical equipment. The main purpose of the algorithm based on image processing is to extract vehicles from an image correctly. Before the extraction, the algorithm needs the pre-processing such as background subtraction and binary image thresholding. During the pre-processing much noise is brought about because roadside tree and passengers in the sidewalk as well as vehicles are extracted as traffic flow. The noise undermines the overall accuracy of the algorithm. In this research, most of the noise could be removed by extracting the exact road area which does not include sidewalk or roadside tree. To extract the exact road area, traffic lanes in the image were used. Algorithm speed also increased. In addition, with the ratio between the sequential images, the problem caused by vehicles' shadow was minimized.

  • PDF

Resynthesis of Logic Gates on Mapped Circuit for Low Power (저전력 기술 매핑을 위한 논리 게이트 재합성)

  • 김현상;조준동
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.35C no.11
    • /
    • pp.1-10
    • /
    • 1998
  • The advent of deep submicron technologies in the age of portable electronic systems creates a moving target for CAB algorithms, which now need to reduce power as well as delay and area in the existing design methodology. This paper presents a resynthesis algorithm for logic decomposition on mapped circuits. The existing algorithm uses a Huffman encoding, but does not consider glitches and effects on logic depth. The proposed algorithm is to generalize the Huffman encoding algorithm to minimize the switching activity of non-critical subcircuits and to preserve a given logic depth. We show how to obtain a transition-optimum binary tree decomposition for AND tree with zero gate delay. The algorithm is tested using SIS (logic synthesizer) and Level-Map (LUT-based FPGA lower power technology mapper) and shows 58%, 8% reductions on power consumptions, respectively.

  • PDF