• Title/Summary/Keyword: 동적 정보

Search Result 5,261, Processing Time 0.144 seconds

Low Cost and Acceptable Delay Unicast Routing Algorithm Based on Interval Estimation (구간 추정 기반의 지연시간을 고려한 저비용 유니캐스트 라우팅 방식)

  • Kim, Moon-Seong;Bang, Young-Cheol;Choo, Hyun-Seung
    • The KIPS Transactions:PartC
    • /
    • v.11C no.2
    • /
    • pp.263-268
    • /
    • 2004
  • The end-to-end characteristic Is an important factor for QoS support. Since network users and required bandwidths for applications increase, the efficient usage of networks has been intensively investigated for the better utilization of network resources. The distributed adaptive routing is the typical routing algorithm that is used in the current Internet. The DCLC(Delay Constrained 1.east Cost) path problem has been shown to be NP-hard problem. The path cost of LD path is relatively more expensive than that of LC path, and the path delay of LC path is relatively higher than that of LD path in DCLC problem. In this paper, we investigate the performance of heuristic algorithm for the DCLC problem with new factor which is probabilistic combination of cost and delay. Recently Dr. Salama proposed a polynomial time algorithm called DCUR. The algorithm always computes a path, where the cost of the path is always within 10% from the optimal CBF. Our evaluation showed that heuristic we propose is more than 38% better than DCUR with cost when number of nodes is more than 200. The new factor takes in account both cost and delay at the same time.

A Region-based Comparison Algorithm of k sets of Trapezoids (k 사다리꼴 셋의 영역 중심 비교 알고리즘)

  • Jung, Hae-Jae
    • The KIPS Transactions:PartA
    • /
    • v.10A no.6
    • /
    • pp.665-670
    • /
    • 2003
  • In the applications like automatic masks generation for semiconductor production, a drawing consists of lots of polygons that are partitioned into trapezoids. The addition/deletion of a polygon to/from the drawing is performed through geometric operations such as insertion, deletion, and search of trapezoids. Depending on partitioning algorithm being used, a polygon can be partitioned differently in terms of shape, size, and so on. So, It's necessary to invent some comparison algorithm of sets of trapezoids in which each set represents interested parts of a drawing. This comparison algorithm, for example, may be used to verify a software program handling geometric objects consisted of trapezoids. In this paper, given k sets of trapezoids in which each set forms the regions of interest of each drawing, we present how to compare the k sets to see if all k sets represent the same geometric scene. When each input set has the same number n of trapezoids, the algorithm proposed has O(2$^{k-2}$ $n^2$(log n+k)) time complexity. It is also shown that the algorithm suggested has the same time complexity O( $n^2$ log n) as the sweeping-based algorithm when the number k(<< n) of input sets is small. Furthermore, the proposed algorithm can be kn times faster than the sweeping-based algorithm when all the trapezoids in the k input sets are almost the same.

Online Reorganization of B+ tree in a Scalable and Highly Available Database Cluster (확장 가능한 고가용 데이터베이스 클러스터에서 B+ 트리 색인의 온-라인 재조직 기법)

  • Lee, Chung-Ho;Bae, Hea-Young
    • The KIPS Transactions:PartD
    • /
    • v.9D no.5
    • /
    • pp.801-812
    • /
    • 2002
  • On-line reorganization in a shared nothing database cluster is crucial to the performance of the database system in a dynamic environment like WWW where the number of users grows rapidly and changing access patterns may exhibit high skew. In the existing method of on-line reorganization have a drawback that needs excessive data migrations in case more than two nodes within a cluster have overload at the same time. In this paper, we propose an advanced B$^{+}$ tree based on-line reorganization method that solves data skew on multi-nodes. Our method facilitates fast and efficient data migration by including spare nodes that are added to cluster through on-line scaling. Also we apply CSB$^{+}$ tree (Cache Sensitive B$^{+}$ tree) to our method instead of B$^{+}$ tree for fast select and update queries. We conducted performance study and implemented the method on Ultra Fault-Tolerant Database Cluster developed for high scalability and availability. Empirical results demonstrate that our proposed method is indeed effective and fast than the existing method. method.

Improving Naïve Bayes Text Classifiers with Incremental Feature Weighting (점진적 특징 가중치 기법을 이용한 나이브 베이즈 문서분류기의 성능 개선)

  • Kim, Han-Joon;Chang, Jae-Young
    • The KIPS Transactions:PartB
    • /
    • v.15B no.5
    • /
    • pp.457-464
    • /
    • 2008
  • In the real-world operational environment, most of text classification systems have the problems of insufficient training documents and no prior knowledge of feature space. In this regard, $Na{\ddot{i}ve$ Bayes is known to be an appropriate algorithm of operational text classification since the classification model can be evolved easily by incrementally updating its pre-learned classification model and feature space. This paper proposes the improving technique of $Na{\ddot{i}ve$ Bayes classifier through feature weighting strategy. The basic idea is that parameter estimation of $Na{\ddot{i}ve$ Bayes considers the degree of feature importance as well as feature distribution. We can develop a more accurate classification model by incorporating feature weights into Naive Bayes learning algorithm, not performing a learning process with a reduced feature set. In addition, we have extended a conventional feature update algorithm for incremental feature weighting in a dynamic operational environment. To evaluate the proposed method, we perform the experiments using the various document collections, and show that the traditional $Na{\ddot{i}ve$ Bayes classifier can be significantly improved by the proposed technique.

An Organization Framework for Role-based Adaptive Distributed Systems (역할기반 적응형 분산 시스템을 위한 조직 프레임워크)

  • Hwang, Seong-Yun;Jung, Jong-Yun;Lee, Jung-Tae;Ryu, Ki-Yeol
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.2 no.9
    • /
    • pp.369-380
    • /
    • 2013
  • Recently, role-based distributed system models have been proposed to support adaptive interactions in ubiquitous application environment. A Role-based distributed model regards an application as an organization composed of roles, and separate players running role processes from the roles. When an application is running, it binds a role with a player, and the player runs dynamically assigned role processes provided by an application for supporting adaptability. However, there has not been much attention on researches about development and runtime environment for role-based distributed systems. In this paper we suggest an application framework as an environment for developing and executing role-base distributed systems. The application framework is divided into two parts: an organization framework to manage and construct an organization composed of roles necessary in the application, and a player framework to provide running environment for players. In this paper, we focus on the organization framework which supports the creation and management of organizations, directory service for players and allocation of players to roles, and message brokering between roles and players. The proposed framework makes developers to be able to develop highly adaptive distributed systems in the ubiquitous environment.

A Hybrid Mapping Technique for Logical Volume Manager in SAN Environments (SAN 논리볼륨 관리자를 위한 혼합 매핑 기법)

  • 남상수;피준일;송석일;유재수;최영희;이병엽
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.10 no.1
    • /
    • pp.99-113
    • /
    • 2004
  • A new architecture called SAN(Storage Area Network) was developed in response to the requirements of high availability of data, scalable growth, and system performance. In order to use SAN more efficiently, most of SAN operating softwares support storage virtualization concepts that allow users to view physical storage devices attached to SAN as a large volume virtually h logical volume manager plays a key role in storage virtualization. It realizes the storage virtualization by mapping logical addresses to physical addresses. A logical volume manager also supports a snapshot that preserves a volume image at certain time and on-line reorganization to allow users to add/remove storage devices to/from SAN even while the system is running. To support the snapshot and the on-line reorganization, most logical volume managers have used table based mapping methods. However, it is very difficult to manage mapping table because the mapping table is large in proportion to a storage capacity. In this paper, we design and implement an efficient and flexible hybrid mapping method based on mathematical equations. The mapping method in this paper supports a snapshot and on-line reorganization. The proposed snapshot and on-line reorganization are performed on the reserved area which is separated from data area of a volume. Due to this strategy normal I/O operations are not affected by snapshot and reorganization. Finally, we show the superiority of our proposed mapping method through various experiments.

Extraction of Classes and Hierarchy from Procedural Software (절차지향 소프트웨어로부터 클래스와 상속성 추출)

  • Choi, Jeong-Ran;Park, Sung-Og;Lee, Moon-Kun
    • Journal of KIISE:Software and Applications
    • /
    • v.28 no.9
    • /
    • pp.612-628
    • /
    • 2001
  • This paper presents a methodology to extract classes and inheritance relations from procedural software. The methodology is based on the idea of generating all groups of class candidates, based on the combinatorial groups of object candidates, and their inheritance with all possible combinations and selecting a group of object candidates, and their inheritance with all possible combinations and selecting a group with the best or optimal combination of candidates with respect to the degree of relativity and similarity between class candidates in the group and classes in a domain model. The methodology has innovative features in class candidates in the group and classes in a domain model. The methodology has innovative features in class and inheritance extraction: a clustering method based on both static (attribute) and dynamic (method) clustering, the combinatorial cases of grouping class candidate cases based on abstraction, a signature similarity measurement for inheritance relations among n class candidates or m classes, two-dimensional similarity measurement for inheritance relations among n class candidates or m classes, two-dimensional similarity measurement, that is, the horizontal measurement for overall group similarity between n class candidates and m classes, and the vertical measurement for specific similarity between a set of classes in a group of class candidates and a set of classes with the same class hierarchy in a domain model, etc. This methodology provides reengineering experts with a comprehensive and integrated environment to select the best or optimal group of class candidates.

  • PDF

Partitioning and Merging an Index for Efficient XML Keyword Search (효율적 XML키워드 검색을 인덱스 분할 및 합병)

  • Kim, Sung-Jin;Lee, Hyung-Dong;Kim, Hyoung-Joo
    • Journal of KIISE:Databases
    • /
    • v.33 no.7
    • /
    • pp.754-765
    • /
    • 2006
  • In XML keyword search, a search result is defined as a set of the smallest elements (i.e., least common ancestors) containing all query keywords and a granularity of indexing is an XML element instead of a document. Under the conventional index structure, all least common ancestors produced by the combination of the elements, each of which contains a query keyword, are considered as a search result. In this paper, to avoid unnecessary operations of producing the least common ancestors and reduce query process time, we describe a way to construct a partitioned index composed of several partitions and produce a search result by merging those partitions if necessary. When a search result is restricted to be composed of the least common ancestors whose depths are higher than a given minimum depth, under the proposed partitioned index structure, search systems can reduce the query process time by considering only combinations of the elements belonging to the same partition. Even though the minimum depth is not given or unknown, search systems can obtain a search result with the partitioned index, which requires the same query process time to obtain the search result with non-partitioned index. Our experiment was conducted with the XML documents provided by the DBLP site and INEX2003, and the partitioned index could reduce a substantial amount of query processing time when the minimum depth is given.

ea­-RED++: Adding Prediction Algorithm for ea­-RED Router Buffer Management Algorithm (ea-­RED++ : 예측 알고리즘을 적용한 ea-­RED 알고리즘)

  • Lee, Jong-Hyun;Lim, Hye-Young;Hwang, Jun;Kim, Young-Chan
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2003.10c
    • /
    • pp.298-300
    • /
    • 2003
  • ea­RED(Efficient Adaptive RED)[1][2]는 다수의 TCP 커넥션이 경쟁하는 병목구간에서 인터넷 라우터 버퍼를 능동적으로 관리하는 다양한 AQM(Active Queue Management) 알고리즘 중의 하나로 RED 라우터 버퍼 관리 알고리즘의 성능을 개선한 라우터 버퍼 관리 알고리즘이다. RED 라우터가 TD 라우터와 같은 네트워크 퍼포먼스를 유지하면서 TCP 커넥션 간 페어니스를 향상시키기 위해서는 link bandwidth. active 커넥션 수. congestion level 등에 대한 네트워크 상태를 고려하여 파라미터에 적절한 값을 설정해야만 한다. 문제는 다이내믹하게 변하는 네트워크 상황에 적합한 파라미터 값을 초기에 설정해주는 것이 매우 어렵다는 점이다. [3]. ea­RED는 max threshold와 min threshold 값을 네트워크 상황에 따라 동적으로 조절함으로써 이런 문제를 해결했고, 기존 RED에 비해 라우터 버퍼는 50% 정도만 사용하면서도, 페어니스 인덱스(Fairness Index)[4]가 최대 41.42% 개선되었다. [1] [2] 그러나 송신 TCP 커넥션의 수가 늘어날수록 성능향상에 대한 효과가 감소되었고, 드롭 패킷수가 TD나 RED 라우터 버퍼관리 알고리즘에 비해 많았기 때문에 라우터의 출력(output) 총 패킷 용량이 최대 약 2.3% 정도 TD나 RED 라우터 버퍼관리 알고리즘에 비해 적었다. 이 부분을 개선하기 위해 기존 ea­RED 알고리즘에 LR_Lines 예측 알고리즘을 적용한 ea­RED++ 알고리즘을 구현하였고, 실험 결과 페어니스 인덱스는 기존 ea­RED에 비해 최대 약 30% 정도 향상되었고, 총 output 패킷 용량의 손실률은 최대 50%정도 감소하여 기존 ea­RED에 비해 향상된 성능을 보여주었다.웍스 네트워크상의 다양한 디바이스들간의 네트워크 다양화와 분산화 기능을 얻을 수 있었고, 기존의 고가의 해외 솔루션인 Echelon사의 LonMaker 소프트웨어를 사용하지 않고도 국내의 순수 솔루션인 리눅스 기반의 LonWare 3.0 다중 바인딩 기능을 통해 저 비용으로 홈 네트워크 구성 관리 서버 시스템 개발에 대한 비용을 줄일 수 있다. 기대된다.e 함량이 대체로 높게 나타났다. 점미가 수가용성분에서 goucose대비 용출함량이 고르게 나타나는 경향을 보였고 흑미는 알칼리가용분에서 glucose가 상당량(0.68%) 포함되고 있음을 보여주었고 arabinose(0.68%), xylose(0.05%)도 다른 종류에 비해서 다량 함유한 것으로 나타났다. 흑미는 총식이섬유 함량이 높고 pectic substances, hemicellulose, uronic acid 함량이 높아서 콜레스테롤 저하 등의 효과가 기대되며 고섬유식품으로서 조리 특성 연구가 필요한 것으로 사료된다.리하였다. 얻어진 소견(所見)은 다음과 같았다. 1. 모년령(母年齡), 임신회수(姙娠回數), 임신기간(姙娠其間), 출산시체중등(出産時體重等)의 제요인(諸要因)은 주산기사망(周産基死亡)에 대(對)하여 통계적(統計的)으로 유의(有意)한 영향을 미치고 있어 $25{\sim}29$세(歲)의 연령군에서, 2번째 임신과 2번째의 출산에서 그리고 만삭의 임신 기간에, 출산시체중(出産時體重) $3.50{\sim}3.99kg$사이의 아이에서 그 주산기사망률(周産基死亡率)이 각각 가장 낮았다. 2. 사산(死産)과 초생아사망(初生兒死亡)을 구분(區分)하여 고려해 볼때 사산(死産)은 모성(母性)의 임신력(姙娠歷)과 매우 밀접한 관련이 있는 것으로 사료(思料)되었고 초생아사망(初生兒死亡)은 미숙아(未熟兒)와 이에 관련된 병

  • PDF

News Video Shot Boundary Detection using Singular Value Decomposition and Incremental Clustering (특이값 분해와 점증적 클러스터링을 이용한 뉴스 비디오 샷 경계 탐지)

  • Lee, Han-Sung;Im, Young-Hee;Park, Dai-Hee;Lee, Seong-Whan
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.2
    • /
    • pp.169-177
    • /
    • 2009
  • In this paper, we propose a new shot boundary detection method which is optimized for news video story parsing. This new news shot boundary detection method was designed to satisfy all the following requirements: 1) minimizing the incorrect data in data set for anchor shot detection by improving the recall ratio 2) detecting abrupt cuts and gradual transitions with one single algorithm so as to divide news video into shots with one scan of data set; 3) classifying shots into static or dynamic, therefore, reducing the search space for the subsequent stage of anchor shot detection. The proposed method, based on singular value decomposition with incremental clustering and mercer kernel, has additional desirable features. Applying singular value decomposition, the noise or trivial variations in the video sequence are removed. Therefore, the separability is improved. Mercer kernel improves the possibility of detection of shots which is not separable in input space by mapping data to high dimensional feature space. The experimental results illustrated the superiority of the proposed method with respect to recall criteria and search space reduction for anchor shot detection.