• Title/Summary/Keyword: 부분 중복 제거

Search Result 48, Processing Time 0.038 seconds

An Algorithm of Solution for the Exceptional Field Problem in the Speculative Partial Redundancy Elimination(SPRE) Optimization (추론적 부분 중복 제거의 최적화 예외 영역 문제 해결 알고리즘)

  • Shin, Hyun-Deok;Ahn, Heui-Hak
    • The KIPS Transactions:PartA
    • /
    • v.13A no.6 s.103
    • /
    • pp.489-494
    • /
    • 2006
  • This paper improves the algorithms for Speculative Partial Redundancy Elimination(SPRE) proposed by Knoop et al. This research brings up an issue concerning a field to which SPRE cannot be applied, and suggests a solution to the problem. The Improved SPRE algorithm performs the execution speed optimization based on the information on the execution frequency from profiling and the memory space optimization.

Optimization Using Partial Redundancy Elimination in SSA Form (SSA Form에서 부분 중복 제거를 이용한 최적화)

  • Kim, Ki-Tae;Yoo, Weon-Hee
    • The KIPS Transactions:PartD
    • /
    • v.14D no.2
    • /
    • pp.217-224
    • /
    • 2007
  • In order to determine the value and type statically. CTOC uses the SSA Form which separates the variable according to assignment. The SSA Form is widely being used as the intermediate expression of the compiler for data flow analysis as well as code optimization. However, the conventional SSA Form is more associated with variables rather than expressions. Accordingly, the redundant expressions are eliminated to optimize expressions of the SSA From. This paper defines the partial redundant expression to obtain a more optimized code and also implements the technique for eliminating such expressions.

De-Duplication Performance Test for Massive Data (대용량 데이터의 중복제거(De-Duplication) 성능 실험)

  • Lee, Choelmin;Kim, Jai-Hoon;Kim, Young Gyu
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.11a
    • /
    • pp.271-273
    • /
    • 2012
  • 중복 제거(De-duplication) 여러 데이터를 저장한 스토리지에서 같은 내용을 담고 있는 파일자체나 블록단위의 chunk 등을 찾아 중복된 내용을 제거하여 중복된 부분은 하나의 데이터 단위를 유지함으로써 스토리지 공간을 절약할 수 있다. 본 논문에서는 실험적인 데이터가 아닌 실제 업무 환경에서 적용될만한 대용량의 데이터 백업을 가정한 상황에 대해 중복 제거 기법을 테스트해봄으로써 중복제거율과 성능을 측정하였으며 이를 시각적으로 표현하는 방법을 제안함으로써 평가자 및 사용자가 알아보기 쉽게 하였다.

Resource Redundancy Elimination by Bridging the Semantic Gap in Virtualized Systems (가상 환경에서 시멘틱 갭 연결을 통한 자원 중복성 제거)

  • Kim, Inhyeok;Kim, Taehyoung;Eom, Young Ik
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2011.11a
    • /
    • pp.185-187
    • /
    • 2011
  • 전통적인 가상화 기술들은 하나의 물리 머신에서 서로 다른 운영체제들을 동작시키기 위해 여러 개의 가상 머신을 제공하고 있다. 이러한 가상화 기술들은 운영체제를 소스 수정없이, 혹은 약간의 수정만으로 물리 머신에서 동작하는 것과 동일한 방식으로 동작할 수 있도록 지원하기 때문에 호스트와 게스트 간의 깊은 시멘틱 갭을 발생시킨다. 이러한 시멘틱 갭은 여러 컴퓨팅 자원에 대한 공간 중복과 접근 중복의 중요한 원인이 되고, 이러한 자원 중복은 대규모 가상화 시스템의 확장성에 제약한다. 이러한 자원 중복을 제거하기 위해 컨텐츠 기반 페이지 공유 등 다양한 연구들이 진행되어 왔지만 운영체제 수정없이 지원하는 정책으로 인한 시멘틱 갭은 여전히 제안 기법들을 제한하게 된다. 이에 우리들은 운영체제의 자원 관리 부분을 수정하여 근본적인 시멘틱 갭을 제거함으로써 메모리와 스토리지의 공간 및 접근 중복성을 제거할 수 있는 파일시스템을 제안하였다. 그리고 실험을 통해 기존의 가상 블록 장치를 사용하는 파일시스템과 비교 및 평가하여 제안 시스템이 페이지 캐시 공유 및 스토리지 접근 중복 제거에 효과적임을 입증하였다.

A Clustering File Backup Server Using Multi-level De-duplication (다단계 중복 제거 기법을 이용한 클러스터 기반 파일 백업 서버)

  • Ko, Young-Woong;Jung, Ho-Min;Kim, Jin
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.7
    • /
    • pp.657-668
    • /
    • 2008
  • Traditional off-the-shelf file server has several potential drawbacks to store data blocks. A first drawback is a lack of practical de-duplication consideration for storing data blocks, which leads to worse storage capacity waste. Second drawback is the requirement for high performance computer system for processing large data blocks. To address these problems, this paper proposes a clustering backup system that exploits file fingerprinting mechanism for block-level de-duplication. Our approach differs from the traditional file server systems in two ways. First, we avoid the data redundancy by multi-level file fingerprints technology which enables us to use storage capacity efficiently. Second, we applied a cluster technology to I/O subsystem, which effectively reduces data I/O time and network bandwidth usage. Experimental results show that the requirement for storage capacity and the I/O performance is noticeably improved.

A Study of Method to Restore Deduplicated Files in Windows Server 2012 (윈도우 서버 2012에서 데이터 중복 제거 기능이 적용된 파일의 복원 방법에 관한 연구)

  • Son, Gwancheol;Han, Jaehyeok;Lee, Sangjin
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.27 no.6
    • /
    • pp.1373-1383
    • /
    • 2017
  • Deduplication is a function to effectively manage data and improve the efficiency of storage space. When the deduplication is applied to the system, it makes it possible to efficiently use the storage space by dividing the stored file into chunks and storing only unique chunk. However, the commercial digital forensic tool do not support the file system analysis, and the original file extracted by the tool can not be executed or opened. Therefore, in this paper, we analyze the process of generating chunks of data for a Windows Server 2012 system that can apply deduplication, and the structure of the resulting file(Chunk Storage). We also analyzed the case where chunks that are not covered in the previous study are compressed. Based on these results, we propose the method to collect deduplicated data and reconstruct the original file for digital forensic investigation.

The Consistency Management Using Trees of Replicated Data Items in Partially Replicated Database (부분 중복 데이터베이스에서 중복 데이터의 트리를 이용한 일관성 유지)

  • Bae, Mi-Sook;Hwang, Bu-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.10D no.4
    • /
    • pp.647-654
    • /
    • 2003
  • The replication of data is used to increase its availability and to improve the performance of a system. The distributed database system has to maintain both the database consistency and the replica consistency. This paper proposes an algorithm which resolves the conflict of the operations by using the mechanism based on the structure that the replicas of each data item are hierarchically organized. Each update is propagated along the tree based on the fact that the root of each data item is the primary replica in partially replicated databases. The use of a hierarchy of data may eliminate useless propagation since the propagation can be done only to sites having the replicas. In consequence, the propagation delay of updates may be reduced. By using the timestamp and a compensating transaction, our algorithm resolves the non-serializability problem caused by the conflict of operations that can happen on the way of the update propagation due to the lazy propagation. This resolution also guarantees the data consistency.

Prediction of Unsatisfied Customers Using Machine Learning (기계학습을 이용한 불만족 고객의 예측)

  • Oh, Se-Chang;Choi, Min
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2016.04a
    • /
    • pp.667-670
    • /
    • 2016
  • 많은 기계학습 문제에서 특정 선택 문제는 전체적인 성능을 좌우하는 중요한 부분이다. 이는 불만족 고객의 식별 문제와 같이 수 많은 특징을 사용하는 문제에서 더욱 절실하다. 본 연구에서는 중요한 특징을 찾고 중복성을 제거하기 위한 몇 가지 대표적인 방법들을 불만족 고객의 식별 문제에 적용하였다. 이를 통해 먼저 정보 획득량 지표로 의미 있는 특징들을 선별하고, PCA를 사용해서 남아있는 중복성을 줄이는 방법이 가장 좋은 결과를 얻었다.

Wavelet Denoising Using Region Merging (영역 병합을 이용한 웨이블릿 잡음 제거)

  • Eom Il kyu;Kim Yoo shin
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.3C
    • /
    • pp.119-124
    • /
    • 2005
  • In this paper, we propose a novel algorithm for determining the variable size of locally adaptive window using region-merging method. A region including a denoising point is partitioned to disjoint sub-regions. Locally adaptive window for denoising is obtained by selecting Proper sub-lesions. In our method, nearly arbitrarily shaped window is achieved. Experimental results show that our method outperforms other critically sampled wavelet denoising scheme.

Processing of Code Optimization in CTOC (CTOC에서 코드 최적화 수행)

  • Kim, Ki-Tae;Yoo, Weon-Hee
    • Journal of Korea Multimedia Society
    • /
    • v.10 no.5
    • /
    • pp.687-697
    • /
    • 2007
  • Although the Java bytecode has numerous advantages, there are also shortcomings such as slow execution speed and difficulty in analysis. Therefore, in order for the Java class file to be effectively executed under the execution environment such as the network, it is necessary to convert it into optimized code. We implemented CTOC for transforming to optimized code. In Optimization, SSA Form that distinguish variable by assignment is used to determine value and type statically. Copy propagation, dead code elimination optimization is applied to SSA Form. However, existing SSA Form is related to variable than expression. Therefore, in this paper, to performing optimization to SSA Form expression, after performing copy propagation and dead code elimination, in addition to that, partial redundant expression elimination is performed

  • PDF