• Title/Summary/Keyword: 데이타 캐쉬

Search Result 83, Processing Time 0.028 seconds

Hybrid Scheme of Data Cache Design for Reducing Energy Consumption in High Performance Embedded Processor (고성능 내장형 프로세서의 에너지 소비 감소를 위한 데이타 캐쉬 통합 설계 방법)

  • Shim, Sung-Hoon;Kim, Cheol-Hong;Jhang, Seong-Tae;Jhon, Chu-Shik
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.33 no.3
    • /
    • pp.166-177
    • /
    • 2006
  • The cache size tends to grow in the embedded processor as technology scales to smaller transistors and lower supply voltages. However, larger cache size demands more energy. Accordingly, the ratio of the cache energy consumption to the total processor energy is growing. Many cache energy schemes have been proposed for reducing the cache energy consumption. However, these previous schemes are concerned with one side for reducing the cache energy consumption, dynamic cache energy only, or static cache energy only. In this paper, we propose a hybrid scheme for reducing dynamic and static cache energy, simultaneously. for this hybrid scheme, we adopt two existing techniques to reduce static cache energy consumption, drowsy cache technique, and to reduce dynamic cache energy consumption, way-prediction technique. Additionally, we propose a early wake-up technique based on program counter to reduce penalty caused by applying drowsy cache technique. We focus on level 1 data cache. The hybrid scheme can reduce static and dynamic cache energy consumption simultaneously, furthermore our early wake-up scheme can reduce extra program execution cycles caused by applying the hybrid scheme.

A Concurrency Control Method using Data Group Information in Mobile Computing Environments (이동 컴퓨팅 환경에서 데이타 그룰 정보를 이용한 동시성 제어 방법)

  • Kim, Dae-In;Hwang, Bu-Hyun
    • Journal of KIISE:Databases
    • /
    • v.32 no.3
    • /
    • pp.315-325
    • /
    • 2005
  • In mobile computing environments, a mobile host caches the data items to use the bandwidth efficiently and improve the response time of transactions. If the data items cached in mobile host are updated in the server, the server broadcasts an invalidation report for maintaining the cache consistency of mobile hosts. However, this method has a problem that the response time of mobile transactions can be long since their commit decision is delayed until receiving the invalidation report from the server In this paper, we propose the USR-MT method for improving the response time of mobile transactions. As the UGR-MT method can make a commit decision by using the data group information before receiving the invalidation report, the response time of mobile transactions can be improved. Also our method can improve the cache's efficiency since it prevents all the contents of a cache from being invalidated in the case that the disconnection of a mobile host is longer than the broadcast period of invalidation report.

Advanced Victim Cache with Processor Reuse Information (프로세서의 재사용 정보를 이용하는 개선된 고성능 희생 캐쉬)

  • Kwak Jong Wook;Lee Hyunbae;Jhang Seong Tae;Jhon Chu Shik
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.12
    • /
    • pp.704-715
    • /
    • 2004
  • Recently, a single or multi processor system uses the hierarchical memory structure to reduce the time gap between processor clock rate and memory access time. A cache memory system includes especially two or three levels of caches to reduce this time gap. Moreover, one of the most important things In the hierarchical memory system is the hit rate in level 1 cache, because level 1 cache interfaces directly with the processor. Therefore, the high hit rate in level 1 cache is critical for system performance. A victim cache, another high level cache, is also important to assist level 1 cache by reducing the conflict miss in high level cache. In this paper, we propose the advanced high level cache management scheme based on the processor reuse information. This technique is a kind of cache replacement policy which uses the frequency of processor's memory accesses and makes the higher frequency address of the cache location reside longer in cache than the lower one. With this scheme, we simulate our policy using Augmint, the event-driven simulator, and analyze the simulation results. The simulation results show that the modified processor reuse information scheme(LIVMR) outperforms the level 1 with the simple victim cache(LIV), 6.7% in maximum and 0.5% in average, and performance benefits become larger as the number of processors increases.

Log Indexing for Consistency Control of Cached Spatial Data of Mobile Clients (모바일 클라이언트에 캐쉬된 공간 데이타의 일관성 제어를 위한 로그 인덱싱)

  • 차지태;안경환;전봉기;홍봉희
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.04b
    • /
    • pp.178-180
    • /
    • 2002
  • 모바일 클라이언트(Mobile Client, MC) - 서버 환경에서 캐쉬된 공간 데이타는 용량과 가용성이 크고 긴단절을 가지는 특징이 있다. 이러한 긴 단절을 거친 MC의 일관성 제어시 캐쉬된 데이타 전체가 무효화되어 재전송 하는 것은 비효율적이므로 서버는 변경된 객체 정보를 로그로 기록하여, 캐쉬된 영역에 속하는 최근 변경 정보만을 전송하는 것이 필요하다. 또한 일반적으로 로그의 순차 검색과 불필요한 로그의 누적으로 인해 일관성 제어시 통신 비용이 증가하게 되므로, 서버의 로그 검색 시간과 전송량을 최소화 하기위한 기법이 필요하다. 이를 위해 본 논문에서는 로그를 이용안 공간 로그 인덱스(Spatial Log Index, SLI) 기법과 공간 로그 관리(Spatial Log Management, SLM) 기법을 제시하고 있다.

  • PDF

Efficient Cache Management Scheme in Database based on Block Classification (블록 분류에 기반한 데이타베이스의 효율적 캐쉬 관리 기법)

  • Sin, Il-Hoon;Koh, Kern
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.7
    • /
    • pp.369-376
    • /
    • 2002
  • Although LRU is not adequate for database that has non-uniform reference pattern, it has been adopted in most database systems due to the absence of the proper alternative. We analyze database block reference pattern with the realistic database trace. Based on this analysis, we propose a new cache replacement policy. Trace analysis shows that extremely non-popular blocks take up about 70 % of the entire blocks. The influence of recency on blocks' re-reference likelihood is at first strong due to temporal locality, however, it rapidly decreases and eventually becomes negligible as stack distance increases. Based on this observation, RCB(Reference Characteristic Based) cache replacement policy, which we propose in this paper, classifies the entire blocks into four block groups by blocks' recency and re-reference likelihood, and operates different priority evaluation methods for each block group. RCB policy evicts non-popular blocks more quickly than the others and evaluates the priority of the block by frequency that has not been referenced for a long time. In a trace-driven simulation, RCB delivers a better performance than the existing polices(LRU, 2Q, LRU-K, LRFU). Especially compared to LRU. It reduces miss count by 5~l2.7%. Time complexity of RCB is O(1), which is the same with LRU and 2Q and superior to LRU-K(O(log$_2$N)) and LRFU(O(l) ~ O(log$_2$N)).

Client Cache Management Scheme For Data Broadcasting Environments (LRU-CFP: 데이터 방송 환경을 위한 클라이언트 캐쉬 관리 기법)

  • Kwon, Hyeok-Min
    • The KIPS Transactions:PartD
    • /
    • v.10D no.6
    • /
    • pp.961-970
    • /
    • 2003
  • In data broadcasting environments, the server periodically broadcasts data items in the broadcast channel. When each client wants to access any data item, it should monitor the broadcast channel and wait for the desired item to arrive. Client data caching is a very effective technique for reducing the time spent waiting for the desired item to be broadcastted. This paper proposes a new client cache management scheme, named LRU-CFP, to reduce this waiting time ans evaluates its performance on the basis of a simulation model. The performance results indicate that LRU-CFP scheme shows superior performance over LRU, GRAY and CF in the average response time.

A Strategy for Efficiently Maintaining Cache Consistency in Mobile Computing Environments of the Asynchronous Broadcasting, (비동기적 방송을 하는 이동 컴퓨팅 환경에서 효율적인 캐쉬 일관성 유지 정책)

  • 김대옹;박성배;김길삼;황부현
    • Journal of the Korea Society of Computer and Information
    • /
    • v.4 no.3
    • /
    • pp.78-92
    • /
    • 1999
  • In mobile computing environments, to efficiently use the narrow bandwidth of wireless networks a mobile host caches the data that are frequently accessed. To guarantee the correctness of the mobile transaction, the data cached in a mobile host must be consistent with the data in a server. This paper proposes a new strategy which maintains cache consistency efficiently when the data cached in a mobile host are inconsistent with the data in a server by the mobility of the mobile host at the asynchronous mobile environment. In this strategy, the size of the invalidation message is relatively small and is independent of the number of data to be invalidated under conditions of variable update rates/patterns. So this strategy uses the narrow bandwidth of wireless networks efficiently and reduces the communication cost.

A Cache Consistency Control for B-Tree Indices in a Database Sharing System (데이타베이스 공유 시스템에서 B-트리 인덱스를 위한 캐쉬 일관성 제어)

  • On, Gyeong-O;Jo, Haeng-Rae
    • The KIPS Transactions:PartD
    • /
    • v.8D no.5
    • /
    • pp.593-604
    • /
    • 2001
  • A database sharing system (DSS) refers to a system for high performance transaction processing. In the DSS, the processing nodes are coupled via a high speed network and share a common database at the disk level. Each node has a local memory and a separate copy of operating system. To reduce the number of disk accesses, the node caches data pages and index pages in its memory buffer. In general, B-tree index pages are accessed more often and thus cached at more processing nodes, than their corresponding data pages. There are also complicated operations in the B-tree such as Fetch, Fetch Next, Insertion and Deletion. Therefore, an efficient cache consistency scheme supporting high level concurrency is required. In this paper, we propose cache consistency schemes using identifiers of index pages and page_LSN of leaf page. The propose schemes can improve the system throughput by reducing the required message traffic between nodes and index re-traversal.

  • PDF

2Q-CFP: A Client Cache Management Scheme for Broadcast-based Information Systems (2Q-CFP: 방송에 기초한 정보 시스템을 위한 클라이언트 캐쉬 관리 기법)

  • 권혁민
    • Journal of KIISE:Databases
    • /
    • v.30 no.6
    • /
    • pp.561-572
    • /
    • 2003
  • Broadcast-based data delivery has attracted a lot of attention as an efficient way of disseminating data to very large client populations. The main motivation of broadcast-based information systems (BBISs) is that the number of clients that they serve can grow arbitrarily large without any effect on their performance. The performance of BBISs depends mainly on client caching strategies and on data broadcast scheduling mechanisms. This paper addresses the former issue and proposes a new client cache management scheme, named 2Q-CFP, that is suitable to BBISs. This paper also evaluates the performance of 2Q-CFP on the basis of a simulation model. The performance results indicate that 2Q-CFP scheme shows superior performances over GRAY, LRU and CF in the average response time.

Extending SQL CLI for Large Object Caching (LOB 캐쉬를 위한 SQL CLI의 확장)

  • 이종민;강현철
    • The KIPS Transactions:PartD
    • /
    • v.8D no.1
    • /
    • pp.1-9
    • /
    • 2001
  • The SQL CLl (Call Level Interface) is a database application programming interface (API) that fits the client-server environment, and its de facto standards such as ODBC and JDBC are widely employed in various applications. The current SQL CLl standards, however, do not specify enough features for efficient retrieval of large objects (LOBs) that constitute the multimedia data. In this paper, we propose the LOB caching through the SQL CLl and describe the extension of the SQL eLl to achieve such a goal. We implement our proposal by extending the SQL CLl library for BADA - II DBMS developed at ETRI, and evaluate the performance of LOB retrieval through the cache compared to that of LOB retrieval conducted solely with the functions specified in the current SQL CLl standards.

  • PDF