• 제목/요약/키워드: Data Memory

검색결과 3,302건 처리시간 0.025초

Clothing Memory in the Apparel Purchase Behavior (의복 구매 행동에 영향을 미치는 의복 관련 기억)

  • 유연실;이은영
    • Journal of the Korean Society of Clothing and Textiles
    • /
    • 제26권7호
    • /
    • pp.982-992
    • /
    • 2002
  • The purpose of this study is to investigate the clothing memory which affects apparel purchase behavior. The construction of clothing memory, the relationship between clothing memory and apparel purchase behavior, and the difference of clothing memory among persons by demographic variables are examined 746 female apparel consumers answer the questionnaire and data are analyzed by factor analysis, canonical correlation, Pearson correlation, and t-test. The result of this study are as follows; 1. There are three dimensions in the clothing memory which memory apparel purchase behavior; general memory about clothing, memory of appropriate clothing for oneself and memory of apparel purchase. 2. nothing memory affects apparel purchase behavior. General memory about clothing plays the most important role in the apparel purchase behavior especially in the amount of apparel purchase. 3. Unmarried young women with high income, high education level, and decent job have large amount of clothing memory.

Study on the Performance Evaluation and Analysis of Mobile Cache Memory

  • Lee, Sangmin;Kim, Jongwan;Kim, Ji Young;Oh, Dukshin
    • Journal of the Korea Society of Computer and Information
    • /
    • 제25권6호
    • /
    • pp.99-107
    • /
    • 2020
  • In this paper, we analyze the characteristics of mobile cache, which is used to improve the data access speed when executing applications on mobile devices, and verify the importance of mobile cache through a cache data access experiment. The mobile device market has grown at a fast pace over the past decade; however, battery limitations and size, price considerations restrict the usage of fast hardware. Thus, their performance are supplemented by using a memory buffer structure such as the cache memory. The analysis mainly focuses on cache size, hierarchical structure of cache, cache replacement policy, and the effect these features has on mobile performance. For the experimental data, we applied a data set from a microprocessor system study, originally used to test the cache performance. In the experimental results, the average data access speed on a mobile device showed a performance improvement of up to 10 times with the presence of cache memory than without. Accordingly, the cache memory was helpful for the performance improvement of a mobile device when the specifications were identical.

Design of a Data Analysis System with Wireless LAN for a Train Operating (열차운행시 무선LAN을 적용한 데이터 분석시스템의 설계)

  • 이우철;서상준;박계서
    • Proceedings of the KSR Conference
    • /
    • 한국철도학회 2000년도 춘계학술대회 논문집
    • /
    • pp.180-187
    • /
    • 2000
  • This paper presents the system of analyzing data in memory of TCMS. This system can show physical data in memory to text and graphic : format. To transfer data from TCMS to this system, a large number of system have used to memory IC card. the method of using memory IC card as a intermediation has many points at issue, that is a speed of transmitting data, life time of IC card and identity of train system each other. So this paper proposes that the method of wireless LAM be adopted by this system to improve the week point of previous method and other method to better the method of wireless LAN.

  • PDF

Flash Memory Shadow Paging Scheme Using Deferred Cleaning List for Portable Databases (휴대용 데이터베이스를 위한 지연된 소거 리스트를 이용하는 플래시 메모리 쉐도우 페이징 기법)

  • Byun Si-Woo
    • Journal of Information Technology Applications and Management
    • /
    • 제13권2호
    • /
    • pp.115-126
    • /
    • 2006
  • Recently, flash memories are one of best media to support portable computer's storages in mobile computing environment. We propose a new transaction recovery scheme for a flash memory database environment which is based on a flash media file system. We improved traditional shadow paging schemes by reusing old data pages which are supposed to be invalidated in the course of writing a new data page in the flash file system environment. In order to reuse these data pages, we exploit deferred cleaning list structure in our flash memory shadow paging (FMSP) scheme. FMSP scheme removes the additional storage overhead for keeping shadow pages and minimizes the I/O performance degradation caused by data page distribution phenomena of traditional shadow paging schemes. We also propose a simulation model to show the performance of FMSP. Based on the results of the performance evaluation, we conclude that FMSP outperforms the traditional scheme.

  • PDF

The Efficient Buffer Size in A Dual Flash Memory Structure with Buffer System (이중 NAND 플래시 구조의 버퍼시스템에서 효율적 버퍼 크기)

  • Jung, Bo-Sung;Lee, Jung-Hoon
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • 제6권6호
    • /
    • pp.383-391
    • /
    • 2011
  • As we know the effects of cache memory research, instruction and data caches can be separated for higher performance with Harvard CPUs. In this paper, we shows the efficiency of buffer system in the instruction and data flash storage medium. And we analyzed characteristics of the data and instruction flash and evaluated the performance. Finally, we propose the best buffer structure with an optimal block size and buffer size for the instruction and data flash.

Time-Aware Wear Leveling by Combining Garbage Collector and Static Wear Leveler for NAND Flash Memory System

  • Hwang, Sang-Ho;Kwak, Jong Wook
    • Journal of the Korea Society of Computer and Information
    • /
    • 제22권3호
    • /
    • pp.1-8
    • /
    • 2017
  • In this paper, we propose a new hybrid wear leveling technique for NAND Flash memory, called Time-Aware Wear Leveling (TAWL). Our proposal prolongs the lifetime of NAND Flash memory by using dynamic wear leveling technique which considers the wear level of hot blocks as well as static wear leveling technique which considers the wear level of the whole blocks. TAWL also reduces the overhead of garbage collection by separating hot data and cold data using update frequency rate. We showed that TAWL enhanced the lifetime of NAND flash memory up to 220% compared with previous wear leveling techniques and our technique also reduced the number of copy operations of garbage collections by separating hot and cold data up to 45%.

Simulation Study on the Stream Server for Deciding the Priority for Using Resources (스트림 서버에서 자원 사용 우선순위 결정을 위한 시뮬레이션 연구)

  • 박진원
    • Journal of the Korea Society for Simulation
    • /
    • 제12권4호
    • /
    • pp.95-102
    • /
    • 2003
  • Stream servers are for supplying multimedia stream data to users through the internet such as movies and music without discontinuation. A typical stream server is designed roughly by considering the characteristics of stream services and by employing processors, memory, PCI bus, Ethernet, TOE and disks. This study focuses on deciding the priority for using resources such as PCI bus, buffer memory and TOE buffer, which have limited capacities in a typical stream server. When the priorities for using limited resources are not given properly, the stream servers may not even function as originally designed. The simulation study shows that the top priority for using PCI bus for normal streaming services should be given to the operation that sends data from buffer memory to TOE buffer. Giving priority for using PCI bus to other operation such as sending data from disks to memory results in a deadlock phenomenon.

  • PDF

A Design and Implementation on Large Data File Management Using Buffer Cache and Virtual Memory File (버퍼 캐쉬와 가상메모리 파일을 이용한 대형 데이터화일의 처리방법 설계 및 구현)

  • 김병철;신병석;조동섭;황희영
    • The Transactions of the Korean Institute of Electrical Engineers
    • /
    • 제41권7호
    • /
    • pp.784-792
    • /
    • 1992
  • In this paper we design and implement a method for application programs to allow handling of large data files in DOS environment. In this method we use extended memory and hard disk as a data buffer. And we use a part of the conventional DOS memory as a buffer cache which allows the application program to use extended memory and hard disks transparently. Using buffer cache also allows us some speed improvement for the application program.

  • PDF

Simulation Study on the Stream Server for Deciding the Priority for Using Resources (스트림 서버에서 자원 사용 우선순위 결정을 위한 시뮬레이션 연구)

  • 박진원
    • Proceedings of the Korea Society for Simulation Conference
    • /
    • 한국시뮬레이션학회 2003년도 추계학술대회 및 정기총회
    • /
    • pp.67-74
    • /
    • 2003
  • Stream servers are for supplying multimedia stream data to users through the internet such as movies and musics without discontinuation. A typical stream server is designed roughly by considering the characteristics of stream services and by employing processors, memory, PCI bus, Ethernet, TOE and disks. This study focuses on deciding the priority for using resources such as PCI bus, buffer memory and TOE buffer, which have limited capacities in a typical stream server. The simulation study shows that the top priority for using PCI bus for normal streaming services should be given to the operation that sends data from buffer memory to TOE buffer Giving priority for using PCI bus to other operation such as sending data from disks to memory results in deadlock Phenomenon.

  • PDF

Locally weighted linear regression prefetching method for hybrid memory system (하이브리드 메모리 시스템의 지역 가중 선형회귀 프리페치 방법)

  • Tang, Qian;Kim, Jeong-Geun;Kim, Shin-Dug
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 한국정보처리학회 2020년도 추계학술발표대회
    • /
    • pp.12-15
    • /
    • 2020
  • Data access characteristics can directly affect the efficiency of the system execution. This research is to design an accurate predictor by using historical memory access information, where highly accessible data can be migrated from low-speed storage (SSD/HHD) to high-speed memory (Memory/CPU Cache) in advance, thereby reducing data access latency and further improving overall performance. For this goal, we design a locally weighted linear regression prefetch scheme to cope with irregular access patterns in large graph processing applications for a DARM-PCM hybrid memory structure. By analyzing the testing result, the appropriate structural parameters can be selected, which greatly improves the cache prefetching performance, resulting in overall performance improvement.