• Title/Summary/Keyword: Memory allocation

Search Result 204, Processing Time 0.033 seconds

The Design and Implementation of Memory Allocation using Max Heap Algorithm on Real-time Operating System (실시간 운영체제에서 최대 힙 알고리즘을 이용한 메모리 할당 기법 설계 및 구현)

  • 이정원;최인범;김용희;이철훈
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2004.10a
    • /
    • pp.634-636
    • /
    • 2004
  • 실시간 운영체제는 멀티태스킹 및 ITC(Inter Task Communication)를 제공한다는 면에서는 범용운영체제와 비슷하나, 시간 결정성을 보장해야 한다는 면에서는 일반 운영체제와 다르다. 실시간 시스템에서는 메모리를 할당하는데 있어서 시간 제약을 어기지 않아야 하기 때문에 동적 메모리 할당은 효율적으로 구성되어야 한다. 본 논문에서는 실시간 운영체제 $_{1}$RTOS$^{TM}$에서 메모리 할당에 소요되는 시간을 향상시키기 위해 최대 힙 알고리즘을 적용한 메모리 할당 기법을 설계 및 구현하였다.

  • PDF

The Verification Using CBMC about Triple DES (CBMC를 이용한 트리플 DES의 검증)

  • 강미영;유희준;최진영
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2004.10a
    • /
    • pp.754-756
    • /
    • 2004
  • CBMC는 ANSI-C 프로그램과 베릴로그 서킷사이의 일치성을 검증하는 툴이다. 입력된 서킷과 코드를 CNF로 변환 과정에서 C 코드는 중첩 루프, pointer, dynamic memory allocation등에 대한 변환의 문제점이 있다. 본 논문에서는 CBMC에서 C 코드의 CNF로 변환하는 과정의 문제점들을 동일한 식(equation)의 변환 과정에 대하여 설명하고 상용적인 트리플 DES를 CBMC로 검증하는 과정을 제시한다

  • PDF

Development Visual Tools of that shows memory allocation of JVM (자바가상기계 메모리 할당을 보여주는 비주얼 도구의 개발)

  • 이수진;정민수;김우완;이공선;윤기송
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10a
    • /
    • pp.409-411
    • /
    • 1999
  • 본 논문에서 소개하고 있는 비주얼 자바가상기계 시뮬레이터는 자바 컴파일러에 의해 컴파일된 결과인 바이트코드를 분석하고, 그 분석된 결과가 자바가상기계내부구조에 어떻게 할당되는지를 시각적으로 보여줌으로써 자바 소스 코드가 보여주지 못하는 가상기계 내부의 메소드 영역, 자바 스택 영역, 힙 영역에 할당되는 정보를 통해 자바 소스 프로그램의 보다 명확하고 쉬운 이해가 가능하도록 한다.

  • PDF

A Study on Effect of Code Distribution and Data Replication for Multicore Computing Architectures

  • Cho, Doosan
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.282-287
    • /
    • 2021
  • A multicore system must be able to take full advantage of the program's instruction and data parallelism. This study introduces the data replication technique as a support technique to maximize the program's instruction and data parallelism. Instruction level parallelism can be limited by data dependency. In this case, if data is replicated to each processor core and used, instruction level parallelism can be used to the maximum. The technique proposed in this study can maximize the performance improvement effect when applied to scientific applications such as matrix multiplication operation.

High Throughput Parallel KMP Algorithm Considering CPU-GPU Memory Hierarchy (CPU-GPU 메모리 계층을 고려한 고처리율 병렬 KMP 알고리즘)

  • Park, Soeun;Kim, Daehee;Lee, Myungho;Park, Neungsoo
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.67 no.5
    • /
    • pp.656-662
    • /
    • 2018
  • Pattern matching algorithm is widely used in many application fields such as bio-informatics, intrusion detection, etc. Among many string matching algorithms, KMP (Knuth-Morris-Pratt) algorithm is commonly used because of its fast execution time when using large texts. However, the processing speed of KMP algorithm is also limited when the text size increases significantly. In this paper, we propose a high throughput parallel KMP algorithm considering CPU-GPU memory hierarchy based on OpenCL in GPGPU (General Purpose computing on Graphic Processing Unit). We focus on the optimization for the allocation of work-times and work-groups, the local memory copy of the pattern data and the failure table, and the overlapping of the data transfer with the string matching operations. The experimental results show that the execution time of the optimized parallel KMP algorithm is about 3.6 times faster than that of the non-optimized parallel KMP algorithm.

Analyzing Fine-Grained Resource Utilization for Efficient GPU Workload Allocation (GPU 작업 배치의 효율화를 위한 자원 이용률 상세 분석)

  • Park, Yunjoo;Shin, Donghee;Cho, Kyungwoon;Bahn, Hyokyung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.1
    • /
    • pp.111-116
    • /
    • 2019
  • Recently, GPU expands application domains from graphic processing to various kinds of parallel workloads. However, current GPU systems focus on the maximization of each workload's parallelism through simplified control rather than considering various workload characteristics. This paper classifies the resource usage characteristics of GPU workloads into computing-bound, memory-bound, and dependency-latency-bound, and quantifies the fine-grained bottleneck for efficient workload allocation. For example, we identify the exact bottleneck resources such as single function unit, double function unit, or special function unit even for the same computing-bound workloads. Our analysis implies that workloads can be allocated together if fine-grained bottleneck resources are different even for the same computing-bound workloads, which can eventually contribute to efficient workload allocation in GPU.

Technique to Reduce Container Restart for Improving Execution Time of Container Workflow in Kubernetes Environments (쿠버네티스 환경에서 컨테이너 워크플로의 실행 시간 개선을 위한 컨테이너 재시작 감소 기법)

  • Taeshin Kang;Heonchang Yu
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.3
    • /
    • pp.91-101
    • /
    • 2024
  • The utilization of container virtualization technology ensures the consistency and portability of data-intensive and memory volatile workflows. Kubernetes serves as the de facto standard for orchestrating these container applications. Cloud users often overprovision container applications to avoid container restarts caused by resource shortages. However, overprovisioning results in decreased CPU and memory resource utilization. To address this issue, oversubscription of container resources is commonly employed, although excessive oversubscription of memory resources can lead to a cascade of container restarts due to node memory scarcity. Container restarts can reset operations and impose substantial overhead on containers with high memory volatility that include numerous stateful applications. This paper proposes a technique to mitigate container restarts in a memory oversubscription environment based on Kubernetes. The proposed technique involves identifying containers that are likely to request memory allocation on nodes experiencing high memory usage and temporarily pausing these containers. By significantly reducing the CPU usage of containers, an effect similar to a paused state is achieved. The suspension of the identified containers is released once it is determined that the corresponding node's memory usage has been reduced. The average number of container restarts was reduced by an average of 40% and a maximum of 58% when executing a high memory volatile workflow in a Kubernetes environment with the proposed method compared to its absence. Furthermore, the total execution time of a container workflow is decreased by an average of 7% and a maximum of 13% due to the reduced frequency of container restarts.

Minimizing Security Hole and Improving Performance in Stateful Inspection for TCP Connections (TCP연결의 스테이트풀 인스펙션에 있어서의 보안 약점 최소화 및 성능 향상 방법)

  • Kim, Hyo-Gon;Kang, In-Hye
    • Journal of KIISE:Information Networking
    • /
    • v.32 no.4
    • /
    • pp.443-451
    • /
    • 2005
  • Stateful inspection devices must maintain flow information. These devices create the flow information also for network attack packets, and it can fatally inflate the dynamic memory allocation on stateful inspection devices under network attacks. The memory inflation leads to memory overflow and subsequent performance degradation. In this paper, we present a guideline to set the flow entry timeout for a stateful inspection device to remove harmful embryonic entries created by network attacks. Considering Transmission Control Protocol (TCP) if utilized by most of these attacks as well as legitimate traffic, we propose a parsimonious memory management guideline based on the design of the TCP and the analysis of real-life Internet traces. In particular, we demonstrate that for all practical purposes one should not reserve memory for an embryonic TCP connection with more than (R+T) seconds of inactivity where R=0, 3, 9 and $1\leqq{T}\leqq{2}$ depending on the load level.

Dynamic Limited Directory Scheme for Distributed Shared Memory Systems (분산공유 메모리 시스템을 위한 동적 제한 디렉터리 기법)

  • Lee, Dong-Gwang;Gwon, Hyeok-Seong;Choe, Seong-Min;An, Byeong-Cheol
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.4
    • /
    • pp.1098-1105
    • /
    • 1999
  • The caches in distributed shared memory systems enhance the performance by reducing memory access latency and communication overhead, but they must solve the cache coherence problem. This paper proposes a new directory protocol to solve the cache coherence problem and to improve the system performance in distributed shared memory systems. To maintain the cache coherence of shared data, processors within a limited distance reduce the communication overhead by using a bit-vector like the full directory scheme. Processors over a limited distance store pointers in a directory pool. Since the bit-vector and the directory pool remove the unnecessary cache invalidations, the proposed scheme reduces the communication traffic and improves the system performance. The dynamic limited directory scheme reduces the communication traffic up to 66 percents compared with the limited directory scheme and the number of directory access up to 27 percents compared with the dynamic pointer allocation scheme.

  • PDF

Meta-cognitive awareness and prospective memory retrieval process: The effect of voluntary attention allocation strategy (메타 인지적 인식과 미래계획기억 인출 과정: 자발적 주의 할당 전략의 효과)

  • Yoon, Yong-Sik;Sohn, Young-Woo
    • Korean Journal of Cognitive Science
    • /
    • v.22 no.2
    • /
    • pp.145-172
    • /
    • 2011
  • Prospective memory (PM) is memory for activities to be performed in the future, such as remembering to purchase a piece of fruit on the way home or remembering to give someone a telephone message. Theoretically, PM retrieval can be accomplished either by controlled monitoring of intended action or by a more automatic process that spontaneously responds to the presence of intended event. The present study examined the effect of meta-cognitive awareness in PM retrieval process. In Experiment 1, we manipulated PM condition with a salient cue to stimulate participants' meta-cognitive awareness. Results showed the salient cue condition spontaneously triggered remembering to perform an intended action. In Experiment 2, we manipulated participants' meta-cognitive awareness about PM task difficulty and about the degree to which cognitive capacity is required to realize the PM task. Results showed that PM retrieval process varied with meta-cognitive awareness (i.e., more resources were allocated if the PM task was expected to be difficult). These results imply that participants' meta-cognitive awareness play an important role in PM retrieval process.

  • PDF