• Title/Summary/Keyword: load aware

Search Result 94, Processing Time 0.019 seconds

A Study on Context Aware Vertical Handover Scheme for Supporting Optimized Flow Multi-Wireless Channel Service based Heterogeneous Networks (이기종 망간의 최적화된 플로우 기반 다중 무선 채널 지원을 위한 상황인지 수직핸드오버 네트워크 연구)

  • Shin, Seungyong;Park, Byungjoo
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.2
    • /
    • pp.1-7
    • /
    • 2019
  • Recently, multimedia streaming service has been activated, and the demand for high-quality multimedia convergence contents services is predicted to increase significantly in the future. The issues of the increasing network load due to the rise of multimedia streaming traffic must be addressed in order to provide QoS guaranteed services. To do this, an efficient network resource management and mobility support technologies are needed through seamless mobility support for heterogeneous networks. Therefore, in this paper, an MIH technology was used to recognize the network situation information in advance and reduce packet loss due to handover delays, and an ACLMIH-FHPMIPv6 is designed that can provide an intelligent interface through introducing a hierarchical mobility management technique in FPMIPv6 integrated network.

Implementation of Multicore-Aware Load Balancing on Clusters through Data Distribution in Chapel (클러스터 상에서 다중 코어 인지 부하 균등화를 위한 Chapel 데이터 분산 구현)

  • Gu, Bon-Gen;Carpenter, Patrick;Yu, Weikuan
    • The KIPS Transactions:PartA
    • /
    • v.19A no.3
    • /
    • pp.129-138
    • /
    • 2012
  • In distributed memory architectures like clusters, each node stores a portion of data. How data is distributed across nodes influences the performance of such systems. The data distribution scheme is the strategy to distribute data across nodes and realize parallel data processing. Due to various reasons such as maintenance, scale up, upgrade, etc., the performance of nodes in a cluster can often become non-identical. In such clusters, data distribution without considering performance cannot efficiently distribute data on nodes. In this paper, we propose a new data distribution scheme based on the number of cores in nodes. We use the number of cores as the performance factor. In our data distribution scheme, each node is allocated an amount of data proportional to the number of cores in it. We implement our data distribution scheme using the Chapel language. To show our data distribution is effective in reducing the execution time of parallel applications, we implement Mandelbrot Set and ${\pi}$-Calculation programs with our data distribution scheme, and compare the execution times on a cluster. Based on experimental results on clusters of 8-core and 16-core nodes, we demonstrate that data distribution based on the number of cores can contribute to a reduction in the execution times of parallel programs on clusters.

Energy-Efficient Routing Protocol based on Interference Awareness for Transmission of Delay-Sensitive Data in Multi-Hop RF Energy Harvesting Networks (다중 홉 RF 에너지 하베스팅 네트워크에서 지연에 민감한 데이터 전송을 위한 간섭 인지 기반 에너지 효율적인 라우팅 프로토콜)

  • Kim, Hyun-Tae;Ra, In-Ho
    • The Journal of the Korea Contents Association
    • /
    • v.18 no.3
    • /
    • pp.611-625
    • /
    • 2018
  • With innovative advances in wireless communication technology, many researches for extending network lifetime in maximum by using energy harvesting have been actively performed on the area of network resource optimization, QoS-guaranteed transmission, energy-intelligent routing and etc. As known well, it is very hard to guarantee end-to-end network delay due to uncertainty of the amount of harvested energy in multi-hop RF(radio frequency) energy harvesting wireless networks. To minimize end-to-end delay in multi-hop RF energy harvesting networks, this paper proposes an energy efficient routing metric based on interference aware and protocol which takes account of various delays caused by co-channel interference, energy harvesting time and queuing in a relay node. The proposed method maximizes end-to-end throughput by performing avoidance of packet congestion causing load unbalance, reduction of waiting time due to exhaustion of energy and restraint of delay time from co-channel interference. Finally simulation results using ns-3 simulator show that the proposed method outperforms existing methods in respect of throughput, end-to-end delay and energy consumption.

Design and Implementation of a Fault-Tolerant Caching System for Dynamic Heterogeneous Cache Server Networks (동적 이기종 캐시 서버 네트워크에서의 내결함성 캐싱 시스템 설계 및 구현)

  • Hyeon-Gi Kim;Gyu-Sik Ham;Jin-Woo Kim;Soo-Young Jang;Chang-Beom Choi
    • Journal of IKEEE
    • /
    • v.28 no.3
    • /
    • pp.458-464
    • /
    • 2024
  • This study proposes a fault-tolerant caching system to address the issue of caching content imbalance caused by the dynamic departure and participation of cache servers in a heterogeneous cache server network, and validates it in both real and virtual environments. With the increase of large-scale media content requiring various types and resolutions, the necessity of cache servers as key components to reduce response time to user requests and alleviate network load has been growing. In particular, research on heterogeneous cache server networks utilizing edge computing and low-power devices has been actively conducted recently. However, in such environments, the irregular departure and participation of cache servers can occur frequently, leading to content imbalance among the cache servers deployed in the network, which can degrade the performance of the cache server network. The fault-tolerant caching algorithm proposed in this study ensures stable service quality by maintaining balance among media contents even when cache servers depart. Experimental results confirmed that the proposed algorithm effectively maintains content distribution despite the departure of cache servers. Additionally, we built a network composed of seven heterogeneous cache servers to verify the practicality of the proposed caching system and demonstrated its performance and scalability through a large-scale cache server network in a virtual environment.