• Title/Summary/Keyword: server performance

Search Result 1,690, Processing Time 0.033 seconds

Back-end Prefetching Scheme for Improving the Performance of Cluster-based Web Servers (클러스터 웹 서버에서 성능 향상을 위한 노드간 선인출 기법)

  • Park, Seon-Yeong;Park, Do-Hyeon;Lee, Joon-Won;Cho, Jung-Wan
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.5
    • /
    • pp.265-273
    • /
    • 2002
  • With the explosive growth of WWW traffic, there is an increasing demand for the high performance Web servers to provide a stable Web service to users. The cluster-based Web server is a solution to core with the heavy access from users, easily scaling the server according to the loads. In the cluster-based Web sewer, a back-end node may not be able to serve some HTTP requests directly because it does not have the requested contents in its main memory. In this case, the back-end node has to retrieve the requested contents from its local disk or other back-end nodes in the cluster. To reduce service latency, we introduce a new prefetch scheme. The back-end nodes predict the next HTTP requests and prefetch the contents of predicted requests before the next requests arrive. We develop three prefetch algorithms bated on some useful information gathered from many clients'HTTP requests. Through trace-driven simulation, the service latency of the prefetch scheme is reduced by 10 ~ 25% as compared with no prefetch scheme. Among the proposed prefetch algorithms, Time and Access Probability-based Prefetch (TAP2) algorithm, which uses the access probability and the inter-reference time of Web object, shows the best performance.

Development of a Kernel Thread Web Accelerator (SCALA-AX) (커널 쓰레드 웹가속기(SCALA-AX) 개발)

  • Park, Jong-Gyu;Min, Byung-Jo;Lim, Han-Na;Park, Jang-Hoon;Chang, Whi;Kim, Hag-Bae
    • The KIPS Transactions:PartA
    • /
    • v.9A no.3
    • /
    • pp.327-332
    • /
    • 2002
  • Conventional proxy web cache, which is generally used to caching server, is a content-copy based system. This method focuses on speeding up the phase delivery not improving the webserver performance. However, if immense clients attempt to connect the webserver simultaneously, the proxy web cache cannot achieve the desired result. In this paper, we propose the web accelerator called the SCALA-AX, whitch improves web server performance by accelerating the delivery contents. The SCALA-AX is built in the Linux-based kernel as a kernel modulo and works in combination with the conventional webserver program. The SCALA-AX speeds up the processing rate of the webserver, because it processes the requests using the kernel thread. The SCALA-AX also applies the well-developed cache algorithm to the processing, and thus it obtains the advantage of the caching server without installing additional hardware. A banchmarking test demonstrates that the SCALA-AX improves webserver performance by up to 500% for content delivery.

Load balancing of a deployment server using P2P (P2P를 이용한 배포 서버의 부하 분산)

  • Son Sei-Il;Lee Suk-Kyoon
    • The KIPS Transactions:PartA
    • /
    • v.13A no.1 s.98
    • /
    • pp.45-52
    • /
    • 2006
  • To perform on-line maintenance for Distributed Information System, it is indispensable to disseminate files to participant nodes in the network. When users' requests for file deployment occur simultaneously in a short period a deployment server falls into overload phase, which is often called Flash Crowds. h common solution to avoid Flash Crowds is to increase hardware capacity. In this paper, we propose a software solution based on P2P, which does not cost any additional expense. In the proposed solution, nodes in the network are grouped into subnetworks one of which is composed of only neighboring nodes. In each subnetwork, copies of deployment files can be transferred to each other. Consequently, it brings about the effect of load balancing in deployment server. To raise the effectiveness, target files for deployment are packed into one package. Before being transferred, each package is divided into multiple equal-sized segments. A deployment server in a normal phase transmits a package requested from nodes in segment units. However a deployment server is overloaded, if segments already exist in the subnetwork participant nodes in the subnetwork receive necessary segments from neighboring nodes. In this paper, we propose data structures and algorithm for this approach and show performance improvement through simulation.

Optimistic Concurrency Control based on 2-Version and TimeStamp for Broadcast Environment : OCC/2VTS (방송환경에서 이중 버전과 타임스탬프에 기반을 둔 낙관적 동시성 제어 기법)

  • Lee, Uk-Hyun;Hwang, Bu-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.8D no.2
    • /
    • pp.132-144
    • /
    • 2001
  • The broadcast environment is asymmetric communication aspect that is typically much greater communication capacity available from server to clients than in the opposite direction. In addition, most of mobile computing systems only allow the generation of read-only transactions from mobile clients for retrieving different types of information such as stock data, traffic information and news updates. Since previous concurrency control protocols, however, do not consider such a particular characteristics, the performance degradation occurs when those schemes are applied to the broadcast environment having quite a high data contention. In this paper, we propose OCC/2VTS (Optimistic Concurrency Control based on 2-Version and TimeStamp) that is most appropriate for broadcast environment. OCC/2VTS lets each client process and commit query transactions for itself by using two version data in cache. If the values of appropriate data items are not changed twice by invalidation report after a query transaction starts, the query transaction is committed safely independent of commitment of update transactions. OCC/2VTS decreases the number of informing server for the purpose of commitment. Due to broadcasting the validation reports including updated recent values, it reduces the opportunity of requesting a recent data values of server as well. As a result, OCC/2VTS makes full use of the asymmetric bandwidth. It also improves transaction throughput by increasing the query transaction commit ratio as much as possible.

  • PDF

A Shared Cache Directory based Wireless Internet Proxy Server Cluster (공유 캐시 디렉토리 기반의 무선 인터넷 프록시 서버 클러스터)

  • Kwak Hu-Keun;Chung Kyu-Sik
    • The KIPS Transactions:PartA
    • /
    • v.13A no.4 s.101
    • /
    • pp.343-350
    • /
    • 2006
  • In this paper, wireless internet proxy server clusters are used for the wireless internet because their caching, distillation, and clustering functions are helpful to overcome the limitations and needs of the wireless internet. A wireless Internet proxy server cluster needs a systematic scalability, simple communication structure, cooperative caching, and serving Hot Spot requests. In our former research, we proposed the CD-A structure which can be scalable in a systematic way and has a simple communication structure but it has no cooperative caching. A hash based load balancing can be used to solve the problem, but it can not deal with Hot Spot request problem. In this paper, we proposed a shared storage based wireless internet proxy server cluster which has a systematic scalability, simple communication structure, cooperative caching, and serving Hot Spot requests. The proposed method shares one cache directory and it has advantages: advantages of the existing CD-A structure, cooperative caching, and serving Hot Spot requests. We performed experiments using 16 PCs and experimental results show high performance improvement of the proposed system compared to the existing systems in Hot Spot requests.

A Performance Evaluation for IPoIB Protocol in Channel based Network (채널기반형 네트웍에서의 IPoIB 프로토콜 성능평가)

  • Jeon, Ki-Man;Min, Soo-Young;Kim, Young-Wan
    • Proceedings of the KIEE Conference
    • /
    • 2004.11c
    • /
    • pp.687-689
    • /
    • 2004
  • As using of network increases rapidly, performance of system has been deteriorating because of the overhead and bottleneck. Nowadays, High speed I/O network standard, that is a sort of PCI Express, HyperTransport, InfiniBand, and so on, has come out to improve the limites of traditional I/O bus. The InfiniBand Architecture(IBA) provides some protocols to service the applications such as SDP, SRP and IPoIB. In our paper, We explain the architecture of IPoIB (IP over InfiniBand) and its features in channel based I/O network. And so we provide a performance evaluation result of IPoIB which is compared with current network protocol. Our experimental results also show that IPoIB is batter than TCP/IP protocol. For this test, We use the dual processor server systems and Linux Redhat 9.0 operating system.

  • PDF

Simulation Analysis for Configuring Web Clusters (웹 클러스터 구성을 위한 시뮬레이션 분석)

  • Kang, Sung-Yeol;Song, Young-Hyo
    • Journal of Digital Convergence
    • /
    • v.6 no.2
    • /
    • pp.117-125
    • /
    • 2008
  • High-volume web sites often use clusters of servers with load balancing as a way to increase the performance, scalability, and availability of the sites. Load balancing, usually performed by load balancer in front of such clusters, is a technique to spread workload between several computers or resources, in order to get optimal resource utilization or response time. In this paper we examine the performance for several configurations of cluster-based web servers using a simulation approach. We investigate two types of buffering scheme (common and local) for web clusters and three load balancing policies (uniformly random, round robin, and least queue first), using response time as a performance measure. We also examine two basic approaches of scaling web clusters: adding more servers of same type or upgrading the capacity of the servers in the clusters.

  • PDF

Simulation Study for Performance Measures of Resources in a Port Container Terminal

  • Choi, Yong-Seok
    • Journal of Navigation and Port Research
    • /
    • v.28 no.7
    • /
    • pp.587-591
    • /
    • 2004
  • In order to measure the performance of resources in a port container terminal, we conducted a state transition network simulation to model various equipment processes. The processes that container cranes and transfer cranes perform are idle, wait, move, and work. Vehicles perform idle, wait, empty travel, and full travel. Because cranes, vehicles, and vessels are movable entities and all equipment is classified as either a customer or server, we separated the various stages of the process based upon the state transition network To validate the simulation results, a real system was used to illustrate the use of various measurements using the state transition network.

M/G/1 Queueing Model for the Performance Estimation of AS/RS (자동창고시스템의 성능평가를 위한 M/G/1 대기모형)

  • Lee, Mun-Hwan;Lim, Si-Yeong;Heo, Seon;Lee, Yeong-Hae
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2000.04a
    • /
    • pp.253-256
    • /
    • 2000
  • In general, Automated Storage/Retrieval Systems (AS/RS ) have racks of equal sized cells to utilize the concept of unit-load. Most of the techniques for the performance estimation of a unit-load AS/RS are a static model or computer simulation. Especially, their models were developed under assumption that the Storage/Retrieval (S/R) machine performs only single command (SC) or dual command (DC). In reality, defending on the operating policy and the status of the system at a particular time, the S/B machine performs a SC or a DC, or becomes .: idle. In order to resolve these weak points, we propose a stochastic model for the performance estimation of unit-load AS/RS by using a single-server queueing model. Expected numbers of waiting storage and retrieval commands are found

  • PDF

Performance Analysis of NTT/BT Protocol (NTT/BT 프로토콜의 성능 분석)

  • 이창훈;백상엽;이동주
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.22 no.2
    • /
    • pp.99-123
    • /
    • 1997
  • Performance analysis of NTT/BT protocol is investigated, which is a GFC (Generic Flow Control) ptotocol in ATM (Asynchronous Transfer Mode ) network and is based on cyclic reset mechanism. THe mean cell delay time is proposed as a performance measure of NTT/BT protocol. The mean cell delay time is defined as the duration from the instant the cell arrives at the transmission buffer until the cell is fully transmitted. The process of cell transmission can be described as a single server queueing modle with two dependent services. By utilizing this model, mean cell delay time is obtained and sensitivity of the factors such as window size and reset period is also analysed.

  • PDF