• Title/Summary/Keyword: Cluster Computing

Search Result 425, Processing Time 0.027 seconds

Visualization System for Natural Disaster Data (자연재난 데이터 실감 가시화 시스템)

  • Kim, Jongyong;Jeong, Seokcheol;Lee, Gyeweon;Cho, Joonyoung;Kim, Dongwook;Park, Sanghun
    • Journal of the Korea Computer Graphics Society
    • /
    • v.24 no.3
    • /
    • pp.21-31
    • /
    • 2018
  • We introduces a system that enables fast and effective visualization of natural disaster data such as typhoons, tsunamis, floods, and flooding to help make informed decisions in disaster situations. Data containing disaster information consists of a few hundred megabytes to many tens and hundreds of gigabytes, which can not be handled by a PC. This system was implemented in the form of a client-server based service to generate and output results from high-performance servers. The server in a built-in, high-performance cluster handles client requests and sends the result of visualization to the client. Clients can receive the results in any form of images, videos, or 3D graphic model by specifying a desired time frame, effectively viewing the results with a user-friendly GUI.

Development of a VR Juggler-based Virtual Reality Interface for Scientific Visualization Application (과학적 가시화 어플리케이션을 위한 VR Juggler 기반 가상현실 인터페이스 개발)

  • Gu, Gibeom;Hwang, Gyuhyun;Hur, YoungJu
    • KIISE Transactions on Computing Practices
    • /
    • v.22 no.10
    • /
    • pp.488-496
    • /
    • 2016
  • In this paper, we introduce a virtual reality interface for scientific visualization applications. Our VR interface is based on an open-source framework called VR Juggler. Although VR Juggler has its own advantages, it lacks some of the important functionalities needed for practical applications - event handling, synchronization and data sharing among cluster nodes, to name a few. We explain how these issues are resolved while developing the VR interface. Also, a new interface with a smart device, which replaces the virtual reality input device, is introduced. Finally, system usability test results are provided to prove the effectiveness of the proposed interfaces.

Rhipe Platform for Big Data Processing and Analysis (빅데이터 처리 및 분석을 위한 Rhipe 플랫폼)

  • Jung, Byung Ho;Shin, Ji Eun;Lim, Dong Hoon
    • The Korean Journal of Applied Statistics
    • /
    • v.27 no.7
    • /
    • pp.1171-1185
    • /
    • 2014
  • Rhipe that integrates R and Hadoop environment, made it possible to process and analyze massive amounts of data using a distributed processing environment. In this paper, we implemented multiple regression analysis using Rhipe with various data sizes of actual data and simulated data. Experimental results for comparing the computing speeds of pseudo-distributed and fully-distributed modes for configuring Hadoop cluster, showed fully-distributed mode was more fast than pseudo-distributed mode and computing speeds of fully-distributed mode were faster as the number of data nodes increases. We also compared the performance of our Rhipe with stats and biglm packages available on bigmemory. The results showed that our Rhipe was more fast than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.

Verification of the PMCEPT Monte Carlo dose Calculation Code for Simulations in Medical Physics (의학물리 분야에 사용하기 위한 PMCEPT 몬테카를로 도즈계산용 코드 검증)

  • Kum, O-Yeon
    • Progress in Medical Physics
    • /
    • v.19 no.1
    • /
    • pp.21-34
    • /
    • 2008
  • The parallel Monte Carlo electron and photon transport (PMCEPT) code [Kum and Lee, J. Korean Phys. Soc. 47, 716 (2006)] for calculating electron and photon beam doses has been developed based on the three dimensional geometry defined by computed tomography (CT) images and implemented on the Beowulf PC cluster. Understanding the limitations of Monte Carlo codes is useful in order to avoid systematic errors in simulations and to suggest further improvement of the codes. We evaluated the PMCEPT code by comparing its normalized depth doses for electron and photon beams with those of MCNP5, EGS4, DPM, and GEANT4 codes, and with measurements. The PMCEPT results agreed well with others in homogeneous and heterogeneous media within an error of $1{\sim}3%$ of the dose maximum. The computing time benchmark has also been performed for two cases, showing that the PMCEPT code was approximately twenty times faster than the MCNP5 for 20-MeV electron beams irradiated on the water phantom. For the 18-MV photon beams irradiated on the water phantom, the PMCEPT was three times faster than the GEANT4. Thus, the results suggest that the PMCEPT code is indeed appropriate for both fast and accurate simulations.

  • PDF

Multi-platform Visualization System for Earth Environment Data (지구환경 데이터를 위한 멀티플랫폼 가시화 시스템)

  • Jeong, Seokcheol;Jung, Seowon;Kim, Jongyong;Park, Sanghun
    • Journal of the Korea Computer Graphics Society
    • /
    • v.21 no.3
    • /
    • pp.36-45
    • /
    • 2015
  • It is important subject of research in engineering and natural science field that creating continuing high-definition image from very large volume data. The necessity of software that helps analyze useful information in data has improved by effectively showing visual image information of high resolution data with visualization technique. In this paper, we designed multi-platform visualization system based on client-server to analyze and express earth environment data effectively constructed with observation and prediction. The visualization server comprised of cluster transfers data to clients through parallel/distributed computing, and the client is developed to be operated in various platform and visualize data. In addition, we aim user-friendly program through multi-touch, sensor and have made realistic simulation image with image-based lighting technique.

Research on An Energy Efficient Triangular Shape Routing Protocol based on Clusters (클러스터에 기반한 에너지 효율적 삼각모양 라우팅 프로토콜에 관한 연구)

  • Nurhayati, Nurhayati;Lee, Kyung-Oh
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.9
    • /
    • pp.115-122
    • /
    • 2011
  • In this paper, we propose an efficient dynamic workload balancing strategy which improves the performance of high-performance computing system. The key idea of this dynamic workload balancing strategy is to minimize execution time of each job and to maximize the system throughput by effectively using system resource such as CPU, memory. Also, this strategy dynamically allocates job by considering demanded memory size of executing job and workload status of each node. If an overload node occurs due to allocated job, the proposed scheme migrates job, executing in overload nodes, to another free nodes and reduces the waiting time and execution time of job by balancing workload of each node. Through simulation, we show that the proposed dynamic workload balancing strategy based on CPU, memory improves the performance of high-performance computing system compared to previous strategies.

Clustering Algorithm for Extending Lifetime of Wireless Sensor Networks (무선 센서 네트워크의 수명연장을 위한 클러스터링 알고리즘)

  • Kim, Sun-Chol;Choi, Seung-Kwon;Cho, Yong-Hwan
    • Journal of the Korea Society of Computer and Information
    • /
    • v.20 no.4
    • /
    • pp.77-85
    • /
    • 2015
  • Recently, wireless sensor network(WSN) have been used in various fields to implement ubiquitous computing environment. WSN uses small, low cost and low power sensors in order to collect information from the sensor field. This paper proposes a clustering algorithm for energy efficiency of sensor nodes. The proposed algorithm is based on conventional LEACH, the representative clustering protocol for WSN and it prolongs network and nodes life time using sleep technique and changable transmission mode. The nodes of the proposed algorithm first calculate their clustering participation value based on the distance to the neighbor nodes. The nodes located in high density area will have clustering participation value and it can turn to sleep mode. Besides, proposed algorithm can change transmission method from conventional single-hop transmission to multi-hop transmission according to the energy level of cluster head. Simulation results show that the proposed clustering algorithm outperforms conventional LEACH, especially non-uniformly deployed network.

Workflow-based Bio Data Analysis System for HPC (HPC 환경을 위한 워크플로우 기반의 바이오 데이터 분석 시스템)

  • Ahn, Shinyoung;Kim, ByoungSeob;Choi, Hyun-Hwa;Jeon, Seunghyub;Bae, Seungjo;Choi, Wan
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.2
    • /
    • pp.97-106
    • /
    • 2013
  • Since human genome project finished, the cost for human genome analysis has decreased very rapidly. This results in the sharp increase of human genome data to be analyzed. As the need for fast analysis of very large bio data such as human genome increases, non IT researchers such as biologists should be able to execute fast and effectively many kinds of bio applications, which have a variety of characteristics, under HPC environment. To accomplish this purpose, a biologist need to define a sequence of bio applications as workflow easily because generally bio applications should be combined and executed in some order. This bio workflow should be executed in the form of distributed and parallel computing by allocating computing resources efficiently under HPC cluster system. Through this kind of job, we can expect better performance and fast response time of very large bio data analysis. This paper proposes a workflow-based data analysis system specialized for bio applications. Using this system, non-IT scientists and researchers can analyze very large bio data easily under HPC environment.

Efficient Processing of Multiple Group-by Queries in MapReduce for Big Data Analysis (맵리듀스에서 빅데이터 분석을 위한 다중 Group-by 질의의 효율적인 처리 기법)

  • Park, Eunju;Park, Sojeong;Oh, Sohyun;Choi, Hyejin;Lee, Ki Yong;Shim, Junho
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.5
    • /
    • pp.387-392
    • /
    • 2015
  • MapReduce is a framework used to process large data sets in parallel on a large cluster. A group-by query is a query that partitions the input data into groups based on the values of the specified attributes, and then evaluates the value of the specified aggregate function for each group. In this paper, we propose an efficient method for processing multiple group-by queries using MapReduce. Instead of computing each group-by query independently, the proposed method computes multiple group-by queries in stages with one or more MapReduce jobs in order to reduce the total execution cost. We compared the performance of this method with the performance of a less sophisticated method that computes each group-by query independently. This comparison showed that the proposed method offers better performance in terms of execution time.

Distributed data deduplication technique using similarity based clustering and multi-layer bloom filter (SDS 환경의 유사도 기반 클러스터링 및 다중 계층 블룸필터를 활용한 분산 중복제거 기법)

  • Yoon, Dabin;Kim, Deok-Hwan
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.14 no.5
    • /
    • pp.60-70
    • /
    • 2018
  • A software defined storage (SDS) is being deployed in cloud environment to allow multiple users to virtualize physical servers, but a solution for optimizing space efficiency with limited physical resources is needed. In the conventional data deduplication system, it is difficult to deduplicate redundant data uploaded to distributed storages. In this paper, we propose a distributed deduplication method using similarity-based clustering and multi-layer bloom filter. Rabin hash is applied to determine the degree of similarity between virtual machine servers and cluster similar virtual machines. Therefore, it improves the performance compared to deduplication efficiency for individual storage nodes. In addition, a multi-layer bloom filter incorporated into the deduplication process to shorten processing time by reducing the number of the false positives. Experimental results show that the proposed method improves the deduplication ratio by 9% compared to deduplication method using IP address based clusters without any difference in processing time.