• Title/Summary/Keyword: 분산 클라우드

Search Result 269, Processing Time 0.027 seconds

Design and performance evaluation of a storage cloud service model over KREONET (KREONET 기반의 스토리지 클라우드 서비스 모델 설계 및 성능평가)

  • Hong, Wontaek;Chung, Jinwook
    • Journal of the Korea Convergence Society
    • /
    • v.8 no.7
    • /
    • pp.29-37
    • /
    • 2017
  • Compared to the commercial networks, R&E networks have the strength such as flexible network engineering and design. Based on those features of R&E networks, we propose our storage cloud service model which supports general-purpose network users in a central region and experimental network users in distributed regions simultaneously. We prototype our service model utilizing multiple proxy controllers of OpenStack Swift service in order to deploy several regions via experimental backbone networks. Our experiments on the influence of the network latency and the size of data to be transmitted show that the bigger size of data is preferable to the smaller size of data in an experimental backbone network where the network latency increases within 10ms because the rate of throughput decline in the bigger object is comparatively small. It means that our service model is appropriate for experimental network users who directly access the service in order to move intermittently high volume of data as well as normal users in the central region who access the service frequently.

A Virtual Machine Allocation Scheme based on CPU Utilization in Cloud Computing (클라우드 컴퓨팅에서 CPU 사용률을 고려한 가상머신 할당 기법)

  • Bae, Jun-Sung;Lee, Bong-Hwan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.3
    • /
    • pp.567-575
    • /
    • 2011
  • The two most popular virtual machine allocation schemes, both match making and round robin, do consider hardware specifications such as CPU, RAM, and HDD, but not CPU usage, which results in balanced resource distribution, but not in balanced resource usage. Thus, in this paper a new virtual machine allocation scheme considering current CPU usage rate is proposed while retaining even distribution of node resources. In order to evaluate the performance of the proposed scheme, a cloud computing platform composed of three cloud nodes and one front end is implemented. The proposed allocation scheme was compared with both match making and round robin schemes. Experimental results show that the proposed scheme performs better in even distribution of overall CPU usage, which results in efficient load balancing.

The CloudHIS System for Personal Healthcare Information Integration Scheme of Cloud Computing (클라우드 컴퓨팅 환경에서 개인의료정보를 통합한 CloudHIS 시스템)

  • Cho, Young-Bok;Woo, Sung-Hee;Lee, Sang-Ho
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.5
    • /
    • pp.27-35
    • /
    • 2014
  • The characteristics of today's health care industry, based on the state of the art IT can be represented as a paradigm of human-oriented ubiquitous and accessible as possible by U-Health care. In addition, the healthcare industry is information and communication technologies (ICT) developments regarding the many advances and applications based on the research being carried out actively. Medical information system has been developed toward combining information systems of medical IT and it sets its sights on the fusion of developed IT and u-healthcare system. So changing distributed medical information systems into a safe PHR integrated system based on IaaS cloud computing is suggested in order to forge u-healthcare system with the times in this paper. Our experimental results show that our proposed system increased the data access time by about 24% and reduces the waiting time for processing service by about 4.3% over the web-based PHR.

Hierarchical Visualization of Cloud-Based Social Network Service Using Fuzzy (퍼지를 이용한 클라우드 기반의 소셜 네트워크 서비스 계층적 시각화)

  • Park, Sun;Kim, Yong-Il;Lee, Seong Ro
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.38B no.7
    • /
    • pp.501-511
    • /
    • 2013
  • Recently, the visualization method of social network service have been only focusing on presentation of visualizing network data, which the methods do not consider an efficient processing speed and computational complexity for increasing at the ratio of arithmetical of a big data regarding social networks. This paper proposes a cloud based on visualization method to visualize a user focused hierarchy relationship between user's nodes on social network. The proposed method can intuitionally understand the user's social relationship since the method uses fuzzy to represent a hierarchical relationship of user nodes of social network. It also can easily identify a key role relationship of users on social network. In addition, the method uses hadoop and hive based on cloud for distributed parallel processing of visualization algorithm, which it can expedite the big data of social network.

Tracking Data through Tracking Data Server in Edge Computing (엣지 컴퓨팅 환경에서 추적 데이터 서버를 통한 데이터 추적)

  • Lim, Han-wool;Byoun, Won-jun;Yun, Joobeom
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.31 no.3
    • /
    • pp.443-452
    • /
    • 2021
  • One of the key technologies in edge computing is that it always provides services close to the user by moving data between edge servers according to the user's movements. As such, the movement of data between edge servers is frequent. As IoT technology advances and usage areas expand, the data generated also increases, requiring technology to accurately track and process each data to properly manage the data present in the edge computing environment. Currently, cloud systems do not have data disposal technology based on tracking technology for data movement and distribution in their environment, so users cannot see where it is now, whether it is properly removed or not left in the cloud system if users request it to be deleted. In this paper, we propose a tracking data server to create and manage the movement and distribution of data for each edge server and data stored in the central cloud in an edge computing environment.

Log Management System of Web Server Based on Blockchain in Cloud Environment (클라우드 환경에서 블록체인 기반의 웹서버 로그 관리 시스템)

  • Son, Yong-Bum;Kim, Young-Hak
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.7
    • /
    • pp.143-148
    • /
    • 2020
  • Recently, web services have been expanded to various areas with the evolution of cloud environment. Whenever a user accesses a web service, the user's log information is stored in the web server. This log information is used as data to analyze the user's web service tendencies and is also used as important data to track the user's system access when a security problem in the system occurs. Currently, most web servers manage user log information in a centralized manner. When user log information is managed in a centralized manner, it is simple in the side of operation, but has a disadvantage of being very vulnerable to external malicious attacks. In the case of centralized management, user log information stored in the web server can be arbitrarily manipulated by external attacks, and in severe cases, the manipulated information can be leaked. In this case, it not only decreases the trust of the web service, but also makes it difficult to trace the source and cause of the attack on the web server. In order to solve these problems, this paper proposes a new method of managing user log information in a cloud environment by applying blockchain technology as an alternative to the existing centralized log management method. The proposed method can manage log information safely from external attacks because user log information is distributed and stored in blockchain on a private network with cloud environment.

Implementation of a DB-Based Virtual File System for Lightweight IoT Clouds (경량 사물 인터넷 클라우드를 위한 DB 기반 가상 파일 시스템 구현)

  • Lee, Hyung-Bong;Kwon, Ki-Hyeon
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.3 no.10
    • /
    • pp.311-322
    • /
    • 2014
  • IoT(Internet of Things) is a concept of connected internet pursuing direct access to devices or sensors in fused environment of personal, industrial and public area. In IoT environment, it is possible to access realtime data, and the data format and topology of devices are diverse. Also, there are bidirectional communications between users and devices to control actuators in IoT. In this point, IoT is different from the conventional internet in which data are produced by human desktops and gathered in server systems by way of one-sided simple internet communications. For the cloud or portal service of IoT, there should be a file management framework supporting systematic naming service and unified data access interface encompassing the variety of IoT things. This paper implements a DB-based virtual file system maintaining attributes of IoT things in a UNIX-styled file system view. Users who logged in the virtual shell are able to explore IoT things by navigating the virtual file system, and able to access IoT things directly via UNIX-styled file I O APIs. The implemented virtual file system is lightweight and flexible because it maintains only directory structure and descriptors for the distributed IoT things. The result of a test for the virtual shell primitives such as mkdir() or chdir() shows the smooth functionality of the virtual file system, Also, the exploring performance of the file system is better than that of Window file system in case of adopting a simple directory cache mechanism.

A Study on Security Improvement in Hadoop Distributed File System Based on Kerberos (Kerberos 기반 하둡 분산 파일 시스템의 안전성 향상방안)

  • Park, So Hyeon;Jeong, Ik Rae
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.23 no.5
    • /
    • pp.803-813
    • /
    • 2013
  • As the developments of smart devices and social network services, the amount of data has been exploding. The world is facing Big data era. For these reasons, the Big data processing technology which is a new technology that can handle such data has attracted much attention. One of the most representative technologies is Hadoop. Hadoop Distributed File System(HDFS) designed to run on commercial Linux server is an open source framework and can store many terabytes of data. The initial version of Hadoop did not consider security because it only focused on efficient Big data processing. As the number of users rapidly increases, a lot of sensitive data including personal information were stored on HDFS. So Hadoop announced a new version that introduces Kerberos and token system in 2009. However, this system is vulnerable to the replay attack, impersonation attack and other attacks. In this paper, we analyze these vulnerabilities of HDFS security and propose a new protocol which complements these vulnerabilities and maintains the performance of Hadoop.

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.

Design and Implementation of Sensor Information Management System based on Celery-MongoDB (Celery-MongoDB 를 활용한 센서정보 관리시스템 설계 및 구현)

  • Kang, Yun-Hee
    • Journal of Platform Technology
    • /
    • v.9 no.2
    • /
    • pp.3-9
    • /
    • 2021
  • The management of sensor information requires the functions for registering, modifying and deleting rapidly sensor information about various many sensors. In this research, Celery and MongoDB are used for developing a sensory data management system. Celery supplies a queue structure based on asynchronous communication in Python. Celery is a distributed simple job-queue but reliable distributed system suitable for processing large message. MongoDB is a NoSQL database that is capable of managing various informal information. In this experiment, we have checked that variety of sensor information can be processed with this system in a IoT environment. To improve the performance for handling a message with sensory data, this system will be deployed in the edge of a cloud infrastructure.