• Title/Summary/Keyword: server performance

Search Result 1,690, Processing Time 0.024 seconds

Design and Implementation of SDR-based Multi-Constellation Multi-Frequency Real-Time A-GNSS Receiver Utilizing GPGPU

  • Yoo, Won Jae;Kim, Lawoo;Lee, Yu Dam;Lee, Taek Geun;Lee, Hyung Keun
    • Journal of Positioning, Navigation, and Timing
    • /
    • v.10 no.4
    • /
    • pp.315-333
    • /
    • 2021
  • Due to the Global Navigation Satellite System (GNSS) modernization, recently launched GNSS satellites transmit signals at various frequency bands such as L1, L2 and L5. Considering the Korean Positioning System (KPS) signal and other GNSS augmentation signals in the future, there is a high probability of applying more complex communication techniques to the new GNSS signals. For the reason, GNSS receivers based on flexible Software Defined Radio (SDR) concept needs to be developed to evaluate various experimental communication techniques by accessing each signal processing module in detail. This paper proposes a novel SDR-based A-GNSS receiver capable of processing multi-GNSS/RNSS signals at multi-frequency bands. Due to the modular structure, the proposed receiver has high flexibility and expandability. For real-time implementation, A-GNSS server software is designed to provide immediate delivery of satellite ephemeris data on demand. Due to the sampling bandwidth limitation of RF front-ends, multiple SDRs are considered to process the multi-GNSS/RNSS multi-frequency signals simultaneously. To avoid the overflow problem of sampled RF data, an efficient memory buffer management strategy was considered. To collect and process the multi-GNSS/RNSS multi-frequency signals in real-time, the proposed SDR A-GNSS receiver utilizes multiple threads implemented on a CPU and multiple NVIDIA CUDA GPGPUs for parallel processing. To evaluate the performance of the proposed SDR A-GNSS receiver, several experiments were performed with field collected data. By the experiments, it was shown that A-GNSS requirements can be satisfied sufficiently utilizing only milliseconds samples. The continuous signal tracking performance was also confirmed with the hundreds of milliseconds data for multi-GNSS/RNSS multi-frequency signals and with the ten-seconds data for multi-GNSS/RNSS single-frequency signals.

Performance Evaluation of Smoothing Algorithm Considering Network Bandwidth in IoT Environment (IoT 환경에서 네트워크 대역폭을 고려한 스무딩 알고리즘의 성능 평가)

  • Lee, MyounJae
    • Journal of Internet of Things and Convergence
    • /
    • v.8 no.4
    • /
    • pp.41-47
    • /
    • 2022
  • Smoothing is a transmission plan that converts video data stored at a variable bit rate into a fixed bit rate. Algorithms for smoothing include CBA, which aims to minimize the number of transmission rate increases, MCBA, which minimizes the number of transmission rate changes, and MVBA algorithms that minimize the amount of transmission rate change. This paper compares the proposed algorithm with the CBA algorithm with various video data, buffer size, and performance evaluation factors as a follow-up to the proposed smoothing algorithm that minimizes (maximizes) the transmission rate increase (decrease) when the server requires more bandwidth The evaluation factors used were compared with the number of changes in the fps rate, the minimum fps, the average fps, fps variability, and the number of frames to be discarded. As a result of the comparison, the proposed algorithm showed superiority in comparing the number of fps rate changes and the number of frames discarded.

Implementation of an open API-based virtual network provisioning automation platform for large-scale data transfer (대용량 데이터 전송을 위한 오픈 API 기반 가상 네트워크 프로비저닝 자동화 플랫폼 구현)

  • Kim, Yong-hwan;Park, Seongjin;Kim, Dongkyun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.9
    • /
    • pp.1320-1329
    • /
    • 2022
  • Currently, advanced national research network groups are continuously conducting R&D for the requirement to provide SDN/NFV-based network automation and intelligence technology for R&E users. In addition, the requirement for providing large-scale data transmission with the high performance networking facility, compared to general network environments, is gradually increasing in the advanced national research networks. Accordingly, in this paper, we propose an open API-based virtual network provisioning automation platform for large data transmission researched and developed to respond to the networking requirements of the national research network and present the implementation results. The platform includes the KREONET-S VDN system that provides SDN-based network virtualization technology, and the Kubernetes system that provides container-oriented server virtualization technology, and the Globus Online, a high-performance data transmission system. In this paper, the environment configurations, the system implemetation results for the interworking between the heterogeneous systems, and the automated virtual network provisioning implementation results are presented.

Real-time Task Aware Memory Allocation Techniques for Heterogeneous Mobile Multitasking Environments (이종 모바일 멀티태스킹 환경을 위한 실시간 작업 인지형 메모리 할당 기술 연구)

  • Bahn, Hyokyung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.3
    • /
    • pp.43-48
    • /
    • 2022
  • Recently, due to the rapid performance improvement of smartphones and the increase in background executions of mobile apps, multitasking has become common on mobile platforms. Unlike traditional desktop and server apps, response time is important in most mobile apps as they are interactive tasks, and some apps are classified as real-time tasks with deadlines. In this paper, we discuss how to meet the requirements of heterogeneous multitasking in managing memory of real-time and interactive tasks when they are executed together on a smartphone. To do so, we analyze the memory requirement of real-time tasks, and propose a model that has the ability of allocating memory to multitasking tasks on a smartphone. Trace-driven simulations with real-world storage access traces captured by heterogeneous apps show that the proposed model provides reasonable performance for interactive tasks while guaranteeing the requirement of real-time tasks.

A Typo Correction System Using Artificial Neural Networks for a Text-based Ornamental Fish Search Engine

  • Hyunhak Song;Sungyoon Cho;Wongi Jeon;Kyungwon Park;Jaedong Shim;Kiwon Kwon
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.8
    • /
    • pp.2278-2291
    • /
    • 2023
  • Imported ornamental fish should be quarantined because they can have dangerous diseases depending on their habitat. The quarantine requires a lot of time because quarantine officers collect various information on the imported ornamental fish. Inefficient quarantine processes reduce its work efficiency and accuracy. Also, long-time quarantine causes the death of environmentally sensitive ornamental fish and huge financial losses. To improve existing quarantine systems, information on ornamental fish was collected and structured, and a server was established to develop quarantine performance support software equipped with a text search engine. However, the long names of ornamental fish in general can cause many typos and time bottlenecks when we type search words for the target fish information. Therefore, we need a technique that can correct typos. Typical typo character calibration compares input text with all characters in a calibrated candidate text dictionary. However, this approach requires computational power proportional to the number of typos, resulting in slow processing time and low calibration accuracy performance. Therefore, to improve the calibration accuracy of characters, we propose a fusion system of simple Artificial Neural Network (ANN) models and character preprocessing methods that accelerate the process by minimizing the computation of the models. We also propose a typo character generation method used for training the ANN models. Simulation results show that the proposed typo character correction system is about 6 times faster than the conventional method and has 10% higher accuracy.

Smoothing Algorithm Considering Server Bandwidth and Network Traffic in IoT Environments (IoT 환경에서 서버 대역폭과 네트워크 트래픽을 고려한 스무딩 알고리즘)

  • Lee, MyounJae
    • Journal of Internet of Things and Convergence
    • /
    • v.8 no.1
    • /
    • pp.53-58
    • /
    • 2022
  • Smoothing is a transmission plan that converts video data stored at a variable bit rate into a constant bit rate. In the study of [6-7], when a data rate increase is required, the frame with the smallest increase is set as the start frame of the next transmission rate section, when a data tate decrease is required. the frame with the largest decrease is set as the start frame of the next transmission rate section, And the smoothing algorithm was proposed and performance was evaluated in an environment where network traffic is not considered. In this paper, the smoothing algorithm of [6-7] evaluates the adaptive CBA algorithm and performance with minimum frame rate, average frame rate, and frame rate variation from 512KB to 32MB with E.T 90 video data in an environment that considers network traffic. As a result of comparison, the smoothing algorithm of [6-7] showed superiority in the comparison of the minimum refresh rate.

Quantitative Queue Estimation and Improvement of Drive-Through with Queuing (대기행렬을 적용한 승차 구매점의 정량적인 대기열 산정과 개선방안)

  • Lee, SeungWon;Huh, SeungHa;Yoon, KyoungIl;Kim, JaeJun
    • Korean Journal of Construction Engineering and Management
    • /
    • v.24 no.1
    • /
    • pp.21-30
    • /
    • 2023
  • Excessive complaints and traffic jams occurred as customers who visited the Drive-thru waited in a long line. Company S recommends DT Pass to reduce the queues. Therefore, this study confirmed the improvement in performance of the queue increasing the number of stores operated by two servers insteaol of one using a queue model. And then confirmed performance improvement by dividing them into DT and DT Pass. After that, the L value derived through the queue model and the number of queues in each store were compared to calculate the number of queues to be additionally provided. Through this, the validity of selecting the minimum number of queues in the future is verified based on the results derived in this study.

Survey on the Performance Enhancement in Serverless Computing: Current and Future Directions (성능 향상을 위한 서버리스 컴퓨팅 동향과 발전 방향)

  • Eunyoung Lee
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.2
    • /
    • pp.60-75
    • /
    • 2024
  • The demand of users, who want to focus on the core functionality of their applications without having to manage complex virtual environments in the cloud environment, has created a new computing model called serverless computing. Within the serverless paradigm, resource provisioning and server administration tasks are delegated to cloud services, facilitating application development exclusively focused on program logic. Serverless computing has upgraded the utilization of cloud computing by reducing the burden on cloud service users, and it is expected to become the basic model of cloud computing in the future. A serverless platform is responsible for managing the cloud virtual environment on behalf of users, and it is also responsible for executing serverless functions that compose applications in the cloud environment. Considering the characteristics of serverless computing in which users are billed in proportion to the resources used, the efficiency of the serverless platform is a very important factor for both users and service providers. This paper aims to identify various factors that affect the performance of serverless computing and analyze the latest research trends related to it. Drawing upon the analysis, the future directions for serverless computing that address key challenges and opportunities in serverless computing are proposed.

Queue Lengths and Sojourn Time Analysis of Discrete-time BMAP/G/1 Queue under the Workload Control (일량제어정책을 갖는 이산시간 BMAP/G/1 대기행렬의 고객수와 체재시간 분석)

  • Se Won Lee
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.29 no.1
    • /
    • pp.63-76
    • /
    • 2024
  • In this study, we analyzed queue length and sojourn time of discrete-time BMAP/G/1 queues under the workload control. Group customers (packets) with correlations arrive at the system following a discrete-time Markovian arrival process. The server starts busy period when the total service time of the arrived customers exceeds a predetermined workload threshold D and serves customers until the system is empty. From the analysis of workload and waiting time, distributions of queue length at the departure epoch and arbitrary time epoch and system sojourn time are derived. We also derived the mean value as a performance measure. Through numerical examples, we confirmed that we can obtain results represented by complex forms of equations, and we verified the validity of the theoretical values by comparing them with simulation results. From the results, we can obtain key performance measures of complex systems that operate similarly in various industrial fields and to analyze various optimization problems.

Boot storm Reduction through Artificial Intelligence Driven System in Virtual Desktop Infrastructure

  • Heejin Lee;Taeyoung Kim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.7
    • /
    • pp.1-9
    • /
    • 2024
  • In this paper, we propose BRAIDS, a boot storm mitigation plan consisting of an AI-based VDI usage prediction system and a virtual machine boot scheduler system, to alleviate boot storms and improve service stability. Virtual Desktop Infrastructure (VDI) is an important technology for improving an organization's work productivity and increasing IT infrastructure efficiency. Boot storms that occur when multiple virtual desktops boot simultaneously cause poor performance and increased latency. Using the xgboost algorithm, existing VDI usage data is used to predict future VDI usage. In addition, it receives the predicted usage as input, defines a boot storm considering the hardware specifications of the VDI server and virtual machine, and provides a schedule to sequentially boot virtual machines to alleviate boot storms. Through the case study, the VDI usage prediction model showed high prediction accuracy and performance improvement, and it was confirmed that the boot storm phenomenon in the virtual desktop environment can be alleviated and IT infrastructure can be utilized efficiently through the virtual machine boot scheduler.