• Title/Summary/Keyword: Single Time Schemes

Search Result 137, Processing Time 0.023 seconds

Transmission Techniques for Downlink Multi-Antenna MC-CDMA Systems in a Beyond-3G Context

  • Portier Fabrice;Raos Ivana;Silva Adao;Baudais Jean-Yves;Helard Jean-Francois;Gameiro Atilio;Zazo Santiago
    • Journal of Communications and Networks
    • /
    • v.7 no.2
    • /
    • pp.157-170
    • /
    • 2005
  • The combination of multiple antennas and multi-carrier code division multiple-access (MC-CDMA) is a strong candidate for the downlink of the next generation mobile communications. The study of such systems in scenarios that model real-life trans-missions is an additional step towards an optimized achievement. We consider a realistic MIMO channel with two or four transmit antennas and up to two receive antennas, and channel state information (CSI) mismatches. Depending on the mobile terminal (MT) class, its number of antennas or complexity allowed, different data-rates are proposed with turbo-coding and asymptotic spectral efficiencies from 1 to 4.5 bit/s/Hz, using three algorithms developed within the European IST-MATRICE project. These algorithms can be classified according to the degree of CSI at base-station (BS): i) Transmit space-frequency prefiltering based on constrained zero-forcing algorithm with complete CSI at BS; ii) transmit beamforming based on spatial correlation matrix estimation from partial CSI at BS; iii) orthogonal space-time block coding based on Alamouti scheme without CSI at BS. All presented schemes require a reasonable complexity at MT, and are compatible with a single-antenna receiver. A choice between these algorithms is proposed in order to significantly improve the performance of MC-CDMA and to cover the different environments considered for the next generation cellular systems. For beyond-3G, we propose prefiltering for indoor and pedestrian microcell environments, beamforming for suburban macrocells including high-speed train, and space-time coding for urban conditions with moderate to high speeds.

Complexity Estimation Based Work Load Balancing for a Parallel Lidar Waveform Decomposition Algorithm

  • Jung, Jin-Ha;Crawford, Melba M.;Lee, Sang-Hoon
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.6
    • /
    • pp.547-557
    • /
    • 2009
  • LIDAR (LIght Detection And Ranging) is an active remote sensing technology which provides 3D coordinates of the Earth's surface by performing range measurements from the sensor. Early small footprint LIDAR systems recorded multiple discrete returns from the back-scattered energy. Recent advances in LIDAR hardware now make it possible to record full digital waveforms of the returned energy. LIDAR waveform decomposition involves separating the return waveform into a mixture of components which are then used to characterize the original data. The most common statistical mixture model used for this process is the Gaussian mixture. Waveform decomposition plays an important role in LIDAR waveform processing, since the resulting components are expected to represent reflection surfaces within waveform footprints. Hence the decomposition results ultimately affect the interpretation of LIDAR waveform data. Computational requirements in the waveform decomposition process result from two factors; (1) estimation of the number of components in a mixture and the resulting parameter estimates, which are inter-related and cannot be solved separately, and (2) parameter optimization does not have a closed form solution, and thus needs to be solved iteratively. The current state-of-the-art airborne LIDAR system acquires more than 50,000 waveforms per second, so decomposing the enormous number of waveforms is challenging using traditional single processor architecture. To tackle this issue, four parallel LIDAR waveform decomposition algorithms with different work load balancing schemes - (1) no weighting, (2) a decomposition results-based linear weighting, (3) a decomposition results-based squared weighting, and (4) a decomposition time-based linear weighting - were developed and tested with varying number of processors (8-256). The results were compared in terms of efficiency. Overall, the decomposition time-based linear weighting work load balancing approach yielded the best performance among four approaches.

Multi-task Learning Based Tropical Cyclone Intensity Monitoring and Forecasting through Fusion of Geostationary Satellite Data and Numerical Forecasting Model Output (정지궤도 기상위성 및 수치예보모델 융합을 통한 Multi-task Learning 기반 태풍 강도 실시간 추정 및 예측)

  • Lee, Juhyun;Yoo, Cheolhee;Im, Jungho;Shin, Yeji;Cho, Dongjin
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1037-1051
    • /
    • 2020
  • The accurate monitoring and forecasting of the intensity of tropical cyclones (TCs) are able to effectively reduce the overall costs of disaster management. In this study, we proposed a multi-task learning (MTL) based deep learning model for real-time TC intensity estimation and forecasting with the lead time of 6-12 hours following the event, based on the fusion of geostationary satellite images and numerical forecast model output. A total of 142 TCs which developed in the Northwest Pacific from 2011 to 2016 were used in this study. The Communications system, the Ocean and Meteorological Satellite (COMS) Meteorological Imager (MI) data were used to extract the images of typhoons, and the Climate Forecast System version 2 (CFSv2) provided by the National Center of Environmental Prediction (NCEP) was employed to extract air and ocean forecasting data. This study suggested two schemes with different input variables to the MTL models. Scheme 1 used only satellite-based input data while scheme 2 used both satellite images and numerical forecast modeling. As a result of real-time TC intensity estimation, Both schemes exhibited similar performance. For TC intensity forecasting with the lead time of 6 and 12 hours, scheme 2 improved the performance by 13% and 16%, respectively, in terms of the root mean squared error (RMSE) when compared to scheme 1. Relative root mean squared errors(rRMSE) for most intensity levels were lessthan 30%. The lower mean absolute error (MAE) and RMSE were found for the lower intensity levels of TCs. In the test results of the typhoon HALONG in 2014, scheme 1 tended to overestimate the intensity by about 20 kts at the early development stage. Scheme 2 slightly reduced the error, resulting in an overestimation by about 5 kts. The MTL models reduced the computational cost about 300% when compared to the single-tasking model, which suggested the feasibility of the rapid production of TC intensity forecasts.

Performance Analysis of Slave-Side Arbitration Schemes for the Multi-Layer AHB BusMatrix (ML-AHB 버스 매트릭스를 위한 슬레이브 중심 중재 방식의 성능 분석)

  • Hwang, Soo-Yun;Park, Hyeong-Jun;Jhang, Kyoung-Son
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.34 no.5_6
    • /
    • pp.257-266
    • /
    • 2007
  • In On-Chip bus, the arbitration scheme is one of the critical factors that decide the overall system performance. The arbitration scheme used in traditional shared bus is the master-side arbitration based on the request and grant signals between multiple masters and single arbiter. In the case of the master-side arbitration, only one master and one slave can transfer the data at a time. Therefore the throughput of total bus system and the utilization of resources are decreased in the master-side arbitration. However in the slave-side arbitration, there is an arbiter at each slave port and the master just starts a transaction and waits for the slave response to proceed to the next transfer. Thus, the unit of arbitration can be a transaction or a transfer. Besides the throughput of total bus system and the utilization of resources are increased since the multiple masters can simultaneously perform transfers with independent slaves. In this paper, we implement and analyze the arbitration schemes for the Multi-Layer AHB BusMatrix based on the slave-side arbitration. We implement the slave-side arbitration schemes based on fixed priority, round robin and dynamic priority and accomplish the performance simulation to compare and analyze the performance of each arbitration scheme according to the characteristics of the master and slave. With the performance simulation, we observed that when there are few masters on critical path in a bus system, the arbitration scheme based on dynamic priority shows the maximum performance and in other cases, the arbitration scheme based on round robin shows the highest performance. In addition, the arbitration scheme with transaction based multiplexing shows higher performance than the same arbitration scheme with single transfer based switching in an application with frequent accesses to the long latency devices or memories such as SDRAM. The improvements of the arbitration scheme with transaction based multiplexing are 26%, 42% and 51%, respectively when the latency times of SDRAM are 1, 2 and 3 clock cycles.

Comparison and analysis of data-derived stage prediction models (자료 지향형 수위예측 모형의 비교 분석)

  • Choi, Seung-Yong;Han, Kun-Yeun;Choi, Hyun-Gu
    • Journal of Wetlands Research
    • /
    • v.13 no.3
    • /
    • pp.547-565
    • /
    • 2011
  • Different types of schemes have been used in stage prediction involving conceptual and physical models. Nevertheless, none of these schemes can be considered as a single superior model. To overcome disadvantages of existing physics based rainfall-runoff models for stage predicting because of the complexity of the hydrological process, recently the data-derived models has been widely adopted for predicting flood stage. The objective of this study is to evaluate model performance for stage prediction of the Neuro-Fuzzy and regression analysis stage prediction models in these data-derived methods. The proposed models are applied to the Wangsukcheon in Han river watershed. To evaluate the performance of the proposed models, fours statistical indices were used, namely; Root mean square error(RMSE), Nash Sutcliffe efficiency coefficient(NSEC), mean absolute error(MAE), adjusted coefficient of determination($R^{*2}$). The results show that the Neuro-Fuzzy stage prediction model can carry out the river flood stage prediction more accurately than the regression analysis stage prediction model. This study can greatly contribute to the construction of a high accuracy flood information system that secure lead time in medium and small streams.

Two-Way Donation Locking for Transaction Management in Distributed Database Systems (분산환경에서 거래관리를 위한 두단계 기부 잠금규약)

  • Rhee, Hae-Kyung;Kim, Ung-Mo
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.12
    • /
    • pp.3447-3455
    • /
    • 1999
  • Database correctness is guaranteed by standard transaction scheduling schemes like two-phase locking for the context of concurrent execution environment in which short-lived ones are normally mixed with long-lived ones. Traditional syntax-oriented serializability notions are considered to be not enough to handle in particular various types of transaction in terms of duration of execution. To deal with this situation, altruistic locking has attempted to reduce delay effect associated with lock release moment by use of the idea of donation. An improved form of altruism has also been deployed in extended altruistic locking in a way that scope of data to be early released is enlarged to include even data initially not intended to be donated. In this paper, we first of all investigated limitations inherent in both altruistic schemes from the perspective of alleviating starvation occasions for transactions in particular of short-lived nature. The idea of two-way donation locking(2DL) has then been experimented to see the effect of more than single donation in distributed database systems. Simulation experiments shows that 2DL outperforms the conventional two-phase locking in terms of the degree of concurrency and average transaction waiting time under the circumstances that the size of long-transaction is in between 5 and 9.

  • PDF

Distributed Address Configuration Mechanism Using Address Pool in MANET (MANET 환경에서 주소 영역기반 주소할당 분산기법)

  • Ahn, Sang-Hyun;Lim, Yu-Jin;Yu, Hyun
    • The KIPS Transactions:PartC
    • /
    • v.16C no.6
    • /
    • pp.753-760
    • /
    • 2009
  • As it becomes increasingly important that Internet access is available anywhere at any time, providing MANET (Mobile Ad-Hoc Network) with the Internet access attracts more attention. The existing DHCP (Dynamic Host Configuration Protocol) address configuration schemes require message exchanges between MANET nodes and the DHCP server through multi-hop routes. Messages can be easily dropped in the wireless multi-hop communication environment and the address configuration may be instable and need long delay. In this paper, we propose a new address configuration scheme using the concept of address pool allocation. In the proposed scheme, the DHCP server assigns a part of its address pool to a node instead of a single address and the node can assign a part of its own address pool to its neighbor nodes. Through simulation studies, we show that our scheme yields about 77% of the address configuration delay and 61% of the control message overhead of the existing DHCP based mechanism.

Hybrid Data Recovery Technique of RAID Level 5 Architecture (RAID 레벨 5 구조의 혼합형 데이타 복구 기법)

  • Jeon, Sang-Hoon;Ahn, Byung-Chul
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.27 no.7
    • /
    • pp.675-683
    • /
    • 2000
  • For real time services of multimedia applications, storage systems should provide data without degrading their performance. Since it is very important to recover data immediately at a disk failure, the disk recovery system is required. This paper presents a hybrid recovery scheme which prevents degraded the performance on a single disk failure at RAID level 5 architecture until the failed disk is replaced with a new disk. The proposed scheme is very economical compared to previous spare disk schemes because it does not use extra disks. The performance of the proposed scheme is evaluated and analyzed with that of the RAID level 5 for various requested sizes through the simulation. The results show that the performance of the proposed scheme is improved up to 20 percents at the failure mode and 80 percents at the reconfigured mode. After a multimedia server system has been built with a RAID controller and hard disks, the data recovery performance of the propose scheme are compared with the results of simulation.

  • PDF

Reliable Data Transmission Based on Erasure-resilient Code in Wireless Sensor Networks

  • Lei, Jian-Jun;Kwon, Gu-In
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.4 no.1
    • /
    • pp.62-77
    • /
    • 2010
  • Emerging applications with high data rates will need to transport bulk data reliably in wireless sensor networks. ARQ (Automatic Repeat request) or Forward Error Correction (FEC) code schemes can be used to provide reliable transmission in a sensor network. However, the naive ARQ approach drops the whole frame, even though there is a bit error in the frame and the FEC at the bit level scheme may require a highly complex method to adjust the amount of FEC redundancy. We propose a bulk data transmission scheme based on erasure-resilient code in this paper to overcome these inefficiencies. The sender fragments bulk data into many small blocks, encodes the blocks with LT codes and packages several such blocks into a frame. The receiver only drops the corrupted blocks (compared to the entire frame) and the original data can be reconstructed if sufficient error-free blocks are received. An incidental benefit is that the frame error rate (FER) becomes irrelevant to frame size (error recovery). A frame can therefore be sufficiently large to provide high utilization of the wireless channel bandwidth without sacrificing the effectiveness of error recovery. The scheme has been implemented as a new data link layer in TinyOS, and evaluated through experiments in a testbed of Zigbex motes. Results show single hop transmission throughput can be improved by at least 20% under typical wireless channel conditions. It also reduces the transmission time of a reasonable range of size files by more than 30%, compared to a frame ARQ scheme. The total number of bytes sent by all nodes in the multi-hop communication is reduced by more than 60% compared to the frame ARQ scheme.

Lifetime Maximizing Routing Algorithm for Multi-hop Wireless Networks (다중-홉 무선 네트워크 환경에서 수명 최대화를 위한 라우팅 알고리즘)

  • Lee, Keon-Taek;Han, Seung-Jae;Park, Sun-Ju
    • Journal of KIISE:Information Networking
    • /
    • v.35 no.4
    • /
    • pp.292-300
    • /
    • 2008
  • In multi-hop wireless networks like Wireless Mesh Networks (WMN) and Wireless Sensor Networks (WSN), nodes often rely on batteries as their power source. In such cases, energy efficient routing is critical. Many schemes have been proposed to find the most energy efficient path, but most of them do not achieve optimality on network lifetime. Once found, the energy efficient path is constantly used such that the energy of the nodes on the path is depleted quickly. As an alternative, the approaches that dynamically change the path at run time have also been proposed. These approaches, however, involve high overhead of establishing multiple paths. In this paper, we first find an optimal multi-path routing using LP. Then we apply an approximation algorithm to derive a near-optimal solution for single-path routing. We compare the performance of the proposed scheme with several other existing algorithms through simulation.