• Title/Summary/Keyword: linear block code

Search Result 57, Processing Time 0.019 seconds

Implementation of Fixslicing AES-CTR Speed Optimized Using Pre-Computed on 32-Bit RISC-V (32-bit RISC-V 상에서의 사전 연산을 활용한 Fixslicing AES-CTR 속도 최적화 구현)

  • Eum, Si-Woo;Kim, Hyun-Jun;Sim, Min-Joo;Song, Gyeong-Ju;Seo, Hwa-Jeong
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.32 no.1
    • /
    • pp.1-9
    • /
    • 2022
  • Fixslicing AES is a technique that omits the Shiftrows step to minimize the cost of the linear layer of Bitsliced AES, showing a 30% performance over the Bitsliced technique. However, the amount of code increases to compensate for the omitted shiftrows. Therefore, it is proposed to be divided into Semi-Fixsliced in which only half of shiftrows are omitted and Fully-Fixsliced in which Shiftrows are omitted completely. In this paper, we propose a CTR mode implementation of Fixslicing AES on RISC-V using the pre-computed table technique. By utilizing the characteristics of the CTR mode, it is possible to perform fast encryption by omitting up to the second round SubBytes from the encryption process through pre-computed up to the second round SubBytes operation. Using this technique, it was confirmed that Semi-Fixsliced has a performance of 1,345 cycles per block and a performance improvement of 7% compared to the previous performance result, and Fully-Fixsliced has a performance of 1,283 cycles per block and a performance of 9% compared to the previous performance result on 32-bit RISC-V.

Progressive Edge-Growth Algorithm for Low-Density MIMO Codes

  • Jiang, Xueqin;Yang, Yi;Lee, Moon Ho;Zhu, Minda
    • Journal of Communications and Networks
    • /
    • v.16 no.6
    • /
    • pp.639-644
    • /
    • 2014
  • In low-density parity-check (LDPC) coded multiple-input multiple-output (MIMO) communication systems, probabilistic information are exchanged between an LDPC decoder and a MIMO detector. TheMIMO detector has to calculate probabilistic values for each bit which can be very complex. In [1], the authors presented a class of linear block codes named low-density MIMO codes (LDMC) which can reduce the complexity of MIMO detector. However, this code only supports the outer-iterations between the MIMO detector and decoder, but does not support the inner-iterations inside the LDPC decoder. In this paper, a new approach to construct LDMC codes is introduced. The new LDMC codes can be encoded efficiently at the transmitter side and support both of the inner-iterations and outer-iterations at the receiver side. Furthermore they can achieve the design rates and perform very well over MIMO channels.

A Modified Product Code Over ℤ4 in Steganography with Large Embedding Rate

  • Zhang, Lingyu;Chen, Deyuan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.7
    • /
    • pp.3353-3370
    • /
    • 2016
  • The way of combination of Product Perfect Codes (PPCs) is based on the theory of short codes constructing long codes. PPCs have larger embedding rate than Hamming codes by expending embedding columns in a coding block, and they have been proven to enhance the performance of the F5 steganographic method. In this paper, the proposed modified product codes called MPCs are introduced as an efficient way to embed more data than PPCs by increasing 2r2-1-r2 embedding columns. Unlike PPC, the generation of the check matrix H in MPC is random, and it is different from PPC. In addition a simple solving way of the linear algebraic equations is applied to figure out the problem of expending embedding columns or compensating cases. Furthermore, the MPCs over ℤ4 have been proposed to further enhance not only the performance but also the computation speed which reaches O(n1+σ). Finally, the proposed ℤ4-MPC intends to maximize the embedding rate with maintaining less distortion , and the performance surpasses the existing improved product perfect codes. The performance of large embedding rate should have the significance in the high-capacity of covert communication.

On the Adaptive 3-dimensional Transform Coding Technique Employing the Variable Length Coding Scheme (가변 길이 부호화를 이용한 적응 3차원 변환 부호화 기법)

  • 김종원;이신호;이상욱
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.30B no.7
    • /
    • pp.70-82
    • /
    • 1993
  • In this paper, employing the 3-dimensional discrete cosine transform (DCT) for the utilization of the temporal correlation, an adaptive motion sequence coding technique is proposed. The energy distribution in a 3-D DCT block, due to the nonstationary nature of the image data, varies along the veritical, horizontal and temporal directions. Thus, aiming an adaptive system to local variations, adaptive procedures, such as the 3-D classification, the classified linear scanning technique and the VLC table selection scheme, have been implemented in our approach. Also, a hybrid structure which adaptively combines inter-frame coding is presented, and it is found that the adaptive hybrid frame coding technique shows a significant performance gain for a moving sequence which contains a relatively small moving area. Through an intensive computer simulation, it is demonstrated that, the performance of the proposed 3-D transform coding technique shows a close relation with the temporal variation of the sequence to be code. And the proposed technique has the advantages of skipping the computationally complex motion compensation procedure and improving the performance over the 2-D motion compensated transform coding technique for rates in the range of 0.5 ~ 1.0 bpp.

  • PDF

Prediction of Equivalent Stress Block Parameters for High Strength Concrete (고강도 콘크리트의 등가응력 매개변수 추정에 관한 연구)

  • Lee, Do Hyung;Jeon, Jeongmoon;Jeong, Minchul;Kong, Jungsik
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.31 no.3A
    • /
    • pp.227-234
    • /
    • 2011
  • Recently, a high strength concrete of more than 40 MPa has been increasingly used in practice. However, use of the high strength concrete may influence on design parameters, particularly stress distribution. This is very true since the current everyday practice employs equivalent rectangular stress distribution that is derived from normal strength concrete. Subsequently, the stress distribution seems to be reevaluated and then a new distribution with new parameters needs to be suggested for the high strength concrete. For this purpose, linear and multiple regression analyses have been carried out in term of using experimental data for the high strength concrete of 40 to 80 MPa available in literatures. Accordingly, new parameters associated with the stress distribution have been proposed and employed for the design of flexural and compressive members. Comparative design examples indicate that designs with new parameters reduce section dimensions compared to those with the current code parameters for concrete strengths of 40 to 70 MPa. In particular, for compressive members, design with new parameters exhibit conservative compressive force compared to those with the current code parameters.

Performance of pilot-assisted coded-OFDM-CDMA using low-density parity-check coding in Rayleigh fading channels (레일리 페이딩 채널에서 파일럿 기법과 LDPC 코딩이 적용된 COFDM-CDMA의 성능 분석)

  • 안영신;최재호
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.5C
    • /
    • pp.532-538
    • /
    • 2003
  • In this paper we have investigated a novel approach applying low-density parity-check coding to a COFDM-CDMA system, which operates in a multi-path fading mobile channel. Developed as a linear-block channel coder, the LDPC code is known for a superior signal reception capability in AWGN and/or flat fading channels with respect to increased encoding rates, however, its performance degrades when the communication channel becomes multi-path fading. For a typical multi-path fading mobile channel with a SNR of 16㏈ or lower. in order to obtain a BER lower than 1 out of 10000, the LDPC code with encoding rates below 1:3 requires not only the inherent parity check information but also the piloting information for refreshing front-end equalizer taps of COFDM-CDMA, periodically. For instance, while the 1:3-rate LDPC coded transmission symbol is consisted of data bits and parity-check bits in 1 to 3 proportion, on the other hand, in the proposed method the same rate LDPC transmission symbol contains data bits, parity check bits, and pilot bits in 1 to 2 to 1 proportion, respectively. The included pilot bits are effective not only for channel estimation and channel equalization but for symbol decoding by assisting the parity-check bits, hence, improving SNR vs BER performance over the conventional 1:3-rate LDPC code. The proposed system performance has been verified using computer simulations in multi-path, Rayleigh fading channels, and the results show us that the proposed method out-performs the general LDPC channel coding methods in terms of SNR vs BER measurements.

Parallel Multithreaded Processing for Data Set Summarization on Multicore CPUs

  • Ordonez, Carlos;Navas, Mario;Garcia-Alvarado, Carlos
    • Journal of Computing Science and Engineering
    • /
    • v.5 no.2
    • /
    • pp.111-120
    • /
    • 2011
  • Data mining algorithms should exploit new hardware technologies to accelerate computations. Such goal is difficult to achieve in database management system (DBMS) due to its complex internal subsystems and because data mining numeric computations of large data sets are difficult to optimize. This paper explores taking advantage of existing multithreaded capabilities of multicore CPUs as well as caching in RAM memory to efficiently compute summaries of a large data set, a fundamental data mining problem. We introduce parallel algorithms working on multiple threads, which overcome the row aggregation processing bottleneck of accessing secondary storage, while maintaining linear time complexity with respect to data set size. Our proposal is based on a combination of table scans and parallel multithreaded processing among multiple cores in the CPU. We introduce several database-style and hardware-level optimizations: caching row blocks of the input table, managing available RAM memory, interleaving I/O and CPU processing, as well as tuning the number of working threads. We experimentally benchmark our algorithms with large data sets on a DBMS running on a computer with a multicore CPU. We show that our algorithms outperform existing DBMS mechanisms in computing aggregations of multidimensional data summaries, especially as dimensionality grows. Furthermore, we show that local memory allocation (RAM block size) does not have a significant impact when the thread management algorithm distributes the workload among a fixed number of threads. Our proposal is unique in the sense that we do not modify or require access to the DBMS source code, but instead, we extend the DBMS with analytic functionality by developing User-Defined Functions.