• Title/Summary/Keyword: BITs

Search Result 1,631, Processing Time 0.028 seconds

A Bit Allocation Algorithm Using Adaptive Bandwidth for DMT (적응적인 대역폭을 이용한 DMT에서의 비트 할당 알고리듬)

  • 최현우;신봉식;정정화
    • Proceedings of the IEEK Conference
    • /
    • 1999.11a
    • /
    • pp.372-375
    • /
    • 1999
  • This paper proposes a bit allocation algorithm using adaptive bandwidth for ADSL that uses the DMT technology. In certain cases for high attenuation loops the conventional algorithms are unable to assign data bits to the higher frequency tones, due to the power spectrum mask limitation recommended by ANSI Standard, even if the total power budget is not expended. In the proposed bit allocation algorithm, adjacent empty tones that would not be used merge into single tone, then additional bits is assigned to the merged empty tones. Because additional bits is allocated, most of the available power is used. The proposed algorithm show that total bit increase in about 2~9% bits more than about conventional "water-filling" and "bit removal" algorithms and that is able to use about 93% of the available budget Power

  • PDF

Adaptive Near-Lossless Image Coding (적응적 준무손실 영상 부호화)

  • Kim, Young-Ro;Yi, Joon-Hwan
    • 전자공학회논문지 IE
    • /
    • v.46 no.4
    • /
    • pp.42-48
    • /
    • 2009
  • In this paper, we propose adaptive near-lossless image coding algorithm according to bandwidth while maintaining image quality. The proposed method adjusts error range using amounts of encoded bits and target bits at a slice encoding interval. Experimental results show that our proposed method not only almost fits compression into bandwidth, but also has better image quality.

High capacity multi-bit data hiding based on modified histogram shifting technique

  • Sivasubramanian, Nandhini;Konganathan, Gunaseelan;Rao, Yeragudipati Venkata Ramana
    • ETRI Journal
    • /
    • v.40 no.5
    • /
    • pp.677-686
    • /
    • 2018
  • A novel data hiding technique based on modified histogram shifting that incorporates multi-bit secret data hiding is proposed. The proposed technique divides the image pixel values into embeddable and nonembeddable pixel values. Embeddable pixel values are those that are within a specified limit interval surrounding the peak value of an image. The limit interval is calculated from the number of secret bits to be embedded into each embeddable pixel value. The embedded secret bits can be perfectly extracted from the stego image at the receiver side without any overhead bits. From the simulation, it is found that the proposed technique produces a better quality stego image compared to other data hiding techniques, for the same embedding rate. Since the proposed technique only embeds the secret bits in a limited number of pixel values, the change in the visual quality of the stego image is negligible when compared to other data hiding techniques.

Quantization of Line Spectrum Pair Frequencies using Lattice Vector Quantizers (격자벡터양자화기를 이용한 음성신호의 LSP 주파수 양자화)

  • 강정원;정재호;정대권
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.21 no.10
    • /
    • pp.2634-2644
    • /
    • 1996
  • Two different low rate speech coders using one of four types of lattice vector quantizers(LVQ's) with fairly low complexity were investigated for an application to mobile communications. More specifically, two-stage vector quantizer-lattic vector quantizer(VQ-LVQ) systems and vector differenctial pulse code modulation(VDPCM)systems with lattice vector quantizers simulated to encode the line spectrum frequencies of various sentences at the rate 22 to 39 bits per 20 msec frame. The simulation results showed that the VDPCM system with the lattice VQ can save up to 10 bits/fram compared to the quantization scheme used in QCELP system. For the VQ-LVQ system, the spherical quasi-uniform LVQ below 36 bits/frame outperformed the other 3 types of LVQ's and the pyramidal quasi-uniform LVQ at 37 bits/frame outperformed the other 3 types of LVQ's with the spectral distortion 0.97.

  • PDF

Guess and Determine Attack on Bivium

  • Rohani, Neda;Noferesti, Zainab;Mohajeri, Javad;Aref, Mohammad Reza
    • Journal of Information Processing Systems
    • /
    • v.7 no.1
    • /
    • pp.151-158
    • /
    • 2011
  • Bivium is a simplified version of Trivium, a hardware profile finalist of the eSTREAM project. Bivium has an internal state size of 177 bits and a key length of 80 bits. In this paper, a guess and determine attack on this cipher is introduced. In the proposed method, the best linear approximations for the updating functions are first defined. Then by using these calculated approximations, a system of linear equations is built. By guessing 30 bits of internal state, the system is solved and all the other 147 remaining bits are determined. The complexity of the attack is O ($2^{30}$), which is an improvement to the previous guess and determine attack with a complexity of order O($2^{52.3}$).

Development of Encryption System for Facsimile (팩시밀리용 암호 시스템 개발)

  • Yoo Byong-wook;Han Sang-soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.2
    • /
    • pp.344-352
    • /
    • 2006
  • We developed the facsimile encryption system that could send/receive the cipher text and plain text automatically. This facsimile encryption system is adopting the 128bits SEED encryption algorithm, 1024bits modular and 256bits exponential Diffie-Hellman key exchange method. We can reduce the cipher text transmitting time by developing the Dual Facsimile Server, it can make the real time encryption communication. Key is safely exchanged by the exchanging the listed serial numbers when the Diffie-Hellman key exchanging.

Considering Read and Write Characteristics of Page Access Separately for Efficient Memory Management

  • Hyokyung Bahn
    • International journal of advanced smart convergence
    • /
    • v.12 no.1
    • /
    • pp.70-75
    • /
    • 2023
  • With the recent proliferation of memory-intensive workloads such as deep learning, analyzing memory access characteristics for efficient memory management is becoming increasingly important. Since read and write operations in memory access have different characteristics, an efficient memory management policy should take into accountthe characteristics of thesetwo operationsseparately. Although some previous studies have considered the different characteristics of reads and writes, they require a modified hardware architecture supporting read bits and write bits. Unlike previous approaches, we propose a software-based management policy under the existing memory architecture for considering read/write characteristics. The proposed policy logically partitions memory space into the read/write area and the write area by making use of reference bits and dirty bits provided in modern paging systems. Simulation experiments with memory access traces show that our approach performs better than the CLOCK algorithm by 23% on average, and the effect is similar to the previous policy with hardware support.

Generalized SCAN Bit-Flipping Decoding Algorithm for Polar Code

  • Lou Chen;Guo Rui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.4
    • /
    • pp.1296-1309
    • /
    • 2023
  • In this paper, based on the soft cancellation (SCAN) bit-flipping (SCAN-BF) algorithm, a generalized SCAN bit-flipping (GSCAN-BF-Ω) decoding algorithm is carried out, where Ω represents the number of bits flipped or corrected at the same time. GSCAN-BF-Ω algorithm corrects the prior information of the code bits and flips the prior information of the unreliable information bits simultaneously to improve the block error rate (BLER) performance. Then, a joint threshold scheme for the GSCAN-BF-2 decoding algorithm is proposed to reduce the average decoding complexity by considering both the bit channel quality and the reliability of the coded bits. Simulation results show that the GSCAN-BF-Ω decoding algorithm reduces the average decoding latency while getting performance gains compared to the common multiple SCAN bit-flipping decoding algorithm. And the GSCAN-BF-2 decoding algorithm with the joint threshold reduces the average decoding latency further by approximately 50% with only a slight performance loss compared to the GSCAN-BF-2 decoding algorithm.

A Fast stream cipher Canon (고속 스트림 암호 Canon)

  • Kim, Gil-Ho
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.17 no.7
    • /
    • pp.71-79
    • /
    • 2012
  • Propose stream cipher Canon that need in Wireless sensor network construction that can secure confidentiality and integrity. Create Canon 128 bits streams key by 128 bits secret key and 128 bits IV, and makes 128 bits cipher text through whitening processing with produced streams key and 128 bits plaintext together. Canon for easy hardware implementation and software running fast algorithm consists only of simple logic operations. In particular, because it does not use S-boxes for non-linear operations, hardware implementation is very easy. Proposed stream cipher Canon shows fast speed test results performed better than AES, Salsa20, and gate number is small than Trivium. Canon purpose of the physical environment is very limited applications, mobile phones, wireless Internet environment, DRM (Digital Right Management), wireless sensor networks, RFID, and use software and hardware implementation easy 128 bits stream ciphers.

UEP Effect Analysis of LDPC Codes for High-Quality Communication Systems (고품질 통신 시스템을 위한 LDPC 부호의 UEP 성능 분석)

  • Yu, Seog Kun;Joo, Eon Kyeong
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.38A no.6
    • /
    • pp.471-478
    • /
    • 2013
  • Powerful error control and increase in the number of bits per symbol should be provided for future high-quality communication systems. Each message bit may have different importance in multimedia data. Hence, UEP(unequal error protection) may be more efficient than EEP(equal error protection) in such cases. And the LDPC(low-density parity-check) code shows near Shannon limit error correcting performance. Therefore, the effect of UEP with LDPC codes is analyzed for high-quality message data in this paper. The relationship among MSE(mean square error), BER(bit error rate) and the number of bits per symbol is analyzed theoretically. Then, total message bits in a symbol are classified into two groups according to importance to prove the relationship by simulation. And the UEP performance is obtained by simulation according to the number of message bits in each group with the constraint of a fixed total code rate and codeword length. As results, the effect of UEP with the LDPC codes is analyzed by MSE according to the number of bits per symbol, the ratio of the message bits, and protection level of the classified groups.