• Title/Summary/Keyword: 블록처리 시간

Search Result 434, Processing Time 0.027 seconds

A Hardware Design of Effective Intra Prediction Angular Mode Decision for HEVC Encoder (HEVC 부호기를 위한 효율적인 화면내 예측 Angular 모드 결정 하드웨어 설계)

  • Park, Seungyong;Choi, Juyong;Ryoo, Kwangki
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.4
    • /
    • pp.767-773
    • /
    • 2017
  • In this paper, we propose a design of Intra prediction angular mode decision for HEVC encoder. Intra prediction coding of HEVC is a method for predicting a current block by referring to samples reconstructed around a current block. Intra prediction supports a total of 35 modes with 1 DC mode, 1 Planar mode, and 33 Angular modes. Intra prediction coding of HEVC works by performing all 35 modes for efficient encoding. However, in order to process all of the 35 modes, the computational complexity and operational time required are high. Therefore, this paper proposes comparing the difference in the value of the original pixel, using an algorithm that determines angular mode efficiently. This new algorithm reduces the Hardware size. The hardware which is proposed was designed using Verilog HDL and was implemented in 65nm technology. Its gate count is 14.9K and operating speed is 2GHz.

Multiresolution Watermarking Scheme on DC Image in DCT Compressed Domain (DCT 압축영역에서의 DC 영상 기반 다해상도 워터마킹 기법)

  • Kim, Jung-Youn;Nam, Je-Ho
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.4
    • /
    • pp.1-9
    • /
    • 2008
  • This paper presents a rapid watermarking algorithm based on DC image, which provides a resilience to geometric distortion. Our proposed scheme is based on $8{\times}8$ block DCT that is widely used in image/video compression techniques (e.g., JPEG and MPEG). In particular, a DC image is analyzed by DWT to embed a watermark. To overcome a quality degradation caused by a watermark insertion into DC components, we discern carefully the intensity and amount of watermark along the different subbands of DWT. Note that the proposed technique supports a high throughput for a real-time watermark insertion and extraction by relying on a partial decoding (i.e., DC components) on $8{\times}8$ block DCT domain. Experimental result shows that the proposed watermarking scheme significantly reduces computation time of 82% compared with existing DC component based algorithm and yet provides invariant properties against various attacks such as geometric distortion and JPEG compression, etc.

A Study on the Abrupt Scene Change Detection Using the Features of B frame in the MPEG Sequence (MPEG에서 B 프레임의 특징을 이용한 급진적 장면전환 검출에 관한 연구)

  • Kim Joong-Heon;Jang Jong-Whan
    • The KIPS Transactions:PartB
    • /
    • v.12B no.5 s.101
    • /
    • pp.617-630
    • /
    • 2005
  • General scene change detection determines the changes of a scene by using feature comparison of two continuous images that are above the fixed threshold. But existing algerian detects scene change that was used in comparing the features of two images continuously, it usually takes a lot of time in decrypting the image data and false-detection problem occurs when there is an object motion or a change of illumination. In this paper, macroblock were used to extract the information directly from the MPEG compression area and suggests algorithm that will detect scene changes more effectively. Existing algorithm have shown numerous arithmetic problems that were improved in the proposed algorithm. The existing algorithm cannot detect the changes of a scene after analyzing the relationship of the previousand futureimages while the algorithm being proposed can detect the changes of a scene continuously and resolves the problem of false-detection. To this end, the data used in general were tested to prove that this algerian would be able to detect the scene changes faster and more correctly than the existing ones. The performance of the suggested algorithm was analyzed basedontheresultsoftheexperiment. .

A Study on the WBI System Design & Implemented based on the Component (컴포넌트기반의 웹 기반 교육시스템 설계에 관한 연구)

  • Jeon, Ju-Hyeon;Hong, Chan-Gi
    • The KIPS Transactions:PartD
    • /
    • v.8D no.6
    • /
    • pp.673-680
    • /
    • 2001
  • When the developers develop the software, the cost and time of the software development can be reduced by using blocks that are implemented previously. We call these implemented blocks components. In the early stage of Web-based Instruction, it didn't gain preference in spite of it's benefit of convenience. The main reason is, I think, the lack of generality at the education system which eventually results in unsatisfactory facilities compared with the requirement of teachers and students. And the early systems don't make good use of the plenty data in distributed environment, and don't show so good reliablity due to lack of systematic design and development. In this paper, we suggest WBI developing technology using the concept of WBSE. WBI developing is consist of component of pre-developed education software, integration of component using its reusability, and production of more requirement-satisfactory education software.

  • PDF

The Fractal Image Compression Based on the Wavelet Transform Using the SAS Techniques (SAS 기법을 이용한 웨이브릿 변환 기반 프랙탈 영상 압축)

  • 정태일;강경원;문광석;권기룡;류권열
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.2 no.1
    • /
    • pp.19-27
    • /
    • 2001
  • The conventional fractal image compression based on wavelet transform has the disadvantage that the encoding takes many time, since it finds the optimum domain for all the range blocks. In this paper, we propose the fractal image compression based on wavelet transform using the SAS(Self Affine System) techniques. It consists of the range and domain blocks in the wavelet transform, and the range blocks select the domain which is located the relatively same position. In the encoding process, the proposed methods introduce SAS techniques that the searching process of the domains blocks is not required. Therefore, it can perform a fast encoding by reducing the computational complexity. And, the image quality is improved using the different scale factors for each level and the sub-tree in the decoding. As a result, the image quality and the compression ratio are adjustable by the scale factors.

  • PDF

A Study On Distributed Remote Lecture Contents for QoS Guarantee Streaming Service (QoS보장형 스트리밍 서비스를 위한 분산 원격강의 컨텐츠에 대한 연구)

  • Choi, Yong-jun;Ku, Ja-hyo;Leem, In-taek;Choi, Byung-do;Kim, Chong-gun
    • The KIPS Transactions:PartA
    • /
    • v.9A no.4
    • /
    • pp.603-614
    • /
    • 2002
  • Delivery efficiency of e-learning media can be influenced by authoring processes. Generally, a moving picture recorded by video camera can be delivered to student by multimedia streaming service, using media server technology. A e-learning media authored by lecture authoring tool is played in a student application by download-based delivery system. Recently, some animation know-how are applied to author e-learning media by hand-operation. In this paper, we suggest a client-based streaming service for the e-leaning media consists of media files and integration data The lecture of e-learning media nay be divided into some time-based small blocks. Each blocks can be located distributed site. The student system gather those blocks by download-scheduling. This is a valid method for QoS guarantee streaming services. In addition to our study, lecturers can author composite e-learning media includes media files and dynamic web pages simply, The distributed e-learning media files of our study is managed by multi-author and updated rapidly.

PRMS: Page Reallocation Method for SSDs (PRMS: SSDs에서의 Page 재배치 방법)

  • Lee, Dong-Hyun;Roh, Hong-Chan;Park, Sang-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.17D no.6
    • /
    • pp.395-404
    • /
    • 2010
  • Solid-State Disks (SSDs) have been currently considered as a promising candidate to replace hard disks, due to their significantly short access time, low power consumption, and shock resistance. SSDs, however, have drawbacks such that their write throughput and life span are decreased by random-writes, nearly regardless of SSDs controller designs. Previous studies have mostly focused on better designs of SSDs controller and reducing the number of write operations to SSDs. We suggest another method that reallocates data pages that tend to be simultaneously written to contiguous blocks. Our method gathers write operations during a period of time and generates write traces. After transforming each trace to a set of transactions, our method mines frequent itemsets from the transactions and reallocates the pages of the frequent itemsets. In addition, we introduce an algorithm that reallocates the pages of the frequent itemsets with moderate time complexity. Experiments using TPC-C workload demonstrated that our method successfully reduce 6% of total logical block access.

A Study for Improving Performance of ATM Multicast Switch (ATM 멀티캐스트 스위치의 성능 향상을 위한 연구)

  • 이일영;조양현;오영환
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.12A
    • /
    • pp.1922-1931
    • /
    • 1999
  • A multicast traffic’s feature is the function of providing a point to multipoints cell transmission, which is emerging from the main function of ATM switch. However, when a conventional point-to-point switch executes a multicast function, the excess load is occurred because unicast cell as well as multicast cell passed the copy network. Additionally, due to the excess load, multicast cells collide with other cells in a switch. Thus a deadlock that losses cells raises, extremely diminishes the performance of switch. An input queued switch also has a defect of the HOL (Head of Line) blocking that less lessens the performance of the switch. In the proposed multicast switch, we use shared memory switch to reduce HOL blocking and deadlock. In order to decrease switch’s complexity and cell's processing time, to improve a throughput, we utilize the method that routes a cell on a separated paths by traffic pattern and the scheduling algorithm that processes a maximum 2N cell at once in the control part. Besides, when cells is congested at an output port, a cell loss probability increases. Thus we use the Output Memory (OM) to reduce the cell loss probability. And we make use of the method that stores the assigned memory (UM, MM) with a cell by a traffic pattern and clears the cell of the Output memory after a fixed saving time to improve the memory utilization rate. The performance of the proposed switch is executed and compared with the conventional policy under the burst traffic condition through both the analysis based on Markov chain and simulation.

  • PDF

A Study on Performance Improvement and Development of Integrity Verification Software of TCP/IP output data of VCS Correlation Block (VCS 상관블록의 TCP/IP 출력데이터의 무결성 검사 소프트웨어의 개발과 성능개선에 관한 연구)

  • Yeom, Jae-Hwan;Roh, Duk-Gyoo;Oh, Chung-Sik;Jung, Jin-Seung;Chung, Dong-Kyu;Oh, Se-Jin
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.13 no.4
    • /
    • pp.211-219
    • /
    • 2012
  • In this paper, we described the software development for verifying the integrity of output data of TCP/IP for VLBI Correlation Subsystem (VCS) correlation block and proposed the performance improvement method in order to prevent the data loss of correlation output. The VCS correlation results are saved at the Data Archive system through TCP/IP packet transmission. In this paper, the integrity verification software is developed so as to confirm the integrity of correlation result saved at the data archive system using TCP/IP packet information of VCS. The 3-step integrity verification process is proposed by using the developed software, its effectiveness was confirmed in consequence of correlation experiments. In addition, TCP/IP packet transmission must be completed within minimum integration period. However, there is not only TCP/IP packet loss occurred but also the problem of correlation result integrity affected in account of a large quantity of packets and data during short integration time. In this paper, the reason of TCP/IP packet loss is analyzed and the modified methods for FPGA(Field Programmable Gate Array) of VCS are proposed, the integrity problem of correlation results will be solved.

The Efficient Merge Operation in Log Buffer-Based Flash Translation Layer for Enhanced Random Writing (임의쓰기 성능향상을 위한 로그블록 기반 FTL의 효율적인 합병연산)

  • Lee, Jun-Hyuk;Roh, Hong-Chan;Park, Sang-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.19D no.2
    • /
    • pp.161-186
    • /
    • 2012
  • Recently, the flash memory consistently increases the storage capacity while the price of the memory is being cheap. This makes the mass storage SSD(Solid State Drive) popular. The flash memory, however, has a lot of defects. In order that these defects should be complimented, it is needed to use the FTL(Flash Translation Layer) as a special layer. To operate restrictions of the hardware efficiently, the FTL that is essential to work plays a role of transferring from the logical sector number of file systems to the physical sector number of the flash memory. Especially, the poor performance is attributed to Erase-Before-Write among the flash memory's restrictions, and even if there are lots of studies based on the log block, a few problems still exists in order for the mass storage flash memory to be operated. If the FAST based on Log Block-Based Flash often is generated in the wide locality causing the random writing, the merge operation will be occur as the sectors is not used in the data block. In other words, the block thrashing which is not effective occurs and then, the flash memory's performance get worse. If the log-block makes the overwriting caused, the log-block is executed like a cache and this technique contributes to developing the flash memory performance improvement. This study for the improvement of the random writing demonstrates that the log block is operated like not only the cache but also the entire flash memory so that the merge operation and the erase operation are diminished as there are a distinct mapping table called as the offset mapping table for the operation. The new FTL is to be defined as the XAST(extensively-Associative Sector Translation). The XAST manages the offset mapping table with efficiency based on the spatial locality and temporal locality.