• Title/Summary/Keyword: Fast Computation

Search Result 750, Processing Time 0.024 seconds

Two-stage Content-based Image Retrieval Using the Dimensionality Condensation of Feature Vector (특징벡터의 차원축약 기법을 이용한 2단계 내용기반 이미지검색 시스템)

  • 조정원;최병욱
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.7C
    • /
    • pp.719-725
    • /
    • 2003
  • The content-based image retrieval system extracts features of color, shape and texture from raw images, and builds the database with those features in the indexing process. The search in the whole retrieval system is defined as a process which finds images that have large similarity to query image using the feature database. This paper proposes a new two-stage search method in the content-based image retrieval system. The method is that the features are condensed and stored by the property of Cauchy-Schwartz inequality in order to reduce the similarity computation time which takes a mostly response time from entering a query to getting retrieval results. By the extensive computer simulations, we have observed that the proposed two-stage search method successfully reduces the similarity computation time while maintaining the same retrieval relevance as the conventional exhaustive search method. We also have observed that the method is more effective as the number of images and dimensions of the feature space increase.

A Study on the Fast Motion Estimation Coding by Moving Region Segmentation (동영역 분할에 의한 고속 움직임 추정 부호화에 관한 연구)

  • Lee, Bong-Ho;Choi, Kyung-Soo;Kwak, No-Youn;Hwang, Byong-Won
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.37 no.3
    • /
    • pp.88-97
    • /
    • 2000
  • This paper presents motion estimation method using region segmentation information Motion estimation which is very difficult to be implemented only by software because of intensive computation cost, is implemented by special-purpose hardware in real-time applications In this paper, we propose region based motion estimation algorithm which can reduce the computation cost by using region segmentation information and setting the variable search window compared with FSMA algorithm Secondly, another proposed algorithm is to segment semantic region like face for selective coding and transfer of semantic region using segmented region information This work alms to improving the subjective quality of skin color region or face region m the picture that has slow motion and IS mainly composed of one or two speakers of video conference and video telephony applications.

  • PDF

Efficiency Pixel Recomposition Algorithm for Fractional Motion Estimation (부화소 움직임 추정을 위한 효과적인 화소 재구성 알고리즘)

  • Shin, Wang-Ho;SunWoo, Myung-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.1
    • /
    • pp.64-70
    • /
    • 2011
  • In an H.264/AVC video encoder, the motion estimation at fractional pixel accuracy improves a coding efficiency and image quality. However, it requires additional computation overheads for fractional search and interpolation, and thus, reducing the computation complexity of fractional search becomes more important. This paper proposes a Pixel Re-composition Fractional Motion Estimation (PRFME) algorithm for an H.264/AVC video encoder. Fractional Motion Estimation performs interpolation for the overlapped pixels which increases the computational complexity. PRFME can reduce the computational complexity by eliminating the overlapped pixel interpolation. Compared with the fast full search, the proposed algorithm can reduce 18.1% of computational complexity, meanwhile, the maximum PSNR degradation is less than 0.067dB. Therefore, the proposed PRFME algorithm is quite suitable for mobile applications requiring low power and complexity.

Enhanced NOW-Sort on a PC Cluster with a Low-Speed Network (저속 네트웍 PC 클러스터상에서 NOW-Sort의 성능향상)

  • Kim, Ji-Hyoung;Kim, Dong-Seung
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.10
    • /
    • pp.550-560
    • /
    • 2002
  • External sort on cluster computers requires not only fast internal sorting computation but also careful scheduling of disk input and output and interprocessor communication through networks. This is because the overall time for the execution is determined by reflecting the times for all the jobs involved, and the portion for interprocessor communication and disk I/O operations is significant. In this paper, we improve the sorting performance (sorting throughput) on a cluster of PCs with a low-speed network by developing a new algorithm that enables even distribution of load among processors, and optimizes the disk read and write operations with other computation/communication activities during the sort. Experimental results support the effectiveness of the algorithm. We observe the algorithm reduces the sort time by 45% compared to the previous NOW-sort[1], and provides more scalability in the expansion of the computing nodes of the cluster as well.

Extracting optimal moving patterns of edge devices for efficient resource placement in an FEC environment (FEC 환경에서 효율적 자원 배치를 위한 엣지 디바이스의 최적 이동패턴 추출)

  • Lee, YonSik;Nam, KwangWoo;Jang, MinSeok
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.1
    • /
    • pp.162-169
    • /
    • 2022
  • In a dynamically changing time-varying network environment, the optimal moving pattern of edge devices can be applied to distributing computing resources to edge cloud servers or deploying new edge servers in the FEC(Fog/Edge Computing) environment. In addition, this can be used to build an environment capable of efficient computation offloading to alleviate latency problems, which are disadvantages of cloud computing. This paper proposes an algorithm to extract the optimal moving pattern by analyzing the moving path of multiple edge devices requiring application services in an arbitrary spatio-temporal environment based on frequency. A comparative experiment with A* and Dijkstra algorithms shows that the proposed algorithm uses a relatively fast execution time and less memory, and extracts a more accurate optimal path. Furthermore, it was deduced from the comparison result with the A* algorithm that applying weights (preference, congestion, etc.) simultaneously with frequency can increase path extraction accuracy.

Performance Analysis for Privacy-preserving Data Collection Protocols (개인정보보호를 위한 데이터 수집 프로토콜의 성능 분석)

  • Lee, Jongdeog;Jeong, Myoungin;Yoo, Jincheol
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.12
    • /
    • pp.1904-1913
    • /
    • 2021
  • With the proliferation of smart phones and the development of IoT technology, it has become possible to collect personal data for public purposes. However, users are afraid of voluntarily providing their private data due to privacy issues. To remedy this problem, mainly three techniques have been studied: data disturbance, traditional encryption, and homomorphic encryption. In this work, we perform simulations to compare them in terms of accuracy, message length, and computation delay. Experiment results show that the data disturbance method is fast and inaccurate while the traditional encryption method is accurate and slow. Similar to traditional encryption algorithms, the homomorphic encryption algorithm is relatively effective in privacy preserving because it allows computing encrypted data without decryption, but it requires high computation costs as well. However, its main cost, arithmetic operations, can be processed in parallel. Also, data analysis using the homomorphic encryption needs to do decryption only once at any number of data.

Multiple Binarization Quadtree Framework for Optimizing Deep Learning-Based Smoke Synthesis Method

  • Kim, Jong-Hyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.4
    • /
    • pp.47-53
    • /
    • 2021
  • In this paper, we propose a quadtree-based optimization technique that enables fast Super-resolution(SR) computation by efficiently classifying and dividing physics-based simulation data required to calculate SR. The proposed method reduces the time required for quadtree computation by downscaling the smoke simulation data used as input data. By binarizing the density of the smoke in this process, a quadtree is constructed while mitigating the problem of numerical loss of density in the downscaling process. The data used for training is the COCO 2017 Dataset, and the artificial neural network uses a VGG19-based network. In order to prevent data loss when passing through the convolutional layer, similar to the residual method, the output value of the previous layer is added and learned. In the case of smoke, the proposed method achieved a speed improvement of about 15 to 18 times compared to the previous approach.

High-Speed Maritime Object Detection Scheme for the Protection of the Aid to Navigation

  • Lee, Hyochan;Song, Hyunhak;Cho, Sungyoon;Kwon, Kiwon;Park, Sunghyun;Im, Taeho
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.692-712
    • /
    • 2022
  • Buoys used for Aid to Navigation systems are widely used to guide the sea paths and are powered by batteries, requiring continuous battery replacement. However, since human labor is required to replace the batteries, humans can be exposed to dangerous situation, including even collision with shipping vessels. In addition, Maritime sensors are installed on the route signs, so that these are often damaged by collisions with small and medium-sized ships, resulting in significant financial loss. In order to prevent these accidents, maritime object detection technology is essential to alert ships approaching buoys. Existing studies apply a number of filters to eliminate noise and to detect objects within the sea image. For this process, most studies directly access the pixels and process the images. However, this approach typically takes a long time to process because of its complexity and the requirements of significant amounts of computational power. In an emergent situation, it is important to alarm the vessel's rapid approach to buoys in real time to avoid collisions between vessels and route signs, therefore minimizing computation and speeding up processes are critical operations. Therefore, we propose Fast Connected Component Labeling (FCCL) which can reduce computation to minimize the processing time of filter applications, while maintaining the detection performance of existing methods. The results show that the detection performance of the FCCL is close to 30 FPS - approximately 2-5 times faster, when compared to the existing methods - while the average throughput is the same as existing methods.

Fast PU Decision Method Using Coding Information of Co-Located Sub-CU in Upper Depth for HEVC (상위깊이의 Sub-CU 부호화 정보를 이용한 HEVC의 고속 PU 결정 기법)

  • Jang, Jae-Kyu;Choi, Ho-Youl;Kim, Jae-Gon
    • Journal of Broadcast Engineering
    • /
    • v.20 no.2
    • /
    • pp.340-347
    • /
    • 2015
  • HEVC (High Efficiency Video Coding) achieves high coding efficiency by employing a quadtree-based coding unit (CU) block partitioning structure and various prediction units (PUs), and the determination of the best CU partition structure and the best PU mode based on rate-distortion (R-D) cost. However, the computation complexity of encoding also dramatically increases. In this paper, to reduce such encoding computational complexity, we propose three fast PU mode decision methods based on encoding information of upper depth as follows. In the first method, the search of PU mode of the current CU is early terminated based on the sub-CBF (Coded Block Flag) of upper depth. In the second method, the search of intra prediction modes of PU in the current CU is skipped based on the sub-Intra R-D cost of upper depth. In the last method, the search of intra prediction modes of PU in the lower depth's CUs is skipped based on the sub-CBF of the current depth's CU. Experimental results show that the three proposed methods reduce the computational complexity of HM 14.0 to 31.4%, 2.5%, and 23.4% with BD-rate increase of 1.2%, 0.11%, and 0.9%, respectively. The three methods can be applied in a combined way to be applied to both of inter prediction and intra prediction, which results in the complexity reduction of 34.2% with 1.9% BD-rate increase.

Hexagon-shape Line Search Algorithm for Fast Motion Estimation on Media Processor (미디어프로세서 상의 고속 움직임 탐색을 위한 Hexagon 모양 라인 탐색 알고리즘)

  • Jung Bong-Soo;Jeon Byeung-Woo
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.4 s.310
    • /
    • pp.55-65
    • /
    • 2006
  • Most of fast block motion estimation algorithms reported so far in literatures aim to reduce the computation in terms of the number of search points, thus do not fit well with multimedia processors due to their irregular data flow. For multimedia processors, proper reuse of data is more important than reducing number of absolute difference operations because the execution cycle performance strongly depends on the number of off-chip memory access. Therefore, in this paper, we propose a Hexagon-shape line search (HEXSLS) algorithm using line search pattern which can increase data reuse from on-chip local buffer, and check sub-sampling points in line search pattern to reduce unnecessary SAD operation. Our experimental results show that the prediction error (MAE) performance of the proposed HEXSLS is similar to that of the full search block matching algorithm (FSBMA), while compared with the hexagon-based search (HEXBS), the HEXSLS outperforms. Also the proposed HEXSLS requires much lesser off-chip memory access than the conventional fast motion estimation algorithm such as the hexagon-based search (HEXBS) and the predictive line search (PLS). As a result, the proposed HEXSLS algorithm requires smaller number of execution cycles on media processor.