• Title/Summary/Keyword: in-memory computing

Search Result 766, Processing Time 0.03 seconds

Random Partial Haar Wavelet Transformation for Single Instruction Multiple Threads (단일 명령 다중 스레드 병렬 플랫폼을 위한 무작위 부분적 Haar 웨이블릿 변환)

  • Park, Taejung
    • Journal of Digital Contents Society
    • /
    • v.16 no.5
    • /
    • pp.805-813
    • /
    • 2015
  • Many researchers expect the compressive sensing and sparse recovery problem can overcome the limitation of conventional digital techniques. However, these new approaches require to solve the l1 norm optimization problems when it comes to signal reconstruction. In the signal reconstruction process, the transform computation by multiplication of a random matrix and a vector consumes considerable computing power. To address this issue, parallel processing is applied to the optimization problems. In particular, due to huge size of original signal, it is hard to store the random matrix directly in memory, which makes one need to design a procedural approach in handling the random matrix. This paper presents a new parallel algorithm to calculate random partial Haar wavelet transform based on Single Instruction Multiple Threads (SIMT) platform.

Modified Kernel PCA Applied To Classification Problem (수정된 커널 주성분 분석 기법의 분류 문제에의 적용)

  • Kim, Byung-Joo;Sim, Joo-Yong;Hwang, Chang-Ha;Kim, Il-Kon
    • The KIPS Transactions:PartB
    • /
    • v.10B no.3
    • /
    • pp.243-248
    • /
    • 2003
  • An incremental kernel principal component analysis (IKPCA) is proposed for the nonlinear feature extraction from the data. The problem of batch kernel principal component analysis (KPCA) is that the computation becomes prohibitive when the data set is large. Another problem is that, in order to update the eigenvectors with another data, the whole eigenspace should be recomputed. IKPCA overcomes these problems by incrementally computing eigenspace model and empirical kernel map The IKPCA is more efficient in memory requirement than a batch KPCA and can be easily improved by re-learning the data. In our experiments we show that IKPCA is comparable in performance to a batch KPCA for the feature extraction and classification problem on nonlinear data set.

3차 저장 장치의 장착을 위한 MIDAS-II의 확장

  • Kim, Yeong-Seong;Gang, Hyeon-Cheol;Kim, Jun
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.6 no.1
    • /
    • pp.21-35
    • /
    • 2000
  • MIDAS-II is the storage system for BADA DBMS developed at ETRI. This paper describes the extension of MIDAS-II for incorporating the tertiary storage device such as an optical disk jukebox or a tape library, enabling MIDAS-II to function as a storage system of the data server that stores a massive amount of multimedia data. The MIDAS-II disk volume structure is extended to efficiently function as a volume for the tertiary storage device with multiple platters, which canstore huge amount of data of the order of tera bytes. The storage structure of the LOB is changed to efficiently manage the LOB data in the tertiary storage device. The data structures of the shared memory, the process structure, and the utilities in MIDAS-II are also extended to efficiently incorporating the tertiary storage device. The functionalities of each MIDAS-II API function are expanded to handle the tertiary storage device, while the prototypes of those functions are intact in order not to affect the existing application programs. The performance evaluation shows that the extended MIDAS-II works effectively with the tertiary storage device. All these extensions and the performance evaluation are conducted in the SunOS 5.4 environment.

  • PDF

A Scalable Change Detection Technique for RDF Data using a Backward-chaining Inference based on Relational Databases (관계형 데이터베이스 기반의 후방향 추론을 이용하는 확장 가능한 RDF 데이타 변경 탐지 기법)

  • Im, Dong-Hyuk;Lee, Sang-Won;Kim, Hyoung-Joo
    • Journal of KIISE:Databases
    • /
    • v.37 no.4
    • /
    • pp.197-202
    • /
    • 2010
  • Recent studies on change detection for RDF data are focused on not only the structural difference but also the semantic-aware difference by computing the closure of RDF models. However, since these techniques which take into account the semantics of RDF model require both RDF models to be memory resident, or they use a forward-chaining strategy which computes the entire closure in advance, it is not efficient to apply them directly to detect changes in large RDF data. In this paper, we propose a scalable change detection technique for RDF data, which uses a backward-chaining inference based on relational database. Proposed method uses a new approach for RDF reasoning that computes only the relevant part of the closure for change detection in a relational database. We show that our method clearly outperforms the previous works through experiment using the real RDF from the bioinformatics domain.

A Survey of Genetic Programming and Its Applications

  • Ahvanooey, Milad Taleby;Li, Qianmu;Wu, Ming;Wang, Shuo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.4
    • /
    • pp.1765-1794
    • /
    • 2019
  • Genetic Programming (GP) is an intelligence technique whereby computer programs are encoded as a set of genes which are evolved utilizing a Genetic Algorithm (GA). In other words, the GP employs novel optimization techniques to modify computer programs; imitating the way humans develop programs by progressively re-writing them for solving problems automatically. Trial programs are frequently altered in the search for obtaining superior solutions due to the base is GA. These are evolutionary search techniques inspired by biological evolution such as mutation, reproduction, natural selection, recombination, and survival of the fittest. The power of GAs is being represented by an advancing range of applications; vector processing, quantum computing, VLSI circuit layout, and so on. But one of the most significant uses of GAs is the automatic generation of programs. Technically, the GP solves problems automatically without having to tell the computer specifically how to process it. To meet this requirement, the GP utilizes GAs to a "population" of trial programs, traditionally encoded in memory as tree-structures. Trial programs are estimated using a "fitness function" and the suited solutions picked for re-evaluation and modification such that this sequence is replicated until a "correct" program is generated. GP has represented its power by modifying a simple program for categorizing news stories, executing optical character recognition, medical signal filters, and for target identification, etc. This paper reviews existing literature regarding the GPs and their applications in different scientific fields and aims to provide an easy understanding of various types of GPs for beginners.

Similarity Measure based on Utilization of Rating Distributions for Data Sparsity Problem in Collaborative Filtering

  • Lee, Soojung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.12
    • /
    • pp.203-210
    • /
    • 2020
  • Memory-based collaborative filtering is one of the representative types of the recommender system, but it suffers from the inherent problem of data sparsity. Although many works have been devoted to solving this problem, there is still a request for more systematic approaches to the problem. This study exploits distribution of user ratings given to items for computing similarity. All user ratings are utilized in the proposed method, compared to previous ones which use ratings for only common items between users. Moreover, for similarity computation, it takes a global view of ratings for items by reflecting other users' ratings for that item. Performance is evaluated through experiments and compared to that of other relevant methods. The results reveal that the proposed demonstrates superior performance in prediction and rank accuracies. This improvement in prediction accuracy is as high as 2.6 times more than that achieved by the state-of-the-art method over the traditional similarity measures.

New Simple Power Analysis on scalar multiplication based on sABS recoding (sABS 형태의 스칼라 곱셈 연산에 대한 새로운 단순전력 공격)

  • Kim, Hee-Seok;Kim, Sung-Kyoung;Kim, Tae-Hyun;Park, Young-Ho;Lim, Jong-In;Han, Dong-Guk
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.17 no.2
    • /
    • pp.115-123
    • /
    • 2007
  • In cryptographic devices like a smart-card whose computing ability and memory are limited, cryptographic algorithms should be performed efficiently. Scalar multiplication is very important operation in Elliptic Curve Cryptosystems, and so must be constructed in safety against side channel attack(SCA). But several countermeasures proposed against SCA are exposed weaknesses by new un-dreamed analysis. 'Double-and-add always scalar multiplication' algorithm adding dummy operation being known to secure against SPA is exposed weakness by Doubling Attack. But Doubling Attack cannot apply to sABS receding proposed by Hedabou, that is another countermeasure against SPA. Our paper proposes new strengthened Doubling Attacks that can break sABS receding SPA-countermeasure and a detailed method of our attacks through experimental result.

Analyses of Security Issues and Vulnerability for Healthcare System For Under Internet of Things (사물인터넷과 융합한 헬스케어 시스템에서의 보안 이슈 및 취약점 분석)

  • Jung Tae Kim
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.699-706
    • /
    • 2023
  • Recently, the 4 generation industry revolution is developed with advanced and combined with a variety of new technologies. Conventional healthcare system is applied with IoT application. It provides many advantages with mobility and swift data transfers to patient and doctor. In despite of these kinds of advantages, it occurred security issues between basic devices and protocols in their applications. Especially, internet of things have restricted and limited resources such as small memory capacity, low capability of computing power, etc. Therefore, we can not utilize conventional mechanism. In this paper, we analyzed attacks and vulnerability in terms of security issues. To analyze security structure, features, demands and requirements, we solve the methods to be reduced security issues.

Performance evaluation of approximate frequent pattern mining based on probabilistic technique (확률 기법에 기반한 근접 빈발 패턴 마이닝 기법의 성능평가)

  • Pyun, Gwangbum;Yun, Unil
    • Journal of Internet Computing and Services
    • /
    • v.14 no.1
    • /
    • pp.63-69
    • /
    • 2013
  • Approximate Frequent pattern mining is to find approximate patterns, not exact frequent patterns with tolerable variations for more efficiency. As the size of database increases, much faster mining techniques are needed to deal with huge databases. Moreover, it is more difficult to discover exact results of mining patterns due to inherent noise or data diversity. In these cases, by mining approximate frequent patterns, more efficient mining can be performed in terms of runtime, memory usage and scalability. In this paper, we study the characteristics of an approximate mining algorithm based on probabilistic technique and run performance evaluation of the efficient approximate frequent pattern mining algorithm. Finally, we analyze the test results for more improvement.

Smart Card User Identification Using Low-sized Face Feature Information (경량화된 얼굴 특징 정보를 이용한 스마트 카드 사용자 인증)

  • Park, Jian;Cho, Seongwon;Chung, Sun-Tae
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.24 no.4
    • /
    • pp.349-354
    • /
    • 2014
  • PIN(Personal Identification Number)-based identification method has been used to identify the user of smart cards. However, this type of identification method has several problems. Firstly, PIN can be forgotten by owners of the card. Secondly, PIN can be used by others illegally. Furthermore, the possibility of hacking PIN can be high because this PIN type matching process is performed on terminal. Thus, in this paper we suggest a new identification method which is performed on smart card using face feature information. The proposed identification method uses low-sized face feature vectors and simple matching algorithm in order to get around the limits in computing capability and memory size of smart card.