• Title/Summary/Keyword: 관계형 메모리

Search Result 67, Processing Time 0.019 seconds

Detection of Complex Event Patterns over Interval-based Events (기간기반 복합 이벤트 패턴 검출)

  • Kang, Man-Mo;Park, Sang-Mu;Kim, Sank-Rak;Kim, Kang-Hyun;Lee, Dong-Hyeong
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.12 no.4
    • /
    • pp.201-209
    • /
    • 2012
  • The point-based complex event processing handled an instantaneous event by using one time stamp in each event. However, the activity period of the event plays the important role in the field which is the same as the finance, multimedia, medicine, and meteorology. The point-based event is insufficient for expressing the complex temporal relationship in this field. In the application field of the real-time world, the event has the period. The events more than two kinds can be temporally overlapped. In addition, one event can include the other event. The relation about the events of kind of these can not be successive like the point-based event. This thesis designs and implements the method detecting the patterns of the complex event by using the interval-based events. The interval-based events can express the overlapping relation between events. Furthermore, it can include the others. By using the end point of beginning and end point of the termination, the operator of interval-based events shows the interval-based events. It expresses the sequence of the interval-based events and can detect the complex event patterns. This thesis proposes the algorithm using the active instance stack in order to raise efficiency of detection of the complex event patterns. When comprising the event sequence, this thesis applies the window push down technique in order to reduce the number of intermediate results. It raises the utility factor of the running time and memory.

Studies on the Improvement of the Productivity of the Purse Seine Fishery - The characteristics on the motion of purse seine in the experimental operation of one boat system - (선망어업의 생산성 향상에 관한 연구 - 단선식 시험조업에 있어서 선망의 운동특성 -)

  • 김석종;최찬문;정용진
    • Journal of the Korean Society of Fisheries and Ocean Technology
    • /
    • v.39 no.2
    • /
    • pp.99-111
    • /
    • 2003
  • The fundamental studies on the productivity improvement of the purse seine fishery are presented in this paper. The experiment on net shooting and hauling was carried out in the near sea of Jeju Island (33$^{\circ}$37.8' N, 126$^{\circ}$31.1' E) by using the Cheju national university training ship (A-Ra, 990tons) which was constructed for the one boat system operation of purse seine. The corkline and leadline of the purse seine used for the experiment were 829.1m and 995.7m in length, respectively. Micro data recorder system, net sonde, and tensiometer were used to measure the depth of leadline and the tension of purse seine. Based on the measurement data, the motion and tension of purse seine at the time of shooting, hauling, and pursing were characterized. The experimental results are summarized as follows ; 1. The shooting and hauling of net were found to be possible in the one boat system experimental operation. 2. At the time of purse seine shooting, the relationship between the depth (Dp) of leadline and elapsed time (Et) was found to be Dp=7.58Et-6.48. 3. At the time of pursing, the relationship between the depth (Dp) of leadline and elapsed time (Et) was found to be Dp=-0.8Et$^2$+7.42Et+92.04. 4. At the time of pursing, the tension (metric tons) of purse seine attained its maximum value (14.7tons) when the elapsed time is 8 minutes. The relationship with the elapsed time was found to be T=-0.13Et$^2$+3.23Et-5.72.

Design and Implementation of Real-Time Operating System for a GPS Navigation Computer (GPS 항법 컴퓨터를 위한 실시간 운영체제의 설계 및 구현)

  • Bae, Jang-Sik;Song, Dae-Gi;Lee, Cheol-Hun;Song, Ho-Jun
    • The KIPS Transactions:PartA
    • /
    • v.8A no.4
    • /
    • pp.429-438
    • /
    • 2001
  • GPS (Global Positioning System) is the most ideal navigation system which can be used on the earth irrespective of time and weather conditions. GPS has been used for various applications such as construction, survey, environment, communication, intelligent vehicles and airplanes and the needs of GPS are increasing in these days. This paper deals with the design and implementation of the RTOS (Real-Time Operating System) for a GPS navigation computer in the GPS/INS integrated navigation system. The RTOS provides the optimal environment for execution and the base platform to develop GPS application programs. The key facilities supplied by the RTOS developed in this paper are priority-based preemptive scheduling policy, dynamic memory management, intelligent interrupt handling, timers and IPC, etc. We also verify the correct operations of all application tasks of the GPS navigation computer on the RTOS and evaluate the performance by measuring the overhead of using the RTOS services.

  • PDF

Discrete-Time Analysis of Throughput and Response Time for LAP Derivative Protocols under Markovian Block-Error Pattern (마르코프 오류모델 하에서의 LAP 계열 프로토콜들의 전송성능과 반응시간에 대한 이산-시간 해석)

  • Cho, Young-Jong;Choi, Dug-Kyoo
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.11
    • /
    • pp.2786-2800
    • /
    • 1997
  • In this paper, we investigate how well the channel memory (statistical dependence in the occurrence of transmission errors) can be used in the evaluation of widely used error control schemes. For this we assume a special case named as the simplest Markovian block-error pattern with two states, in which each block is classified into two classes of whether the block transmission is in error or not. We apply the derived pattern to the performance evaluation of the practical link-level procedures, LAPB/D/M with multi-reject options, and investigate both throughput and user-perceived response time behaviors on the discrete-time domain to determine how much the performance of error recovery action is improved under burst error condition. Through numerical examples, we show that the simplest Markovian block-error pattern tends to be superior in throughput and delay characteristics to the random error case. Also, instead of mean alone, we propose a new measure of the response time specified as mean plus two standard deviations 50 as to consider user-perceived worst cases, and show that it results in much greater sensitivity to parameter variations than does mean alone.

  • PDF

An Efficient Matrix Multiplier Available in Multi-Head Attention and Feed-Forward Network of Transformer Algorithms (트랜스포머 알고리즘의 멀티 헤드 어텐션과 피드포워드 네트워크에서 활용 가능한 효율적인 행렬 곱셈기)

  • Seok-Woo Chang;Dong-Sun Kim
    • Journal of IKEEE
    • /
    • v.28 no.1
    • /
    • pp.53-64
    • /
    • 2024
  • With the advancement of NLP(Natural Language Processing) models, conversational AI such as ChatGPT is becoming increasingly popular. To enhance processing speed and reduce power consumption, it is important to implement the Transformer algorithm, which forms the basis of the latest natural language processing models, in hardware. In particular, the multi-head attention and feed-forward network, which analyze the relationships between different words in a sentence through matrix multiplication, are the most computationally intensive core algorithms in the Transformer. In this paper, we propose a new variable systolic array based on the number of input words to enhance matrix multiplication speed. Quantization maintains Transformer accuracy, boosting memory efficiency and speed. For evaluation purposes, this paper verifies the clock cycles required in multi-head attention and feed-forward network and compares the performance with other multipliers.

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.

A Transmission Electron Microscopy Study on the Crystallization Behavior of In-Sb-Te Thin Films (In-Sb-Te 박막의 결정화 거동에 관한 투과전자현미경 연구)

  • Kim, Chung-Soo;Kim, Eun-Tae;Lee, Jeong-Yong;Kim, Yong-Tae
    • Applied Microscopy
    • /
    • v.38 no.4
    • /
    • pp.279-284
    • /
    • 2008
  • The phase change materials have been extensively used as an optical rewritable data storage media utilizing their phase change properties. Recently, the phase change materials have been spotlighted for the application of non-volatile memory device, such as the phase change random access memory. In this work, we have investigated the crystallization behavior and microstructure analysis of In-Sb-Te (IST) thin films deposited by RF magnetron sputtering. Transmission electron microscopy measurement was carried out after the annealing at $300^{\circ}C$, $350^{\circ}C$, $400^{\circ}C$ and $450^{\circ}C$ for 5 min. It was observed that InSb phases change into $In_3SbTe_2$ phases and InTe phases as the temperature increases. It was found that the thickness of thin films was decreased and the grain size was increased by the bright field transmission electron microscopy (BF TEM) images and the selected area electron diffraction (SAED) patterns. In a high resolution transmission electron microscopy (HRTEM) study, it shows that $350^{\circ}C$-annealed InSb phases have {111} facet because the surface energy of a {111} close-packed plane is the lowest in FCC crystals. When the film was heated up to $400^{\circ}C$, $In_3SbTe_2$ grains have coherent micro-twins with {111} mirror plane, and they are healed annealing at $450^{\circ}C$. From the HRTEM, InTe phase separation was occurred in this stage. It can be found that $In_3SbTe_2$ forms in the crystallization process as composition of the film near stoichiometric composition, while InTe phase separation may take place as the composition deviates from $In_3SbTe_2$.