• Title/Summary/Keyword: Not only SQL

Search Result 51, Processing Time 0.024 seconds

SSQUSAR : A Large-Scale Qualitative Spatial Reasoner Using Apache Spark SQL (SSQUSAR : Apache Spark SQL을 이용한 대용량 정성 공간 추론기)

  • Kim, Jonghoon;Kim, Incheol
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.6 no.2
    • /
    • pp.103-116
    • /
    • 2017
  • In this paper, we present the design and implementation of a large-scale qualitative spatial reasoner, which can derive new qualitative spatial knowledge representing both topological and directional relationships between two arbitrary spatial objects in efficient way using Aparch Spark SQL. Apache Spark SQL is well known as a distributed parallel programming environment which provides both efficient join operations and query processing functions over a variety of data in Hadoop cluster computer systems. In our spatial reasoner, the overall reasoning process is divided into 6 jobs such as knowledge encoding, inverse reasoning, equal reasoning, transitive reasoning, relation refining, knowledge decoding, and then the execution order over the reasoning jobs is determined in consideration of both logical causal relationships and computational efficiency. The knowledge encoding job reduces the size of knowledge base to reason over by transforming the input knowledge of XML/RDF form into one of more precise form. Repeat of the transitive reasoning job and the relation refining job usually consumes most of computational time and storage for the overall reasoning process. In order to improve the jobs, our reasoner finds out the minimal disjunctive relations for qualitative spatial reasoning, and then, based upon them, it not only reduces the composition table to be used for the transitive reasoning job, but also optimizes the relation refining job. Through experiments using a large-scale benchmarking spatial knowledge base, the proposed reasoner showed high performance and scalability.

Normalization of XQuery Queries for Efficient XML Query Processing (효율적인 XML질의 처리를 위한 XQuery 질의의 정규화)

  • 김서영;이기훈;황규영
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.10 no.5
    • /
    • pp.419-433
    • /
    • 2004
  • As XML becomes a standard for data representation, integration, and exchange on the Web, several XML query languages have been proposed. World Wide Web Consortium(W3C) has proposed XQuery as a standard for the XML query language. Like SQL, XQuery allows nested queries. Thus, normalization rules have been proposed to transform nested XQuery queries to semantically equivalent ones that could be executed more efficiently. However, previous normalization rules are applicable only to restricted forms of nested XQuery queries. Specifically, they can not handle FLWR expressions having nested expressions in the where clause. In this paper, we propose normalization rules for XQuery queries by extending those for SQL queries. Our proposed rules can handle FLWR expressions haying nested expressions in every clause. The major contributions of this paper are as follows. First, we classily nesting types of XQuery queries according to the existence of correlation and aggregation. We then propose normalization rules for each nesting type. Second, we propose detailed algorithms that apply the normalization rules to nested XQuery queries.

Implementation of Query Processor for Efficient Vehicle Monitoring and Control in e-Logistics (e-로지스틱스에서 효율적인 차량관제를 위한 질의 처리기 구현)

  • Kim, Dong-Ho;Kim, Jin-Suk;Ryu, Keun-Ho
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.7 no.3
    • /
    • pp.35-47
    • /
    • 2004
  • Telematics and LBS is one of rapidly emerged technology domains. In order to efficiently construct them, moving object technology which manages huge volume of real-time location data is required. Especially, the query which obtains special sorts of information closely related to the detailed applications is required in order to effectively retrieve and analyze the location data for moving object in logistics domain. It has also complex query structure comparing to the conventional database query. The approach using the standard database query language, like SQL, can be considered as an effective alternative choice. In this paper, we not only propose a new query language, entitled as MOQL based on SQL, for the query processing of the vehicle monitoring and control in e-Logistics but also design and implement the query processor.

  • PDF

Development of the Performance Benchmark Tool for Data Stream Management Systems Combined with DBMS (DBMS와 결합된 데이터스트림관리시스템을 위한 성능 평가 도구 개발)

  • Kim, Gyoung-Bae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.8
    • /
    • pp.1-11
    • /
    • 2010
  • Many applications of DSMS(Data Stream Management System) require not only to process real-time stream data efficiently but also to provide high quality services such as data mining and data warehouse combining with DBMS(Database Management System) to users. In this paper we execute the performance benchmark of the combined system of DSMS and DBMS that is developed for high quality services. We use the stream data of network monitoring application system and combine the traditional representative DSMSs and DBMSs in a single system for the performance testing. We develop the total performance benchmark tool implementing JAVA language for the our testing. For our performance testing, we combine DSMS such as STREAM and Coral8 and DBMS such MySQL and Oracle10g respectively.

Design and Implementation of a SQL based Moving Object Query Process System for Controling Transportation Vehicle (물류 차량 관제를 위한 SQL 기반 이동 객체 질의 처리 시스템의 설계 및 구현)

  • Jung, Young-Jin;Ryu, Keun-Ho
    • The KIPS Transactions:PartD
    • /
    • v.12D no.5 s.101
    • /
    • pp.699-708
    • /
    • 2005
  • It becomes easy and generalized to track the cellular phone users and vehicles according to the Progress of wireless telecommunication, the spread of network, and the miniaturization of terminal devices. It has been constantly studied to provide location based services to furnish suitable services depending on the positions of customers. Various vehicle tracking and management systems are developed to utilize and manage the vehicle locations to relieve the congestion of traffic and to smooth transportation. However the designed previous work can not evaluated in real world, because most of previous work is only designed not implemented and it is developed for simple model to handle a point, a line, a polygon object. Therefore, we design a moving object query language and implement a vehicle management system to search the positions and trajectories of vehicles and to analyze the cost of transportation effectively. The designed query language based on a SQL can be utilized to get the trajectories between two specific places, the departure time, the arrival time of vehicles, and the predicted uncertainty positions, etc. In addition, the proposed moving object query language for managing transportation vehicles is useful to analyze the cost of trajectories in a variety of moving object management system containing transportation.

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.

Communication Quality Measurement and Management System in Smartphone (스마트폰에서의 WCDMA 통화품질 측정 및 관리시스템)

  • Choi, Ho-Yeon;Lee, Sang-Min;Lee, Goo-Yeon
    • Journal of Industrial Technology
    • /
    • v.31 no.B
    • /
    • pp.99-105
    • /
    • 2011
  • Recently, smartphones have been widely used in wireless communications which caused severe data overload in cellular networks. Cellular companies consider increasing wireless data capacity in overloaded areas. Determining overloaded areas is mainly made from user reports which are currently collected by calling or simple apps by some active users experiencing overloads. Since the current reporting method is inconvenient and does not contain detail information about the overload conditions, and therefore is used by only some active users, the overloaded areas determined from the reports are not likely to be accurate. Therefore, we need a new reporting method which is convenient enough for even inactive users to willingly report the overload condition and which delivers detail information about the overloads. In this paper, we design and implement a measurement function of smartphone's WCDMA communication quality including overload condition and management system of mobile wireless communication networks about smartphones' communication quality.

  • PDF

Automatic Generation of DB Images for Testing Enterprise Systems (전사적 응용시스템 테스트를 위한 DB이미지 생성에 관한 연구)

  • Kwon, Oh-Seung;Hong, Sa-Neung
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.37-58
    • /
    • 2011
  • In general, testing DB applications is much more difficult than testing other types of software. The fact that the DB states as much as the input data influence and determine the procedures and results of program testing is one of the decisive reasons for the difficulties. In order to create and maintain proper DB states for testing, it not only takes a lot of time and efforts, but also requires extensive IT expertise and business knowledge. Despite the difficulties, there are not enough research and tools for the needed help. This article reports the result of research on automatic creation and maintenance of DB states for testing DB applications. As its core, this investigation develops an automation tool which collects relevant information from a variety of sources such as log, schema, tables and messages, combines collected information intelligently, and creates pre- and post-Images of database tables proper for application tests. The proposed procedures and tool are expected to be greatly helpful for overcoming inefficiencies and difficulties in not just unit and integration tests but including regression tests. Practically, the tool and procedures proposed in this research allows developers to improve their productivity by reducing time and effort required for creating and maintaining appropriate DB sates, and enhances the quality of DB applications since they are conducive to a wider variety of test cases and support regression tests. Academically, this research deepens our understanding and introduces new approach to testing enterprise systems by analyzing patterns of SQL usages and defining a grammar to express and process the patterns.

A Synchronization Algorithm for Mobile Database using Message Digest (메시지 다이제스트를 이용한 모바일 데이타베이스 동기화 알고리즘)

  • Moon, Chang-Joo;Choi, Mi-Young;Kim, Sang-Min;Jung, Jin-Hwan
    • Journal of KIISE:Databases
    • /
    • v.34 no.4
    • /
    • pp.357-368
    • /
    • 2007
  • This paper proposes the SAMD(Synchronization Algorithms based on Message Digest) algorithm based on message digest in order to data synchronization between a database server and a mobile database. The SAMD algorithm makes the images at the database server and the mobile database using message digest tables and compares two images in order to select the rows needed to synchronization. If two images are different, the synchronization is progressed according to synchronization policy. The SAMD algorithm does not use techniques which are dependent the some database vendor and also triggers, stored procedures and timestamps. The SAMD only uses the standard SQL functions for the synchronization. Therefore the SAMD algorithm is used in any combinations of the database server and the mobile database because the SAMD algorithm is not depended on some database venders. Also the SAMD algorithm shows better performance compared with existing synchronization products.

Implementation of a Real-time Network Traffic Management System (실시간 네트워크 트래픽 매니지먼트 시스템 구현)

  • Yong, Ki-Tak;Lee, Chae-Woo
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.45 no.8
    • /
    • pp.106-113
    • /
    • 2008
  • In this paper, we will design and substantiate the Network Traffic management system(MAGI), which makes up for the weak points of existing Network Monitoring tools and realize dynamic firewall; MRTG, ntop, ethereal, and nettest. The Network Traffic Monitoring and analyzing system differs from existing software-based Network Monitoring tools as MAGI is a combination of software and hardware. Not only the innovative web user interface applied to the software makes analyzing traffic simpler, but it also has a function for monitoring and managing conditions of the hardware. The web user interface implemented with PHP and MySQL helps to use the database and as a result, analyzing data becomes easier. On the hardware part, the appliance is configured as a bridge in a network. As a result, managed system does not have overload.