• Title/Summary/Keyword: SQL 분석

Search Result 234, Processing Time 0.027 seconds

A Study on Designing Schema for Process Repository (프로세스 리파지토리를 위한 스키마 설계에 관한 연구)

  • 이재정;정호석;홍순구
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.8 no.2
    • /
    • pp.98-107
    • /
    • 2003
  • The purpose of this study is to propose the schema for process repository that overcomes the limitations of existing process repositories in which it is not easy to compare and analyze such the unstructured data as a process model. To achieve this goal, the previous studies on process repository are reviewed, the schema for process repository is designed, and the tables for relational database are created. For the validation of the schema presented in this paper, the process model stored in relational database is examed using SQL. The contributions of this study are as follows. First, the schema for process repository storing process models is designed and validated. Second, the process model created by the new schema and stored in relational database leads to improve the performance of process analysis.

  • PDF

Implementation of Source Analysis Tool for Vulnerability Search of Web Program (웹 프로그램의 취약점 검색을 위한 소스분석 툴 구현)

  • Kim, Soung-Uk;Hoang, Tae-Moon;Kim, Pan-Kyu;Park, Sang-Su;Lee, Jong-Hyeok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • v.9 no.2
    • /
    • pp.776-779
    • /
    • 2005
  • The Program language for web, such as PHP, JSP, ASP and so on, make it possible to offer more user interactive page by using with HTML. These language and program have been developed with great speed, but security part could not catch up with this development. As a result, it has brought a problem which is expose many server systems to the outside. In this research, we implement Web and SQL analysis program which can analysis hacking causing factor. With this analysis program, we will show you how much efficient it has compared with security patch for server system.

  • PDF

ERD Representation using Auto-Generated Form and SQL (자동 생성 폼과 SQL을 이용한 ERD 표현)

  • Ra, Young-Gook
    • The Journal of the Korea Contents Association
    • /
    • v.9 no.5
    • /
    • pp.61-75
    • /
    • 2009
  • Generally, the development of the database application includes the requirement analysis phase of creating ERD (Entity Relationship Diagram) and process models, coding, and testing. From the above phases, the analysis phase is not most formalized. It is usually hard task because (1) customers don't know the details of the desired system; (2) developers can't with ease understand the business logic of the customers; (3) the outcomes of the analysis, which are ERD and process models, are not easy to understand to the customers. This paper propose that the executional forms, which are better to understand the systems, should be presented to the customers instead of the ERD. These forms should accept the data input so that customers can review the various aspects of the outcome models. The developers should be able to instantly implement the business logic and also should be able to visually demonstrate the logic in order to get the details of it. For this goal, the customer supplied business logic should be able to be implemented by the references between forms, actions, constraints from the perspective of the data flow. The customers try to execute the forms implementing the business logic and review their supplied logic find new necessary business logic of their own. Iterating these processes for the requirement analysis would result in the success of the analysis which is sufficiently detailed without conflicts.

The Model of Network Packet Analysis based on Big Data (빅 데이터 기반의 네트워크 패킷 분석 모델)

  • Choi, Bomin;Kong, Jong-Hwan;Han, Myung-Mook
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.23 no.5
    • /
    • pp.392-399
    • /
    • 2013
  • Due to the development of IT technology and the information age, a dependency of the network over the most of our lives have grown to a greater extent. Although it provides us to get various useful information and service, it also has negative effectiveness that can provide network intruder with vulnerable roots. In other words, we need to urgently cope with theses serious security problem causing service disableness or system connected to network obstacle with exploiting various packet information. Many experts in a field of security are making an effort to develop the various security solutions to respond against these threats, but existing solutions have a lot of problems such as lack of storage capacity and performance degradation along with the massive increase of packet data volume. Therefore we propose the packet analysis model to apply issuing Big Data technology in the field of security. That is, we used NoSQL which is technology of massive data storage to collect the packet data growing massive and implemented the packet analysis model based on K-means clustering using MapReudce which is distributed programming framework, and then we have shown its high performance by experimenting.

Study of MongoDB Architecture by Data Complexity for Big Data Analysis System (빅데이터 분석 시스템 구현을 위한 데이터 구조의 복잡성에 따른 MongoDB 환경 구성 연구)

  • Hyeopgeon Lee;Young-Woon Kim;Jin-Woo Lee;Seong Hyun Lee
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.16 no.5
    • /
    • pp.354-361
    • /
    • 2023
  • Big data analysis systems apply NoSQL databases like MongoDB to store, process, and analyze diverse forms of large-scale data. MongoDB offers scalability and fast data processing speeds through distributed processing and data replication, depending on its configuration. This paper investigates the suitable MongoDB environment configurations for implementing big data analysis systems. For performance evaluation, we configured both single-node and multi-node environments. In the multi-node setup, we expanded the number of data nodes from two to three and measured the performance in each environment. According to the analysis, the processing speeds for complex data structures with three or more dimensions are approximately 5.75% faster in the single-node environment compared to an environment with two data nodes. However, a setting with three data nodes processes data about 25.15% faster than the single-node environment. On the other hand, for simple one-dimensional data structures, the multi-node environment processes data approximately 28.63% faster than the single-node environment. Further research is needed to practically validate these findings with diverse data structures and large volumes of data.

Development of educational programs for managing medical information utilizing medical data generation and analysis techniques (의료 데이터 발생과 분석기술을 활용한 의료정보관리 교육용 프로그램 개발)

  • Choi, Joonyoung
    • Journal of Digital Convergence
    • /
    • v.15 no.10
    • /
    • pp.377-386
    • /
    • 2017
  • This study has developed a medical information management educational program that can improve the management ability of medical information. The educational medical information management program was developed for 8mnths uing VB. The database utilized the ACCESS Database, which allows learners to easily understand and understand the structure of the data. The learners enter data in the discharge analysis and the cancer registration program and the incomplete program after analyze the medical records. After entering and saving data, medical information management programs can be used to understand and analyze the structure of the database to generate medical information. The educational programs can improve the ability of learners to manage medical information by extracting the necessary data from the database directly through SQL and creating various medical information. However, although the medical information management program is an educational program, there is no evaluation system for the learners program operation. Accordingly, the next studies should develop the assessment system of the medical information management program for learners evaluation.

OLAP System and Performance Evaluation for Analyzing Web Log Data (웹 로그 분석을 위한 OLAP 시스템 및 성능 평가)

  • 김지현;용환승
    • Journal of Korea Multimedia Society
    • /
    • v.6 no.5
    • /
    • pp.909-920
    • /
    • 2003
  • Nowadays, IT for CRM has been growing and developed rapidly. Typical techniques are statistical analysis tools, on-line multidimensional analytical processing (OLAP) tools, and data mining algorithms (such neural networks, decision trees, and association rules). Among customer data, web log data is very important and to use these data efficiently, applying OLAP technology to analyze multi-dimensionally. To make OLAP cube, we have to precalculate multidimensional summary results in order to get fast response. But as the number of dimensions and sparse cells increases, data explosion occurs seriously and the performance of OLAP decreases. In this paper, we presented why the web log data sparsity occurs and then what kinds of sparsity patterns generate in the two and t.he three dimensions for OLAP. Based on this research, we set up the multidimensional data models and query models for benchmark with each sparsity patterns. Finally, we evaluated the performance of three OLAP systems (MS SQL 2000 Analysis Service, Oracle Express and C-MOLAP).

  • PDF

A Study on N-IDS Detection and Packet Analysis regarding a DoS attack (DoS공격에 대한 N-IDS 탐지 및 패킷 분석 연구)

  • Chun, Woo-Sung;Park, Dea-Woo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.6
    • /
    • pp.217-224
    • /
    • 2008
  • EThis paper is study regarding banking institution and DoS attack regarding government organization which occurred in 2008. We used a tool aggressive actual DoS You install the N-IDS which used Snort in networks in order to detect a DoS attack. Storages of Winpcap and a packet to detect a packet and MySQL, HSC, to analyze. We install NET Framework etc. E-Watch etc. analyzes Packet regarding a DoS attack of a hacker and TCP, UDP etc. information, Port, MAC and IP information etc. through packet analysis tools. There is a meaning you analyze data regarding the cyber DoS, DDoS attack that is dysfunction of Ubiquitous Information Society, and it generates forensics data regarding an invader and back-tracking analysis data, and to ensure safe Internet information system through this paper study.

  • PDF

Natural Language Query Interpretation System for Biomedical Database Access (구조화된 생물의료 정보의 접근을 위한 자연언어 질의 시스템)

  • 이호동;박종철
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.04b
    • /
    • pp.487-489
    • /
    • 2002
  • 본 논문은 이질적인 데이터베이스에 산재되어 있는 생물의료 정보의 개념적인 접근을 가능하게 하기 위한 자연언어질의 시스템을 설명한다. 이를 위해 본 시스템에서는 질의문을 SQL, OQL, CPL 데이터베이스 정형언어로 변환하는데, 이 과정에서 필요한 질의문의 분석 및 변환과정을 보인다. 제안하는 방법은 구문분석에 의해 도출된 정보를 이용해 직접 다양한 정형언어들로 변환하므로, 시스템의 구조가 간결해지고 모듈화되어 전체 성능과 이식성의 향상을 가져올 수 있다.

  • PDF

Development of an OODBMS Functionality Testing Tool Prototype. (객체지향 DBMS 기능 시험 도구의 프로토타입 개발)

  • 김은영;이상호;전성택
    • The Journal of Information Technology and Database
    • /
    • v.2 no.2
    • /
    • pp.25-34
    • /
    • 1995
  • In this paper, we present design philosophy and implementation issues of a functionality testing tool for object-oriented database systems. A testing tool has been developed to validate UniSQL/X functionalities with C++ interface. A testing tool is designed under consideration of scaleability, simplicity and extendibility. The schema is deliberately constructed to verify the object-oriented functionalities such as abstraction, inheritance and aggregation. Each test item has been derived under various black box techniques such as equivalent partitioning and boundary-value analysis. The testing tool consists of six phases, namely, database creation, database population, construction of testindex, compilation and link, execution and result reporting, and final cleanup. The prototype provides more than 140 test items at 90 programs.

  • PDF