• Title/Summary/Keyword: Generate Data

Search Result 3,066, Processing Time 0.035 seconds

Contract-based Access Control Method for NFT Use Rights

  • Jeong, Yoonsung;Ko, Deokyoon;Seo, Jungwon;Park, Sooyong;Kim, Seong-Jin;Kim, Bum-Soo;Kim, Do-Young
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.11
    • /
    • pp.1-11
    • /
    • 2022
  • In this paper, we propose an NFT(Non-Fungible Token)-based access control method for safely sharing data between users in blockchain environment. Since all data stored in the blockchain can be accessed by anyone due to the nature of the technology, it is necessary to control access except for authorized users when sharing sensitive data. For that, we generate each data as NFT and controls access to the data through the smart contract. In addition, in order to overcome the limitations of single ownership of the existing NFT, we separated the NFT into ownership and use rights, so that data can be safely shared between users. Ownership is represented as an original NFT, use rights is represented as a copied NFT, and all data generated as NFT is encrypted and uploaded, so data can be shared only through the smart contract with access control. To verify this approach, we set up a hypothetical scenario called Building Information Modeling (BIM) data trade, and deployed a smart contract that satisfies 32 function call scenarios that require access control. Also, we evaluated the stability in consideration of the possibility of decryption through brute-force attack. Through our approach, we confirmed that the data can be safely shared between users in blockchain environment.

Efficient Mining of Frequent Subgraph with Connectivity Constraint

  • Moon, Hyun-S.;Lee, Kwang-H.;Lee, Do-Heon
    • Proceedings of the Korean Society for Bioinformatics Conference
    • /
    • 2005.09a
    • /
    • pp.267-271
    • /
    • 2005
  • The goal of data mining is to extract new and useful knowledge from large scale datasets. As the amount of available data grows explosively, it became vitally important to develop faster data mining algorithms for various types of data. Recently, an interest in developing data mining algorithms that operate on graphs has been increased. Especially, mining frequent patterns from structured data such as graphs has been concerned by many research groups. A graph is a highly adaptable representation scheme that used in many domains including chemistry, bioinformatics and physics. For example, the chemical structure of a given substance can be modelled by an undirected labelled graph in which each node corresponds to an atom and each edge corresponds to a chemical bond between atoms. Internet can also be modelled as a directed graph in which each node corresponds to an web site and each edge corresponds to a hypertext link between web sites. Notably in bioinformatics area, various kinds of newly discovered data such as gene regulation networks or protein interaction networks could be modelled as graphs. There have been a number of attempts to find useful knowledge from these graph structured data. One of the most powerful analysis tool for graph structured data is frequent subgraph analysis. Recurring patterns in graph data can provide incomparable insights into that graph data. However, to find recurring subgraphs is extremely expensive in computational side. At the core of the problem, there are two computationally challenging problems. 1) Subgraph isomorphism and 2) Enumeration of subgraphs. Problems related to the former are subgraph isomorphism problem (Is graph A contains graph B?) and graph isomorphism problem(Are two graphs A and B the same or not?). Even these simplified versions of the subgraph mining problem are known to be NP-complete or Polymorphism-complete and no polynomial time algorithm has been existed so far. The later is also a difficult problem. We should generate all of 2$^n$ subgraphs if there is no constraint where n is the number of vertices of the input graph. In order to find frequent subgraphs from larger graph database, it is essential to give appropriate constraint to the subgraphs to find. Most of the current approaches are focus on the frequencies of a subgraph: the higher the frequency of a graph is, the more attentions should be given to that graph. Recently, several algorithms which use level by level approaches to find frequent subgraphs have been developed. Some of the recently emerging applications suggest that other constraints such as connectivity also could be useful in mining subgraphs : more strongly connected parts of a graph are more informative. If we restrict the set of subgraphs to mine to more strongly connected parts, its computational complexity could be decreased significantly. In this paper, we present an efficient algorithm to mine frequent subgraphs that are more strongly connected. Experimental study shows that the algorithm is scaling to larger graphs which have more than ten thousand vertices.

  • PDF

Interactions between Soil Moisture and Weather Prediction in Rainfall-Runoff Application : Korea Land Data Assimilation System(KLDAS) (수리 모형을 이용한 Korea Land Data Assimilation System (KLDAS) 자료의 수문자료에 대한 영향력 분석)

  • Jung, Yong;Choi, Minha
    • 한국방재학회:학술대회논문집
    • /
    • 2011.02a
    • /
    • pp.172-172
    • /
    • 2011
  • The interaction between land surface and atmosphere is essentially affected by hydrometeorological variables including soil moisture. Accurate estimation of soil moisture at spatial and temporal scales is crucial to better understand its roles to the weather systems. The KLDAS(Korea Land Data Assimilation System) is a regional, specifically Korea peninsula land surface information systems. As other prior land data assimilation systems, this can provide initial soil field information which can be used in atmospheric simulations. For this study, as an enabling high-resolution tool, weather research and forecasting(WRF-ARW) model is applied to produce precipitation data using GFS(Global Forecast System) with GFS embedded and KLDAS soil moisture information as initialization data. WRF-ARW generates precipitation data for a specific region using different parameters in physics options. The produced precipitation data will be employed for simulations of Hydrological Models such as HEC(Hydrologic Engineering Center) - HMS(Hydrologic Modeling System) as predefined input data for selected regional water responses. The purpose of this study is to show the impact of a hydrometeorological variable such as soil moisture in KLDAS on hydrological consequences in Korea peninsula. The study region, Chongmi River Basin, is located in the center of Korea Peninsular. This has 60.8Km river length and 17.01% slope. This region mostly consists of farming field however the chosen study area placed in mountainous area. The length of river basin perimeter is 185Km and the average width of river is 9.53 meter with 676 meter highest elevation in this region. We have four different observation locations : Sulsung, Taepyung, Samjook, and Sangkeug observatoriesn, This watershed is selected as a tentative research location and continuously studied for getting hydrological effects from land surface information. Simulations for a real regional storm case(June 17~ June 25, 2006) are executed. WRF-ARW for this case study used WSM6 as a micro physics, Kain-Fritcsch Scheme for cumulus scheme, and YSU scheme for planetary boundary layer. The results of WRF simulations generate excellent precipitation data in terms of peak precipitation and date, and the pattern of daily precipitation for four locations. For Sankeug observatory, WRF overestimated precipitation approximately 100 mm/day on July 17, 2006. Taepyung and Samjook display that WRF produced either with KLDAS or with GFS embedded initial soil moisture data higher precipitation amounts compared to observation. Results and discussions in detail on accuracy of prediction using formerly mentioned manners are going to be presented in 2011 Annual Conference of the Korean Society of Hazard Mitigation.

  • PDF

A Study on the Simulated Radar Terrain Scan Data Generated from Discrete Terrain (이산지형정보에서 생성된 레이다 모의 지형 스캔 정보에 관한 연구)

  • Seunghun, Kang;Sunghyun, Hahn;Jiyeon, Jeon;Dongju, Lim;Sangchul, Lee
    • Journal of Aerospace System Engineering
    • /
    • v.16 no.6
    • /
    • pp.1-7
    • /
    • 2022
  • A simulated radar terrain scan data generation method is employed for terrain following. This method scans the discrete terrain by sequentially radiating beams from the radar to the desired scan area with the same azimuth but varying elevation angles. The terrain data collected from the beam is integrated to generate the simulated radar terrain scan data, which comprises radar-detected points. However, these points can be located far from the beam centerline when the radar is far from them due to beam divergence. This paper proposes a geometry-based terrain scan data generation method for analysing simulated radar terrain scan data. The method involves detecting geometric points along the beam centerline, which forms the geometry-based terrain scan data. The analysis of the simulated radar terrain scan data utilising this method confirms that the beam width effects are accounted for in the results.

A Study on the Decryption Method for Volume Encryption and Backup Applications (볼륨 암호화 및 백업 응용프로그램에 대한 복호화 방안 연구)

  • Gwui-eun Park;Min-jeong Lee;Soo-jin Kang;Gi-yoon Kim;Jong-sung Kim
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.33 no.3
    • /
    • pp.511-525
    • /
    • 2023
  • As awareness of personal data protection increases, various Full Disk Encryption (FDE)-based applications are being developed that real-time encryption or use virtual drive volumes to protect data on user's PC. FDE-based applications encrypt and protect the volume containing user's data. However, as disk encryption technology advances, some users are abusing FDE-based applications to encrypt evidence associated with criminal activities, which makes difficulties in digital forensic investigations. Thus, it is necessary to analyze the encryption process used in FDE-based applications and decrypt the encrypted data. In this paper, we analyze Cryptomator and Norton Ghost, which provide volume encryption and backup functions. We analyze the encrypted data structure and encryption process to classify the main data of each application and identify the encryption algorithm used for data decryption. The encryption algorithms of these applications are recently emergin gor customized encryption algorithms which are analyzed to decrypt data. User password is essential to generate a data encryption key used for decryption, and a password acquisition method is suggested using the function of each application. This supplemented the limitations of password investigation, and identifies user data by decrypting encrypted data based on the acquired password.

Bridge Safety Determination Edge AI Model Based on Acceleration Data (가속도 데이터 기반 교량 안전 판단을 위한 Edge AI 모델)

  • Jinhyo Park;Yong-Geun Hong;Joosang Youn
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.29 no.4
    • /
    • pp.1-11
    • /
    • 2024
  • Bridges crack and become damaged due to age and external factors such as earthquakes, lack of maintenance, and weather conditions. With the number of aging bridge on the rise, lack of maintenance can lead to a decrease in safety, resulting in structural defects and collapse. To prevent these problems and reduce maintenance costs, a system that can monitor the condition of bridge and respond quickly is needed. To this end, existing research has proposed artificial intelligence model that use sensor data to identify the location and extent of cracks. However, existing research does not use data from actual bridge to determine the performance of the model, but rather creates the shape of the bridge through simulation to acquire data and use it for training, which does not reflect the actual bridge environment. In this paper, we propose a bridge safety determination edge AI model that detects bridge abnormalities based on artificial intelligence by utilizing acceleration data from bridge occurring in the field. To this end, we newly defined filtering rules for extracting valid data from acceleration data and constructed a model to apply them. We also evaluated the performance of the proposed bridge safety determination edge AI model based on data collected in the field. The results showed that the F1-Score was up to 0.9565, confirming that it is possible to determine safety using data from real bridge, and that rules that generate similar data patterns to real impact data perform better.

Improving minority prediction performance of support vector machine for imbalanced text data via feature selection and SMOTE (단어선택과 SMOTE 알고리즘을 이용한 불균형 텍스트 데이터의 소수 범주 예측성능 향상 기법)

  • Jongchan Kim;Seong Jun Chang;Won Son
    • The Korean Journal of Applied Statistics
    • /
    • v.37 no.4
    • /
    • pp.395-410
    • /
    • 2024
  • Text data is usually made up of a wide variety of unique words. Even in standard text data, it is common to find tens of thousands of different words. In text data analysis, usually, each unique word is treated as a variable. Thus, text data can be regarded as a dataset with a large number of variables. On the other hand, in text data classification, we often encounter class label imbalance problems. In the cases of substantial imbalances, the performance of conventional classification models can be severely degraded. To improve the classification performance of support vector machines (SVM) for imbalanced data, algorithms such as the Synthetic Minority Over-sampling Technique (SMOTE) can be used. The SMOTE algorithm synthetically generates new observations for the minority class based on the k-Nearest Neighbors (kNN) algorithm. However, in datasets with a large number of variables, such as text data, errors may accumulate. This can potentially impact the performance of the kNN algorithm. In this study, we propose a method for enhancing prediction performance for the minority class of imbalanced text data. Our approach involves employing variable selection to generate new synthetic observations in a reduced space, thereby improving the overall classification performance of SVM.

Classification of Multi-temporal SAR Data by Using Data Transform Based Features and Multiple Classifiers (자료변환 기반 특징과 다중 분류자를 이용한 다중시기 SAR자료의 분류)

  • Yoo, Hee Young;Park, No-Wook;Hong, Sukyoung;Lee, Kyungdo;Kim, Yeseul
    • Korean Journal of Remote Sensing
    • /
    • v.31 no.3
    • /
    • pp.205-214
    • /
    • 2015
  • In this study, a novel land-cover classification framework for multi-temporal SAR data is presented that can combine multiple features extracted through data transforms and multiple classifiers. At first, data transforms using principle component analysis (PCA) and 3D wavelet transform are applied to multi-temporal SAR dataset for extracting new features which were different from original dataset. Then, three different classifiers including maximum likelihood classifier (MLC), neural network (NN) and support vector machine (SVM) are applied to three different dataset including data transform based features and original backscattering coefficients, and as a result, the diverse preliminary classification results are generated. These results are combined via a majority voting rule to generate a final classification result. From an experiment with a multi-temporal ENVISAT ASAR dataset, every preliminary classification result showed very different classification accuracy according to the used feature and classifier. The final classification result combining nine preliminary classification results showed the best classification accuracy because each preliminary classification result provided complementary information on land-covers. The improvement of classification accuracy in this study was mainly attributed to the diversity from combining not only different features based on data transforms, but also different classifiers. Therefore, the land-cover classification framework presented in this study would be effectively applied to the classification of multi-temporal SAR data and also be extended to multi-sensor remote sensing data fusion.

Development of Examination Model of Weather Factors on Garlic Yield Using Big Data Analysis (빅데이터 분석을 활용한 마늘 생산에 미치는 날씨 요인에 관한 영향 조사 모형 개발)

  • Kim, Shinkon
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.19 no.5
    • /
    • pp.480-488
    • /
    • 2018
  • The development of information and communication technology has been carried out actively in the field of agriculture to generate valuable information from large amounts of data and apply big data technology to utilize it. Crops and their varieties are determined by the influence of the natural environment such as temperature, precipitation, and sunshine hours. This paper derives the climatic factors affecting the production of crops using the garlic growth process and daily meteorological variables. A prediction model was also developed for the production of garlic per unit area. A big data analysis technique considering the growth stage of garlic was used. In the exploratory data analysis process, various agricultural production data, such as the production volume, wholesale market load, and growth data were provided from the National Statistical Office, the Rural Development Administration, and Korea Rural Economic Institute. Various meteorological data, such as AWS, ASOS, and special status data, were collected and utilized from the Korea Meteorological Agency. The correlation analysis process was designed by comparing the prediction power of the models and fitness of models derived from the variable selection, candidate model derivation, model diagnosis, and scenario prediction. Numerous weather factor variables were selected as descriptive variables by factor analysis to reduce the dimensions. Using this method, it was possible to effectively control the multicollinearity and low degree of freedom that can occur in regression analysis and improve the fitness and predictive power of regression analysis.

A Study of Measuring Traffic Congestion for Urban Network using Average Link Travel Time based on DTG Big Data (DTG 빅데이터 기반의 링크 평균통행시간을 이용한 도심네트워크 혼잡분석 방안 연구)

  • Han, Yohee;Kim, Youngchan
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.16 no.5
    • /
    • pp.72-84
    • /
    • 2017
  • Together with the Big Data of the 4th Industrial Revolution, the traffic information system has been changed to an section detection system by the point detection system. With DTG(Digital Tachograph) data based on Global Navigation Satellite System, the properties of raw data and data according to processing step were examined. We identified the vehicle trajectory, the link travel time of individual vehicle, and the link average travel time which are generated according to the processing step. In this paper, we proposed a application method for traffic management as characteristics of processing data. We selected the historical data considering the data management status of the center and the availability at the present time. We proposed a method to generate the Travel Time Index with historical link average travel time which can be collected all the time with wide range. We propose a method to monitor the traffic congestion using the Travel Time Index, and analyze the case of intersections when the traffic operation method changed. At the same time, the current situation which makes it difficult to fully utilize DTG data are suggested as limitations.