• Title/Summary/Keyword: Fuzzy data mining

Search Result 90, Processing Time 0.024 seconds

Classification Rue Mining from Fuzzy Data based on Fuzzy Decision Tree (퍼지 데이타에 대한 퍼지 결정트리 기반 분류규칙 마이닝)

  • Lee, Geon-Myeong
    • Journal of KIISE:Software and Applications
    • /
    • v.28 no.1
    • /
    • pp.64-72
    • /
    • 2001
  • 결정트리 생성은 일련의 특징값으로 기술된 사례들로부터 분류 지식을 추출하는 학습 방법중의 하나이다. 현장에서 수집되는 사례들은 관측 오류, 주관적인 판단, 불확실성 등으로 인해서 애매하게 주어지는 경우가 많다. 퍼지숫자나 구간값을 사용함으로써 이러한 애매한 데이타의 수치 속성은 쉽게 표현될 수 있다. 이 논문에서는 수치 속성은 보통값 뿐마아니라 퍼지숫자나 구간값을 갖을 수 있고, 비수치 속서은 보통값을 가지며, 데이터의 클래스는 확신도를 기자는 학습 데이터들로 부터, 분류 규칙을 마이닝하기 위한 퍼지 결정트리 생성 방법을 제안한다. 또한 제안한 방법에 의해 생성된 퍼지 결정트리를 사용하여, 새로운 데이터에 대한 클래스를 결정하는 추론 방법을 소개한다. 한편, 제안된 방법의 유용성을 보이기 위해 수행한 실험의 결과를 보인다.

  • PDF

A XML Schema Matching based on Fuzzy Similarity Measure

  • Kim, Chang-Suk;Sim, Kwee-Bo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1482-1485
    • /
    • 2005
  • An equivalent schema matching among several different source schemas is very important for information integration or mining on the XML based World Wide Web. Finding most similar source schema corresponding mediated schema is a major bottleneck because of the arbitrary nesting property and hierarchical structures of XML DTD schemas. It is complex and both very labor intensive and error prune job. In this paper, we present the first complex matching of XML schema, i.e. XML DTD, inlining two dimensional DTD graph into flat feature values. The proposed method captures not only schematic information but also integrity constraints information of DTD to match different structured DTD. We show the integrity constraints based hierarchical schema matching is more semantic than the schema matching only to use schematic information and stored data.

  • PDF

A XML DTD Matching using Fuzzy Similarity Measure

  • Kim, Chang-Suk;Son, Dong-Cheul;Kim, Dae-Su
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.3 no.1
    • /
    • pp.32-36
    • /
    • 2003
  • An equivalent schema matching among several different source schemas is very important for information integration or mining on the XML based World Wide Web. Finding most similar source schema corresponding mediated schema is a major bottleneck because of the arbitrary nesting property and hierarchical structures of XML DTD schemas. It is complex and both very labor intensive and error prune job. In this paper, we present the first complex matching of XML schema, i.e. XML DTD. The proposed method captures not only schematic information but also integrity constraints information of DTD to match different structured DTD. We show the integrity constraints based hierarchical schema matching is more semantic than the schema matching only to use schematic information and stored data.

Big Numeric Data Classification Using Grid-based Bayesian Inference in the MapReduce Framework

  • Kim, Young Joon;Lee, Keon Myung
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.14 no.4
    • /
    • pp.313-321
    • /
    • 2014
  • In the current era of data-intensive services, the handling of big data is a crucial issue that affects almost every discipline and industry. In this study, we propose a classification method for large volumes of numeric data, which is implemented in a distributed programming framework, i.e., MapReduce. The proposed method partitions the data space into a grid structure and it then models the probability distributions of classes for grid cells by collecting sufficient statistics using distributed MapReduce tasks. The class labeling of new data is achieved by k-nearest neighbor classification based on Bayesian inference.

EXTENDED ONLINE DIVISIVE AGGLOMERATIVE CLUSTERING

  • Musa, Ibrahim Musa Ishag;Lee, Dong-Gyu;Ryu, Keun-Ho
    • Proceedings of the KSRS Conference
    • /
    • 2008.10a
    • /
    • pp.406-409
    • /
    • 2008
  • Clustering data streams has an importance over many applications like sensor networks. Existing hierarchical methods follow a semi fuzzy clustering that yields duplicate clusters. In order to solve the problems, we propose an extended online divisive agglomerative clustering on data streams. It builds a tree-like top-down hierarchy of clusters that evolves with data streams using geometric time frame for snapshots. It is an enhancement of the Online Divisive Agglomerative Clustering (ODAC) with a pruning strategy to avoid duplicate clusters. Our main features are providing update time and memory space which is independent of the number of examples on data streams. It can be utilized for clustering sensor data and network monitoring as well as web click streams.

  • PDF

On development of supporting tool for Folksonomy Mining based on Formal Concept Analysis (형식개념분석을 이용한 폭소노미 마이닝 기법과 지원도구의 개발)

  • Kang, Yu-Kyung;Hwang, Suk-Hyung;Yang, Hae-Sool
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.10 no.8
    • /
    • pp.1877-1893
    • /
    • 2009
  • Folksonomy is a user-generated taxonomy to organize information by which a user assigns tags to resources published on the web. Triadic datas that indicate relations of between users, tags, and resources, are created by collaborative tagging from many users in folksonomy-based system. Such the folksonomy data has been utilized in the field of the semantic web and web2.0 as metadata about web resources. In this paper, we propose FCA-based folksonomy data mining approach in order to extract the useful information from folksonomy data with various points of view. And we developed tool for supporting our approach. In order to verify the usefulness of our proposed approach and FMT, we have done some experiments for data of del.icio.us, which is a popular folksonomy-based bookmarking system. And we report about result of our experiments.

A Study of Short-Term Load Forecasting System Using Data Mining (데이터 마이닝을 이용한 단기 부하 예측 시스템 연구)

  • Joo, Young-Hoon;Jung, Keun-Ho;Kim, Do-Wan;Park, Jin-Bae
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.2
    • /
    • pp.130-135
    • /
    • 2004
  • This paper presents a new design methods of the short-term load forecasting system (STLFS) using the data mining. The structure of the proposed STLFS is divided into two parts: the Takagi-Sugeno (T-S) fuzzy model-based classifier and predictor The proposed classifier is composed of the Gaussian fuzzy sets in the premise part and the linearized Bayesian classifier in the consequent part. The related parameters of the classifier are easily obtained from the statistic information of the training set. The proposed predictor takes form of the convex combination of the linear time series predictors for each inputs. The problem of estimating the consequent parameters is formulated by the convex optimization problem, which is to minimize the norm distance between the real load and the output of the linear time series estimator. The problem of estimating the premise parameters is to find the parameter value minimizing the error between the real load and the overall output. Finally, to show the feasibility of the proposed method, this paper provides the short-term load forecasting example.

A Prediction Model Based on Relevance Vector Machine and Granularity Analysis

  • Cho, Young Im
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.16 no.3
    • /
    • pp.157-162
    • /
    • 2016
  • In this paper, a yield prediction model based on relevance vector machine (RVM) and a granular computing model (quotient space theory) is presented. With a granular computing model, massive and complex meteorological data can be analyzed at different layers of different grain sizes, and new meteorological feature data sets can be formed in this way. In order to forecast the crop yield, a grey model is introduced to label the training sample data sets, which also can be used for computing the tendency yield. An RVM algorithm is introduced as the classification model for meteorological data mining. Experiments on data sets from the real world using this model show an advantage in terms of yield prediction compared with other models.

A new viewpoint on stability theorem for engineering structural and geotechnical parameter

  • Timothy Chen;Ruei-Yuan Wang;Yahui Meng;Z.Y. Chen
    • Geomechanics and Engineering
    • /
    • v.36 no.5
    • /
    • pp.475-487
    • /
    • 2024
  • Many uncertainties affect the stability assessment of rock structures. Some of these factors significantly influence technology decisions. Some of these factors belong to the geological domain, and spatial uncertainty measurements are useful for structural stability analysis. This paper presents an integrated approach to study the stability of rock structures, including spatial factors. This study models two main components: discrete structures (fault zones) and well known geotechnical parameters (rock quality indicators). The geostatistical modeling criterion are used to quantify geographic uncertainty by producing simulated maps and RQD values for multiple equally likely error regions. Slope stability theorem would be demonstrated by modeling local failure zones and RQDs. The approach proided is validated and finally, the slope stability analysis method and fuzzy Laypunov criterion are applied to mining projects with limited measurement data. The goals of this paper are towards access to adequate, safe and affordable housing and basic services, promotion of inclusive and sustainable urbanization and participation, implementation of sustainable and disaster-resilient buildings, sustainable human settlement planning and manage. Simulation results of linear and nonlinear structures show that the proposed method is able to identify structural parameters and their changes due to damage and unknown excitations. Therefore, the goal is believed to achieved in the near future by the ongoing development of AI and fuzzy theory.

Increasing Spatial Resolution of Remotely Sensed Image using HNN Super-resolution Mapping Combined with a Forward Model

  • Minh, Nguyen Quang;Huong, Nguyen Thi Thu
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.31 no.6_2
    • /
    • pp.559-565
    • /
    • 2013
  • Spatial resolution of land covers from remotely sensed images can be increased using super-resolution mapping techniques for soft-classified land cover proportions. A further development of super-resolution mapping technique is downscaling the original remotely sensed image using super-resolution mapping techniques with a forward model. In this paper, the model for increasing spatial resolution of remote sensing multispectral image is tested with real SPOT 5 imagery at 10m spatial resolution for an area in Bac Giang Province, Vietnam in order to evaluate the feasibility of application of this model to the real imagery. The soft-classified land cover proportions obtained using a fuzzy c-means classification are then used as input data for a Hopfield neural network (HNN) to predict the multispectral images at sub-pixel spatial resolution. The 10m SPOT multispectral image was improved to 5m, 3,3m and 2.5m and compared with SPOT Panchromatic image at 2.5m resolution for assessment.Visually, the resulted image is compared with a SPOT 5 panchromatic image acquired at the same time with the multispectral data. The predicted image is apparently sharper than the original coarse spatial resolution image.