• Title/Summary/Keyword: inverse learning

Search Result 205, Processing Time 0.033 seconds

Proposition of balanced comparative confidence considering all available diagnostic tools (모든 가능한 진단도구를 활용한 균형비교신뢰도의 제안)

  • Park, Hee Chang
    • Journal of the Korean Data and Information Science Society
    • /
    • v.26 no.3
    • /
    • pp.611-618
    • /
    • 2015
  • By Wikipedia, big data is a broad term for data sets so large or complex that traditional data processing applications are inadequate. Data mining is the computational process of discovering patterns in huge data sets involving methods at the intersection of association rule, decision tree, clustering, artificial intelligence, machine learning. Association rule is a well researched method for discovering interesting relationships between itemsets in huge databases and has been applied in various fields. There are positive, negative, and inverse association rules according to the direction of association. If you want to set the evaluation criteria of association rule, it may be desirable to consider three types of association rules at the same time. To this end, we proposed a balanced comparative confidence considering sensitivity, specificity, false positive, and false negative, checked the conditions for association threshold by Piatetsky-Shapiro, and compared it with comparative confidence and inversely comparative confidence through a few experiments.

The Design of Polynomial RBF Neural Network by Means of Fuzzy Inference System and Its Optimization (퍼지추론 기반 다항식 RBF 뉴럴 네트워크의 설계 및 최적화)

  • Baek, Jin-Yeol;Park, Byaung-Jun;Oh, Sung-Kwun
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.58 no.2
    • /
    • pp.399-406
    • /
    • 2009
  • In this study, Polynomial Radial Basis Function Neural Network(pRBFNN) based on Fuzzy Inference System is designed and its parameters such as learning rate, momentum coefficient, and distributed weight (width of RBF) are optimized by means of Particle Swarm Optimization. The proposed model can be expressed as three functional module that consists of condition part, conclusion part, and inference part in the viewpoint of fuzzy rule formed in 'If-then'. In the condition part of pRBFNN as a fuzzy rule, input space is partitioned by defining kernel functions (RBFs). Here, the structure of kernel functions, namely, RBF is generated from HCM clustering algorithm. We use Gaussian type and Inverse multiquadratic type as a RBF. Besides these types of RBF, Conic RBF is also proposed and used as a kernel function. Also, in order to reflect the characteristic of dataset when partitioning input space, we consider the width of RBF defined by standard deviation of dataset. In the conclusion part, the connection weights of pRBFNN are represented as a polynomial which is the extended structure of the general RBF neural network with constant as a connection weights. Finally, the output of model is decided by the fuzzy inference of the inference part of pRBFNN. In order to evaluate the proposed model, nonlinear function with 2 inputs, waster water dataset and gas furnace time series dataset are used and the results of pRBFNN are compared with some previous models. Approximation as well as generalization abilities are discussed with these results.

Design and Implementation of Recurrent Time Delayed Neural Network Controller Using Fuzzy Compensator (퍼지 보상기를 사용한 리커런트 시간지연 신경망 제어기 설계 및 구현)

  • Lee, Sang-Yun;Shin, Woo-Jae
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.13 no.3
    • /
    • pp.334-341
    • /
    • 2003
  • In this paper, we proposed a recurrent time delayed neural network(RTDNN) controller which compensate a output of neural network controller. Even if learn by neural network controller, it can occur an bad results from disturbance or load variations. So in order to adjust above case, we used the fuzzy compensator to get an expected results. And the weight of main neural network can be changed with the result of learning a inverse model neural network of plant, so a expected dynamic characteristics of plant can be got. As the results of simulation through the second order plant, we confirmed that the proposed recurrent time delayed neural network controller get a good response compare with a time delayed neural network(TDU) controller. We implemented the controller using the DSP processor and applied in a hydraulic servo system. And then we observed an experimental results.

Selection of Representative GCM Based on Performance Indices (성능지표 기반 대표 GCM 선정)

  • Song, Young Hoon;Chung, Eun Sung;Mang, Ngun Za Luai
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2019.05a
    • /
    • pp.101-101
    • /
    • 2019
  • 전 지구적 기온상승으로 인한 기후변화는 사회적, 수문학적, 다양한 분야에 영향을 미친다. 또한 IPCC(Intergovernmental Panel on Climate Change)의 보고서에 따르면 미래에도 지속적으로 기온상승이 예상되며, 이러한 현상은 인류의 삶에 큰 영향을 미칠것으로 예상된다. 또한 수자원 및 관련 분야에서도 기온 상승에 따른 강수량, 강수의 주기 변동, 극한 기후사상의 심도(severity)와 빈도 변화에 따른 다양한 연구가 진행되고 있으며, 미래의 강우량과 온도를 예측하는 기후변화연구에서는 다양한 기후모형을 고려하여 분석한다. 하지만 모든 기후모형이 우리나라에 적합한 것은 아니므로 과거 기후를 모의한 결과를 토대로 성능이 뛰어난 모형의 결과에 더 높은 가중치를 주고 미래를 예측하는 연구가 활발히 진행되고 있다. 일반적으로 기후모형으로 GCM (General Circulation Model) 모의 결과가 이용되는데 우리나라에 대한 GCM 결과의 정확성을 분석하는 연구는 부족한 실정이다. 따라서 본 연구에서는 21개의 GCM을 대상으로 과거 모의 자료(1970년~2005년)를 실제 관측소에서 관측된 강수량과 비교하여 각 GCM들의 성능을 평가하고 이를 토대로, GCM들의 우선순위를 선정하였다. 또한 격자 기반 GCM 결과를 IDW (Inverse Distance Weighted) 방법을 사용하여 기상관측소로 지역적 상세화를 수행하였으며, GCM과 관측자료 사이의 편이를 보정하기 위해 6가지의 Quantile Mapping 방법과 Random Forest 기법을 사용하였다. 또한 편이 보정 기법 중 성능이 좋은 기법을 선택하여 관측소에 적용하였다. 편이 보정된 GCM 모의결과에 대한 성능을 토대로 우수한 GCM 순위를 도출하기 위해 다기준의사결정기법 중 하나인 TOPSIS (Technique for Order of Preference by Similarity to Ideal Solution)를 이용하였다. 그리고 GCM의 전망기간인 2010년부터 2018년까지의 Machine learning 방법과 Quantile mapping의 기법을 비교 및 성능이 우수한 편이 보정 방법을 선택한 후 전망기간 동안의 GCM 성능의 우선순위를 선정하였다.

  • PDF

Development of machine learning framework to inverse-track a contaminant source of hazardous chemicals in rivers (하천에 유입된 유해화학물질의 역추적을 위한 기계학습 프레임워크 개발)

  • Kwon, Siyoon;Seo, Il Won
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2020.06a
    • /
    • pp.112-112
    • /
    • 2020
  • 하천에서 유해화학물질 유입 사고 발생 시 수환경 피해를 최소화하기 위해 신속한 초기 대응이 필요하다. 따라서, 본 연구에서는 수환경 화학사고 대응 시스템 구축을 위해 하천 실시간 모니터링 지점에서 관측된 유해화학물질의 농도 자료를 이용하여 발생원의 유입 지점과 유입량을 역추적하는 프레임워크를 개발하였다. 본 연구에서 제시하는 프레임워크는 첫 번째로 하천 저장대 모형(Transient Storage Zone Model; TSM)과 HEC-RAS 모형을 이용하여 다양한 유량의 수리 조건에서 화학사고 시나리오를 생성하는 단계, 두번째로 생성된 시나리오의 유입 지점과 유입량에 대한 시간-농도 곡선 (BreakThrough Curve; BTC)을 21개의 곡선특징 (BTC feature)으로 추출하는 단계, 최종적으로 재귀적 특징 선택법(Recursive Feature Elimination; RFE)을 이용하여 의사결정나무 모형, 랜덤포레스트 모형, Xgboost 모형, 선형 서포트 벡터 머신, 커널 서포트 벡터 머신 그리고 Ridge 모형에 대한 모형별 주요 특징을 학습하고 성능을 비교하여 각각 유입 위치와 유입 질량 예측에 대한 최적 모형 및 특징 조합을 제시하는 단계로 구축하였다. 또한, 현장 적용성 제고를 위해 시간-농도 곡선을 2가지 경우 (Whole BTC와 Fractured BTC)로 가정하여 기계학습 모형을 학습시켜 모의결과를 비교하였다. 제시된 프레임워크의 검증을 위해서 낙동강 지류인 감천에 적용하여 모형을 구축하고 시나리오 자료 기반 검증과 Rhodamine WT를 이용한 추적자 실험자료를 이용한 검증을 수행하였다. 기계학습 모형들의 비교 검증 결과, 각 모형은 가중항 기반과 불순도 감소량 기반 특징 중요도 산출 방식에 따라 주요 특징이 상이하게 산출되었으며, 전체 시간-농도 곡선 (WBTC)과 부분 시간-농도 곡선 (FBTC)별 최적 모형도 다르게 산출되었다. 유입 위치 정확도 및 유입 질량 예측에 대한 R2는 대부분의 모형이 90% 이상의 우수한 결과를 나타냈다.

  • PDF

Media-based Analysis of Gasoline Inventory with Korean Text Summarization (한국어 문서 요약 기법을 활용한 휘발유 재고량에 대한 미디어 분석)

  • Sungyeon Yoon;Minseo Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.5
    • /
    • pp.509-515
    • /
    • 2023
  • Despite the continued development of alternative energies, fuel consumption is increasing. In particular, the price of gasoline fluctuates greatly according to fluctuations in international oil prices. Gas stations adjust their gasoline inventory to respond to gasoline price fluctuations. In this study, news datasets is used to analyze the gasoline consumption patterns through fluctuations of the gasoline inventory. First, collecting news datasets with web crawling. Second, summarizing news datasets using KoBART, which summarizes the Korean text datasets. Finally, preprocessing and deriving the fluctuations factors through N-Gram Language Model and TF-IDF. Through this study, it is possible to analyze and predict gasoline consumption patterns.

A New Approach to Automatic Keyword Generation Using Inverse Vector Space Model (키워드 자동 생성에 대한 새로운 접근법: 역 벡터공간모델을 이용한 키워드 할당 방법)

  • Cho, Won-Chin;Rho, Sang-Kyu;Yun, Ji-Young Agnes;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.21 no.1
    • /
    • pp.103-122
    • /
    • 2011
  • Recently, numerous documents have been made available electronically. Internet search engines and digital libraries commonly return query results containing hundreds or even thousands of documents. In this situation, it is virtually impossible for users to examine complete documents to determine whether they might be useful for them. For this reason, some on-line documents are accompanied by a list of keywords specified by the authors in an effort to guide the users by facilitating the filtering process. In this way, a set of keywords is often considered a condensed version of the whole document and therefore plays an important role for document retrieval, Web page retrieval, document clustering, summarization, text mining, and so on. Since many academic journals ask the authors to provide a list of five or six keywords on the first page of an article, keywords are most familiar in the context of journal articles. However, many other types of documents could not benefit from the use of keywords, including Web pages, email messages, news reports, magazine articles, and business papers. Although the potential benefit is large, the implementation itself is the obstacle; manually assigning keywords to all documents is a daunting task, or even impractical in that it is extremely tedious and time-consuming requiring a certain level of domain knowledge. Therefore, it is highly desirable to automate the keyword generation process. There are mainly two approaches to achieving this aim: keyword assignment approach and keyword extraction approach. Both approaches use machine learning methods and require, for training purposes, a set of documents with keywords already attached. In the former approach, there is a given set of vocabulary, and the aim is to match them to the texts. In other words, the keywords assignment approach seeks to select the words from a controlled vocabulary that best describes a document. Although this approach is domain dependent and is not easy to transfer and expand, it can generate implicit keywords that do not appear in a document. On the other hand, in the latter approach, the aim is to extract keywords with respect to their relevance in the text without prior vocabulary. In this approach, automatic keyword generation is treated as a classification task, and keywords are commonly extracted based on supervised learning techniques. Thus, keyword extraction algorithms classify candidate keywords in a document into positive or negative examples. Several systems such as Extractor and Kea were developed using keyword extraction approach. Most indicative words in a document are selected as keywords for that document and as a result, keywords extraction is limited to terms that appear in the document. Therefore, keywords extraction cannot generate implicit keywords that are not included in a document. According to the experiment results of Turney, about 64% to 90% of keywords assigned by the authors can be found in the full text of an article. Inversely, it also means that 10% to 36% of the keywords assigned by the authors do not appear in the article, which cannot be generated through keyword extraction algorithms. Our preliminary experiment result also shows that 37% of keywords assigned by the authors are not included in the full text. This is the reason why we have decided to adopt the keyword assignment approach. In this paper, we propose a new approach for automatic keyword assignment namely IVSM(Inverse Vector Space Model). The model is based on a vector space model. which is a conventional information retrieval model that represents documents and queries by vectors in a multidimensional space. IVSM generates an appropriate keyword set for a specific document by measuring the distance between the document and the keyword sets. The keyword assignment process of IVSM is as follows: (1) calculating the vector length of each keyword set based on each keyword weight; (2) preprocessing and parsing a target document that does not have keywords; (3) calculating the vector length of the target document based on the term frequency; (4) measuring the cosine similarity between each keyword set and the target document; and (5) generating keywords that have high similarity scores. Two keyword generation systems were implemented applying IVSM: IVSM system for Web-based community service and stand-alone IVSM system. Firstly, the IVSM system is implemented in a community service for sharing knowledge and opinions on current trends such as fashion, movies, social problems, and health information. The stand-alone IVSM system is dedicated to generating keywords for academic papers, and, indeed, it has been tested through a number of academic papers including those published by the Korean Association of Shipping and Logistics, the Korea Research Academy of Distribution Information, the Korea Logistics Society, the Korea Logistics Research Association, and the Korea Port Economic Association. We measured the performance of IVSM by the number of matches between the IVSM-generated keywords and the author-assigned keywords. According to our experiment, the precisions of IVSM applied to Web-based community service and academic journals were 0.75 and 0.71, respectively. The performance of both systems is much better than that of baseline systems that generate keywords based on simple probability. Also, IVSM shows comparable performance to Extractor that is a representative system of keyword extraction approach developed by Turney. As electronic documents increase, we expect that IVSM proposed in this paper can be applied to many electronic documents in Web-based community and digital library.

Analysis of ICT Education Trends using Keyword Occurrence Frequency Analysis and CONCOR Technique (키워드 출현 빈도 분석과 CONCOR 기법을 이용한 ICT 교육 동향 분석)

  • Youngseok Lee
    • Journal of Industrial Convergence
    • /
    • v.21 no.1
    • /
    • pp.187-192
    • /
    • 2023
  • In this study, trends in ICT education were investigated by analyzing the frequency of appearance of keywords related to machine learning and using conversion of iteration correction(CONCOR) techniques. A total of 304 papers from 2018 to the present published in registered sites were searched on Google Scalar using "ICT education" as the keyword, and 60 papers pertaining to ICT education were selected based on a systematic literature review. Subsequently, keywords were extracted based on the title and summary of the paper. For word frequency and indicator data, 49 keywords with high appearance frequency were extracted by analyzing frequency, via the term frequency-inverse document frequency technique in natural language processing, and words with simultaneous appearance frequency. The relationship degree was verified by analyzing the connection structure and centrality of the connection degree between words, and a cluster composed of words with similarity was derived via CONCOR analysis. First, "education," "research," "result," "utilization," and "analysis" were analyzed as main keywords. Second, by analyzing an N-GRAM network graph with "education" as the keyword, "curriculum" and "utilization" were shown to exhibit the highest correlation level. Third, by conducting a cluster analysis with "education" as the keyword, five groups were formed: "curriculum," "programming," "student," "improvement," and "information." These results indicate that practical research necessary for ICT education can be conducted by analyzing ICT education trends and identifying trends.

Analysis of teaching and learning contents of matrix in German high school mathematics (독일 고등학교 수학에서 행렬 교수·학습 내용 분석)

  • Ahn, Eunkyung;Ko, Ho Kyoung
    • The Mathematical Education
    • /
    • v.62 no.2
    • /
    • pp.269-287
    • /
    • 2023
  • Matrix theory is widely used not only in mathematics, natural sciences, and engineering, but also in social sciences and artificial intelligence. In the 2009 revised mathematics curriculum, matrices were removed from high school math education to reduce the burden on students, but in anticipation of the age of artificial intelligence, they will be reintegrated into the 2022 revised education curriculum. Therefore, there is a need to analyze the matrix content covered in other countries to suggest a meaningful direction for matrix education and to derive implications for textbook composition. In this study, we analyzed the German mathematics curriculum and standard education curriculum, as well as the matrix units in the German Hesse state mathematics curriculum and textbook, and identified the characteristics of their content elements and development methods. As a result of our analysis, it was found that the German textbooks cover matrices in three categories: matrices for solving linear equations, matrices for explaining linear transformations, and matrices for explaining transition processes. It was also found that the emphasis was on mathematical reasoning and modeling when learning matrices. Based on these findings, we suggest that if matrices are to be reintegrated into school mathematics, the curriculum should focus on deep conceptual understanding, mathematical reasoning, and mathematical modeling in textbook composition.

A New Bias Scheduling Method for Improving Both Classification Performance and Precision on the Classification and Regression Problems (분류 및 회귀문제에서의 분류 성능과 정확도를 동시에 향상시키기 위한 새로운 바이어스 스케줄링 방법)

  • Kim Eun-Mi;Park Seong-Mi;Kim Kwang-Hee;Lee Bae-Ho
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.11
    • /
    • pp.1021-1028
    • /
    • 2005
  • The general solution for classification and regression problems can be found by matching and modifying matrices with the information in real world and then these matrices are teaming in neural networks. This paper treats primary space as a real world, and dual space that Primary space matches matrices using kernel. In practical study, there are two kinds of problems, complete system which can get an answer using inverse matrix and ill-posed system or singular system which cannot get an answer directly from inverse of the given matrix. Further more the problems are often given by the latter condition; therefore, it is necessary to find regularization parameter to change ill-posed or singular problems into complete system. This paper compares each performance under both classification and regression problems among GCV, L-Curve, which are well known for getting regularization parameter, and kernel methods. Both GCV and L-Curve have excellent performance to get regularization parameters, and the performances are similar although they show little bit different results from the different condition of problems. However, these methods are two-step solution because both have to calculate the regularization parameters to solve given problems, and then those problems can be applied to other solving methods. Compared with UV and L-Curve, kernel methods are one-step solution which is simultaneously teaming a regularization parameter within the teaming process of pattern weights. This paper also suggests dynamic momentum which is leaning under the limited proportional condition between learning epoch and the performance of given problems to increase performance and precision for regularization. Finally, this paper shows the results that suggested solution can get better or equivalent results compared with GCV and L-Curve through the experiments using Iris data which are used to consider standard data in classification, Gaussian data which are typical data for singular system, and Shaw data which is an one-dimension image restoration problems.