• Title/Summary/Keyword: 데이터 기반 의사결정

Search Result 783, Processing Time 0.035 seconds

A Study on the Effect of Network Centralities on Recommendation Performance (네트워크 중심성 척도가 추천 성능에 미치는 영향에 대한 연구)

  • Lee, Dongwon
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.23-46
    • /
    • 2021
  • Collaborative filtering, which is often used in personalization recommendations, is recognized as a very useful technique to find similar customers and recommend products to them based on their purchase history. However, the traditional collaborative filtering technique has raised the question of having difficulty calculating the similarity for new customers or products due to the method of calculating similaritiesbased on direct connections and common features among customers. For this reason, a hybrid technique was designed to use content-based filtering techniques together. On the one hand, efforts have been made to solve these problems by applying the structural characteristics of social networks. This applies a method of indirectly calculating similarities through their similar customers placed between them. This means creating a customer's network based on purchasing data and calculating the similarity between the two based on the features of the network that indirectly connects the two customers within this network. Such similarity can be used as a measure to predict whether the target customer accepts recommendations. The centrality metrics of networks can be utilized for the calculation of these similarities. Different centrality metrics have important implications in that they may have different effects on recommended performance. In this study, furthermore, the effect of these centrality metrics on the performance of recommendation may vary depending on recommender algorithms. In addition, recommendation techniques using network analysis can be expected to contribute to increasing recommendation performance even if they apply not only to new customers or products but also to entire customers or products. By considering a customer's purchase of an item as a link generated between the customer and the item on the network, the prediction of user acceptance of recommendation is solved as a prediction of whether a new link will be created between them. As the classification models fit the purpose of solving the binary problem of whether the link is engaged or not, decision tree, k-nearest neighbors (KNN), logistic regression, artificial neural network, and support vector machine (SVM) are selected in the research. The data for performance evaluation used order data collected from an online shopping mall over four years and two months. Among them, the previous three years and eight months constitute social networks composed of and the experiment was conducted by organizing the data collected into the social network. The next four months' records were used to train and evaluate recommender models. Experiments with the centrality metrics applied to each model show that the recommendation acceptance rates of the centrality metrics are different for each algorithm at a meaningful level. In this work, we analyzed only four commonly used centrality metrics: degree centrality, betweenness centrality, closeness centrality, and eigenvector centrality. Eigenvector centrality records the lowest performance in all models except support vector machines. Closeness centrality and betweenness centrality show similar performance across all models. Degree centrality ranking moderate across overall models while betweenness centrality always ranking higher than degree centrality. Finally, closeness centrality is characterized by distinct differences in performance according to the model. It ranks first in logistic regression, artificial neural network, and decision tree withnumerically high performance. However, it only records very low rankings in support vector machine and K-neighborhood with low-performance levels. As the experiment results reveal, in a classification model, network centrality metrics over a subnetwork that connects the two nodes can effectively predict the connectivity between two nodes in a social network. Furthermore, each metric has a different performance depending on the classification model type. This result implies that choosing appropriate metrics for each algorithm can lead to achieving higher recommendation performance. In general, betweenness centrality can guarantee a high level of performance in any model. It would be possible to consider the introduction of proximity centrality to obtain higher performance for certain models.

A Study on Automatic Classification Model of Documents Based on Korean Standard Industrial Classification (한국표준산업분류를 기준으로 한 문서의 자동 분류 모델에 관한 연구)

  • Lee, Jae-Seong;Jun, Seung-Pyo;Yoo, Hyoung Sun
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.221-241
    • /
    • 2018
  • As we enter the knowledge society, the importance of information as a new form of capital is being emphasized. The importance of information classification is also increasing for efficient management of digital information produced exponentially. In this study, we tried to automatically classify and provide tailored information that can help companies decide to make technology commercialization. Therefore, we propose a method to classify information based on Korea Standard Industry Classification (KSIC), which indicates the business characteristics of enterprises. The classification of information or documents has been largely based on machine learning, but there is not enough training data categorized on the basis of KSIC. Therefore, this study applied the method of calculating similarity between documents. Specifically, a method and a model for presenting the most appropriate KSIC code are proposed by collecting explanatory texts of each code of KSIC and calculating the similarity with the classification object document using the vector space model. The IPC data were collected and classified by KSIC. And then verified the methodology by comparing it with the KSIC-IPC concordance table provided by the Korean Intellectual Property Office. As a result of the verification, the highest agreement was obtained when the LT method, which is a kind of TF-IDF calculation formula, was applied. At this time, the degree of match of the first rank matching KSIC was 53% and the cumulative match of the fifth ranking was 76%. Through this, it can be confirmed that KSIC classification of technology, industry, and market information that SMEs need more quantitatively and objectively is possible. In addition, it is considered that the methods and results provided in this study can be used as a basic data to help the qualitative judgment of experts in creating a linkage table between heterogeneous classification systems.

Development of Estimation System for Housing Remodeling Cost through Influence Analysis by Design Elements (설계요소별 영향분석을 통한 공동주택 리모델링 공사비개산견적 산출 시스템 개발)

  • Kim, Jun;Cha, Heesung
    • Korean Journal of Construction Engineering and Management
    • /
    • v.19 no.6
    • /
    • pp.65-78
    • /
    • 2018
  • In As urban apartment are aging, the necessity of reconstruction or remodeling to extend the life of buildings is increasing. In such a case, a co-housing association is formed to implement decisions on reconstruction or remodeling projects. At this time, the most important thing for the co-housing association is the business feasibility based on the input of the construction cost.In the case of reconstruction, it is possible to estimate the construction cost by using the accumulated construction cost data, and then evaluate the feasibility using the construction cost. However, in case of remodeling, it is difficult to calculate the accurate construction cost because the number of accumulated construction cost data is small. In addition, non-specialist clients often require estimates of various design factors, often negatively impacting the accuracy of estimates and the duration of estimates. Therefore, in this study, proposed method to reflect the opinion of the owner who is a non-expert, as a design element, and a method of calculating the expected construction cost according to the design element, and constructed this system so that it can be easily used by the non-specialist owner. In order to clearly reflect the requirements of the non-specialist owner in the estimates, extracts the design elements from the existing remodeling cases, classify them, and suggest a plan for the client to choose. In order to reflect the design factors to the estimates, the existing apartment house remodeling cases were investigated and the design factors were extracted to have a large effect on the construction cost. Finally, developed system based on MS Excel so that the above contents can be easily used by a non-specialist client. In order to verify the accuracy of the proposed estimate in this study, verified the accuracy of 80% of the results by substituting the case of remodeling quotations and obtained a positive result from the questionnaire survey to examine the ease of use of the non-specialist customer. In this study, propose an estimate estimation method using four cases. If the remodeling cases are accumulated continuously, the expected effect of this study will be higher.

Development of a Failure Probability Model based on Operation Data of Thermal Piping Network in District Heating System (지역난방 열배관망 운영데이터 기반의 파손확률 모델 개발)

  • Kim, Hyoung Seok;Kim, Gye Beom;Kim, Lae Hyun
    • Korean Chemical Engineering Research
    • /
    • v.55 no.3
    • /
    • pp.322-331
    • /
    • 2017
  • District heating was first introduced in Korea in 1985. As the service life of the underground thermal piping network has increased for more than 30 years, the maintenance of the underground thermal pipe has become an important issue. A variety of complex technologies are required for periodic inspection and operation management for the maintenance of the aged thermal piping network. Especially, it is required to develop a model that can be used for decision making in order to derive optimal maintenance and replacement point from the economic viewpoint in the field. In this study, the analysis was carried out based on the repair history and accident data at the operation of the thermal pipe network of five districts in the Korea District Heating Corporation. A failure probability model was developed by introducing statistical techniques of qualitative analysis and binomial logistic regression analysis. As a result of qualitative analysis of maintenance history and accident data, the most important cause of pipeline damage was construction erosion, corrosion of pipe and bad material accounted for about 82%. In the statistical model analysis, by setting the separation point of the classification to 0.25, the accuracy of the thermal pipe breakage and non-breakage classification improved to 73.5%. In order to establish the failure probability model, the fitness of the model was verified through the Hosmer and Lemeshow test, the independent test of the independent variables, and the Chi-Square test of the model. According to the results of analysis of the risk of thermal pipe network damage, the highest probability of failure was analyzed as the thermal pipeline constructed by the F construction company in the reducer pipe of less than 250mm, which is more than 10 years on the Seoul area motorway in winter. The results of this study can be used to prioritize maintenance, preventive inspection, and replacement of thermal piping systems. In addition, it will be possible to reduce the frequency of thermal pipeline damage and to use it more aggressively to manage thermal piping network by establishing and coping with accident prevention plan in advance such as inspection and maintenance.

An Analysis of the Dynamics between Media Coverage and Stock Market on Digital New Deal Policy: Focusing on Companies Related to the Fourth Industrial Revolution (디지털 뉴딜 정책에 대한 언론 보도량과 주식 시장의 동태적 관계 분석: 4차산업혁명 관련 기업을 중심으로)

  • Sohn, Kwonsang;Kwon, Ohbyung
    • The Journal of Society for e-Business Studies
    • /
    • v.26 no.3
    • /
    • pp.33-53
    • /
    • 2021
  • In the crossroads of social change caused by the spread of the Fourth Industrial Revolution and the prolonged COVID-19, the Korean government announced the Digital New Deal policy on July 14, 2020. The Digital New Deal policy's primary goal is to create new businesses by accelerating digital transformation in the public sector and industries around data, networks, and artificial intelligence technologies. However, in a rapidly changing social environment, information asymmetry of the future benefits of technology can cause differences in the public's ability to analyze the direction and effectiveness of policies, resulting in uncertainty about the practical effects of policies. On the other hand, the media leads the formation of discourse through communicators' role to disseminate government policies to the public and provides knowledge about specific issues through the news. In other words, as the media coverage of a particular policy increases, the issue concentration increases, which also affects public decision-making. Therefore, the purpose of this study is to verify the dynamic relationship between the media coverage and the stock market on the Korean government's digital New Deal policy using Granger causality, impulse response functions, and variance decomposition analysis. To this end, the daily stock turnover ratio, daily price-earnings ratio, and EWMA volatility of digital technology-based companies related to the digital new deal policy among KOSDAQ listed companies were set as variables. As a result, keyword search volume, daily stock turnover ratio, EWMA volatility have a bi-directional Granger causal relationship with media coverage. And an increase in media coverage has a high impact on keyword search volume on digital new deal policies. Also, the impulse response analysis on media coverage showed a sharp drop in EWMA volatility. The influence gradually increased over time and played a role in mitigating stock market volatility. Based on this study's findings, the amount of media coverage of digital new deals policy has a significant dynamic relationship with the stock market.

Using GA based Input Selection Method for Artificial Neural Network Modeling Application to Bankruptcy Prediction (유전자 알고리즘을 활용한 인공신경망 모형 최적입력변수의 선정: 부도예측 모형을 중심으로)

  • 홍승현;신경식
    • Journal of Intelligence and Information Systems
    • /
    • v.9 no.1
    • /
    • pp.227-249
    • /
    • 2003
  • Prediction of corporate failure using past financial data is a well-documented topic. Early studies of bankruptcy prediction used statistical techniques such as multiple discriminant analysis, logit and probit. Recently, however, numerous studies have demonstrated that artificial intelligence such as neural networks can be an alternative methodology for classification problems to which traditional statistical methods have long been applied. In building neural network model, the selection of independent and dependent variables should be approached with great care and should be treated as model construction process. Irrespective of the efficiency of a teaming procedure in terms of convergence, generalization and stability, the ultimate performance of the estimator will depend on the relevance of the selected input variables and the quality of the data used. Approaches developed in statistical methods such as correlation analysis and stepwise selection method are often very useful. These methods, however, may not be the optimal ones for the development of neural network model. In this paper, we propose a genetic algorithms approach to find an optimal or near optimal input variables fur neural network modeling. The proposed approach is demonstrated by applications to bankruptcy prediction modeling. Our experimental results show that this approach increases overall classification accuracy rate significantly.

  • PDF

Development of GIS-based EEZ Marine Resources Information System (GIS를 이용한 배타적 경제수역 해양자원정보시스템의 구현에 관한 연구)

  • Kim, Kye-Hyun;Kim, Sun-Yong;Park, Eun-Ji;Yoo, Hai-Soo
    • Journal of Korea Spatial Information System Society
    • /
    • v.9 no.2
    • /
    • pp.55-66
    • /
    • 2007
  • There has been increasing concerns regarding marine mineral resources as the land energy resources has been depleting from worldwide energy crisis. Also, all the coastal countries around the world are getting into the high competition as EEZ implemented to widen each country's marine autonomy. Especially, the adoption of EEZ in UN's marine regulation agreement at the January of 1996 has aggravated conflicts among coastal countries and eventually resulted in critical agenda to determine the boundaries of EEZ among such countries. It is imperative for us to have negotiation with neighboring countries to determine the boundaries of EEZ. For the preparation of such negotiation, it is essential to have data such as mineral distribution, deep-sea geology, related agreement and marine laws, etc. Therefore, this study mainly concentrates on analyzing existing data of resources exploration and establishing standards for each type of data and manipulating data based on such standards, thereby building a database for more efficient management of EEZ data from marine resources survey. MRIS has also been developed to diversely analyze and visualize graphic and attribute data considering data usage and inter-relationship in the database. This system can provide various spatial analysis and spatial searching techniques to enable easier comparison of cost-benefit analysis and data provision of any area in EEZ thereby facilitating major policy making. In addition, the system can support sustainable management of marine resources of EEZ regions and data supply for systematic management of national marine resources. Furthermore, this will be very useful for negotiating with neighboring countries to determine EEZ boundaries to lead more favorable results.

  • PDF

An Analysis of ICT-Retail Convergence(IRC) and Consumer Value Creation (소비자 구매단계별 기술-유통 통합(IRC)과 가치에 대한 연구)

  • Park, Sunny;Cho, Eunsun;Rha, Jong-Youn;Lee, Yuri;Kim, Suyoun
    • Journal of Digital Convergence
    • /
    • v.15 no.7
    • /
    • pp.147-157
    • /
    • 2017
  • Recently, ICT Retail Convergence(IRC) has been rapidly increasing to improve consumer satisfaction and consumer experience. In this paper, we aim to diagnose IRC from consumers' point of view by reviewing the present status and value of IRC according to consumer purchase decision making process. Based on the previous studies in retail industry, we classified IRC into 4 types: Experience-specific tech(Virtual Reality and Augmented Reality); Information-specific tech(Artificial Intelligence and Big Data); Location-based tech(Radio Frequency Identification and Beacon); Payment-related tech(Fin-tech and Biometrics). Next, we found that there is a difference in value provided to consumers according to the type of technology, analysing the value by consumer purchase decision making process. This study can be useful to introduce IRC for improving consumer satisfaction as well as ICT and Retail. Also, it can be basic data for future technology studies with a consumer perspective.

The Design of a Complex Event Model for Effective Service Monitoring in Enterprise Systems (엔터프라이즈 시스템에서 효과적인 서비스 모니터링을 위한 복합 이벤트 모델의 설계)

  • Kum, Deuk-Kyu;Lee, Nam-Yong
    • The KIPS Transactions:PartD
    • /
    • v.18D no.4
    • /
    • pp.261-274
    • /
    • 2011
  • In recent competitive business environment each enterprise has to be agile and flexible. For these purposes run-time monitoring ofservices provided by an enterprise and early decision making through this becomes core competition of the enterprise. In addition, in order to process various innumerable events which are generated on enterprise systems techniques which make filtering of meaningful data are needed. However, the existing study related with this is nothing but discovering of service faults by monitoring depending upon API of BPEL engine or middleware, or is nothing but processing of simple events based on low-level events. Accordingly, there would be limitations to provide useful business information. In this paper, through situation detection an extended complex event model is presented, which is possible to provide more valuable and useful business information. Concretely, first of all an event processing architecture in an enterprise system is proposed, and event meta-model which is suitable to the proposed architecture is going to be defined. Based on the defined meta-model, It is presented that syntax and semantics of constructs in our event processing language including various and progressive event operators, complex event pattern, key, etc. In addition, an event context mechanism is proposed to analyze more delicate events. Finally, through application studies application possibility of this study would be shown and merits of this event model would be present through comparison with other event model.

Artificial Intelligence Techniques for Predicting Online Peer-to-Peer(P2P) Loan Default (인공지능기법을 이용한 온라인 P2P 대출거래의 채무불이행 예측에 관한 실증연구)

  • Bae, Jae Kwon;Lee, Seung Yeon;Seo, Hee Jin
    • The Journal of Society for e-Business Studies
    • /
    • v.23 no.3
    • /
    • pp.207-224
    • /
    • 2018
  • In this article, an empirical study was conducted by using public dataset from Lending Club Corporation, the largest online peer-to-peer (P2P) lending in the world. We explore significant predictor variables related to P2P lending default that housing situation, length of employment, average current balance, debt-to-income ratio, loan amount, loan purpose, interest rate, public records, number of finance trades, total credit/credit limit, number of delinquent accounts, number of mortgage accounts, and number of bank card accounts are significant factors to loan funded successful on Lending Club platform. We developed online P2P lending default prediction models using discriminant analysis, logistic regression, neural networks, and decision trees (i.e., CART and C5.0) in order to predict P2P loan default. To verify the feasibility and effectiveness of P2P lending default prediction models, borrower loan data and credit data used in this study. Empirical results indicated that neural networks outperforms other classifiers such as discriminant analysis, logistic regression, CART, and C5.0. Neural networks always outperforms other classifiers in P2P loan default prediction.