• Title/Summary/Keyword: Intelligent Data Analysis

Search Result 1,456, Processing Time 0.034 seconds

Empirical Analysis on Bitcoin Price Change by Consumer, Industry and Macro-Economy Variables (비트코인 가격 변화에 관한 실증분석: 소비자, 산업, 그리고 거시변수를 중심으로)

  • Lee, Junsik;Kim, Keon-Woo;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.195-220
    • /
    • 2018
  • In this study, we conducted an empirical analysis of the factors that affect the change of Bitcoin Closing Price. Previous studies have focused on the security of the block chain system, the economic ripple effects caused by the cryptocurrency, legal implications and the acceptance to consumer about cryptocurrency. In various area, cryptocurrency was studied and many researcher and people including government, regardless of country, try to utilize cryptocurrency and applicate to its technology. Despite of rapid and dramatic change of cryptocurrencies' price and growth of its effects, empirical study of the factors affecting the price change of cryptocurrency was lack. There were only a few limited studies, business reports and short working paper. Therefore, it is necessary to determine what factors effect on the change of closing Bitcoin price. For analysis, hypotheses were constructed from three dimensions of consumer, industry, and macroeconomics for analysis, and time series data were collected for variables of each dimension. Consumer variables consist of search traffic of Bitcoin, search traffic of bitcoin ban, search traffic of ransomware and search traffic of war. Industry variables were composed GPU vendors' stock price and memory vendors' stock price. Macro-economy variables were contemplated such as U.S. dollar index futures, FOMC policy interest rates, WTI crude oil price. Using above variables, we did times series regression analysis to find relationship between those variables and change of Bitcoin Closing Price. Before the regression analysis to confirm the relationship between change of Bitcoin Closing Price and the other variables, we performed the Unit-root test to verifying the stationary of time series data to avoid spurious regression. Then, using a stationary data, we did the regression analysis. As a result of the analysis, we found that the change of Bitcoin Closing Price has negative effects with search traffic of 'Bitcoin Ban' and US dollar index futures, while change of GPU vendors' stock price and change of WTI crude oil price showed positive effects. In case of 'Bitcoin Ban', it is directly determining the maintenance or abolition of Bitcoin trade, that's why consumer reacted sensitively and effected on change of Bitcoin Closing Price. GPU is raw material of Bitcoin mining. Generally, increasing of companies' stock price means the growth of the sales of those companies' products and services. GPU's demands increases are indirectly reflected to the GPU vendors' stock price. Making an interpretation, a rise in prices of GPU has put a crimp on the mining of Bitcoin. Consequently, GPU vendors' stock price effects on change of Bitcoin Closing Price. And we confirmed U.S. dollar index futures moved in the opposite direction with change of Bitcoin Closing Price. It moved like Gold. Gold was considered as a safe asset to consumers and it means consumer think that Bitcoin is a safe asset. On the other hand, WTI oil price went Bitcoin Closing Price's way. It implies that Bitcoin are regarded to investment asset like raw materials market's product. The variables that were not significant in the analysis were search traffic of bitcoin, search traffic of ransomware, search traffic of war, memory vendor's stock price, FOMC policy interest rates. In search traffic of bitcoin, we judged that interest in Bitcoin did not lead to purchase of Bitcoin. It means search traffic of Bitcoin didn't reflect all of Bitcoin's demand. So, it implies there are some factors that regulate and mediate the Bitcoin purchase. In search traffic of ransomware, it is hard to say concern of ransomware determined the whole Bitcoin demand. Because only a few people damaged by ransomware and the percentage of hackers requiring Bitcoins was low. Also, its information security problem is events not continuous issues. Search traffic of war was not significant. Like stock market, generally it has negative in relation to war, but exceptional case like Gulf war, it moves stakeholders' profits and environment. We think that this is the same case. In memory vendor stock price, this is because memory vendors' flagship products were not VRAM which is essential for Bitcoin supply. In FOMC policy interest rates, when the interest rate is low, the surplus capital is invested in securities such as stocks. But Bitcoin' price fluctuation was large so it is not recognized as an attractive commodity to the consumers. In addition, unlike the stock market, Bitcoin doesn't have any safety policy such as Circuit breakers and Sidecar. Through this study, we verified what factors effect on change of Bitcoin Closing Price, and interpreted why such change happened. In addition, establishing the characteristics of Bitcoin as a safe asset and investment asset, we provide a guide how consumer, financial institution and government organization approach to the cryptocurrency. Moreover, corroborating the factors affecting change of Bitcoin Closing Price, researcher will get some clue and qualification which factors have to be considered in hereafter cryptocurrency study.

Building battery deterioration prediction model using real field data (머신러닝 기법을 이용한 납축전지 열화 예측 모델 개발)

  • Choi, Keunho;Kim, Gunwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.243-264
    • /
    • 2018
  • Although the worldwide battery market is recently spurring the development of lithium secondary battery, lead acid batteries (rechargeable batteries) which have good-performance and can be reused are consumed in a wide range of industry fields. However, lead-acid batteries have a serious problem in that deterioration of a battery makes progress quickly in the presence of that degradation of only one cell among several cells which is packed in a battery begins. To overcome this problem, previous researches have attempted to identify the mechanism of deterioration of a battery in many ways. However, most of previous researches have used data obtained in a laboratory to analyze the mechanism of deterioration of a battery but not used data obtained in a real world. The usage of real data can increase the feasibility and the applicability of the findings of a research. Therefore, this study aims to develop a model which predicts the battery deterioration using data obtained in real world. To this end, we collected data which presents change of battery state by attaching sensors enabling to monitor the battery condition in real time to dozens of golf carts operated in the real golf field. As a result, total 16,883 samples were obtained. And then, we developed a model which predicts a precursor phenomenon representing deterioration of a battery by analyzing the data collected from the sensors using machine learning techniques. As initial independent variables, we used 1) inbound time of a cart, 2) outbound time of a cart, 3) duration(from outbound time to charge time), 4) charge amount, 5) used amount, 6) charge efficiency, 7) lowest temperature of battery cell 1 to 6, 8) lowest voltage of battery cell 1 to 6, 9) highest voltage of battery cell 1 to 6, 10) voltage of battery cell 1 to 6 at the beginning of operation, 11) voltage of battery cell 1 to 6 at the end of charge, 12) used amount of battery cell 1 to 6 during operation, 13) used amount of battery during operation(Max-Min), 14) duration of battery use, and 15) highest current during operation. Since the values of the independent variables, lowest temperature of battery cell 1 to 6, lowest voltage of battery cell 1 to 6, highest voltage of battery cell 1 to 6, voltage of battery cell 1 to 6 at the beginning of operation, voltage of battery cell 1 to 6 at the end of charge, and used amount of battery cell 1 to 6 during operation are similar to that of each battery cell, we conducted principal component analysis using verimax orthogonal rotation in order to mitigate the multiple collinearity problem. According to the results, we made new variables by averaging the values of independent variables clustered together, and used them as final independent variables instead of origin variables, thereby reducing the dimension. We used decision tree, logistic regression, Bayesian network as algorithms for building prediction models. And also, we built prediction models using the bagging of each of them, the boosting of each of them, and RandomForest. Experimental results show that the prediction model using the bagging of decision tree yields the best accuracy of 89.3923%. This study has some limitations in that the additional variables which affect the deterioration of battery such as weather (temperature, humidity) and driving habits, did not considered, therefore, we would like to consider the them in the future research. However, the battery deterioration prediction model proposed in the present study is expected to enable effective and efficient management of battery used in the real filed by dramatically and to reduce the cost caused by not detecting battery deterioration accordingly.

Selective Word Embedding for Sentence Classification by Considering Information Gain and Word Similarity (문장 분류를 위한 정보 이득 및 유사도에 따른 단어 제거와 선택적 단어 임베딩 방안)

  • Lee, Min Seok;Yang, Seok Woo;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.105-122
    • /
    • 2019
  • Dimensionality reduction is one of the methods to handle big data in text mining. For dimensionality reduction, we should consider the density of data, which has a significant influence on the performance of sentence classification. It requires lots of computations for data of higher dimensions. Eventually, it can cause lots of computational cost and overfitting in the model. Thus, the dimension reduction process is necessary to improve the performance of the model. Diverse methods have been proposed from only lessening the noise of data like misspelling or informal text to including semantic and syntactic information. On top of it, the expression and selection of the text features have impacts on the performance of the classifier for sentence classification, which is one of the fields of Natural Language Processing. The common goal of dimension reduction is to find latent space that is representative of raw data from observation space. Existing methods utilize various algorithms for dimensionality reduction, such as feature extraction and feature selection. In addition to these algorithms, word embeddings, learning low-dimensional vector space representations of words, that can capture semantic and syntactic information from data are also utilized. For improving performance, recent studies have suggested methods that the word dictionary is modified according to the positive and negative score of pre-defined words. The basic idea of this study is that similar words have similar vector representations. Once the feature selection algorithm selects the words that are not important, we thought the words that are similar to the selected words also have no impacts on sentence classification. This study proposes two ways to achieve more accurate classification that conduct selective word elimination under specific regulations and construct word embedding based on Word2Vec embedding. To select words having low importance from the text, we use information gain algorithm to measure the importance and cosine similarity to search for similar words. First, we eliminate words that have comparatively low information gain values from the raw text and form word embedding. Second, we select words additionally that are similar to the words that have a low level of information gain values and make word embedding. In the end, these filtered text and word embedding apply to the deep learning models; Convolutional Neural Network and Attention-Based Bidirectional LSTM. This study uses customer reviews on Kindle in Amazon.com, IMDB, and Yelp as datasets, and classify each data using the deep learning models. The reviews got more than five helpful votes, and the ratio of helpful votes was over 70% classified as helpful reviews. Also, Yelp only shows the number of helpful votes. We extracted 100,000 reviews which got more than five helpful votes using a random sampling method among 750,000 reviews. The minimal preprocessing was executed to each dataset, such as removing numbers and special characters from text data. To evaluate the proposed methods, we compared the performances of Word2Vec and GloVe word embeddings, which used all the words. We showed that one of the proposed methods is better than the embeddings with all the words. By removing unimportant words, we can get better performance. However, if we removed too many words, it showed that the performance was lowered. For future research, it is required to consider diverse ways of preprocessing and the in-depth analysis for the co-occurrence of words to measure similarity values among words. Also, we only applied the proposed method with Word2Vec. Other embedding methods such as GloVe, fastText, ELMo can be applied with the proposed methods, and it is possible to identify the possible combinations between word embedding methods and elimination methods.

An Optimized Combination of π-fuzzy Logic and Support Vector Machine for Stock Market Prediction (주식 시장 예측을 위한 π-퍼지 논리와 SVM의 최적 결합)

  • Dao, Tuanhung;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.43-58
    • /
    • 2014
  • As the use of trading systems has increased rapidly, many researchers have become interested in developing effective stock market prediction models using artificial intelligence techniques. Stock market prediction involves multifaceted interactions between market-controlling factors and unknown random processes. A successful stock prediction model achieves the most accurate result from minimum input data with the least complex model. In this research, we develop a combination model of ${\pi}$-fuzzy logic and support vector machine (SVM) models, using a genetic algorithm to optimize the parameters of the SVM and ${\pi}$-fuzzy functions, as well as feature subset selection to improve the performance of stock market prediction. To evaluate the performance of our proposed model, we compare the performance of our model to other comparative models, including the logistic regression, multiple discriminant analysis, classification and regression tree, artificial neural network, SVM, and fuzzy SVM models, with the same data. The results show that our model outperforms all other comparative models in prediction accuracy as well as return on investment.

Automatic TV Program Recommendation using LDA based Latent Topic Inference (LDA 기반 은닉 토픽 추론을 이용한 TV 프로그램 자동 추천)

  • Kim, Eun-Hui;Pyo, Shin-Jee;Kim, Mun-Churl
    • Journal of Broadcast Engineering
    • /
    • v.17 no.2
    • /
    • pp.270-283
    • /
    • 2012
  • With the advent of multi-channel TV, IPTV and smart TV services, excessive amounts of TV program contents become available at users' sides, which makes it very difficult for TV viewers to easily find and consume their preferred TV programs. Therefore, the service of automatic TV recommendation is an important issue for TV users for future intelligent TV services, which allows to improve access to their preferred TV contents. In this paper, we present a recommendation model based on statistical machine learning using a collaborative filtering concept by taking in account both public and personal preferences on TV program contents. For this, users' preference on TV programs is modeled as a latent topic variable using LDA (Latent Dirichlet Allocation) which is recently applied in various application domains. To apply LDA for TV recommendation appropriately, TV viewers's interested topics is regarded as latent topics in LDA, and asymmetric Dirichlet distribution is applied on the LDA which can reveal the diversity of the TV viewers' interests on topics based on the analysis of the real TV usage history data. The experimental results show that the proposed LDA based TV recommendation method yields average 66.5% with top 5 ranked TV programs in weekly recommendation, average 77.9% precision in bimonthly recommendation with top 5 ranked TV programs for the TV usage history data of similar taste user groups.

Traffic Vulnerability Analysis of Rural Area using Road Accessibility and Functionality in Cheongju City (도로 접근성과 기능성을 이용한 통합청주시 농촌지역의 교통 취약성 분석)

  • Jeon, Jeongbae;Oh, Hyunkyo;Park, Jinseon;Yoon, Seongsoo
    • Journal of Korean Society of Rural Planning
    • /
    • v.21 no.2
    • /
    • pp.11-21
    • /
    • 2015
  • This study carried out evaluation of vulnerability in accessability and functionality using road network that was extracted from Intelligent Transportation System(ITS) and digital map. It was built in order to figure out accessability that locational data which include community center, public facilities, medical facilities and highway IC. The method for grasping functionality are Digital Elevation Model(DEM) and land slide hazard map provided by Korea Forest Service. The evaluation criteria for figure out accessability was set to related comparison of average time in urban area. Functionality value was calculated by the possibility of backing the vehicle possibility of snowfall and landslides. At last, this research computed weighting value through Analytic Hierarchy Process (AHP), calculated a vulnerable score. As the result, the accessability of rural village came out that would spend more time by 1.4 to 3.2 times in comparison with urban area. Even though, vulnerability of the road by a snowfall was estimated that more than 50% satisfies the first class, however, it show up that the road were still vulnerable due snowing because over the 14% of the road being evaluated the fifth class. The functionality has been satisfied most of the road, however, It was vulnerable around Lake Daechung and Piban-ryung, Yumti-jae, Suriti-jae where on the way Boeun. Also, the fifth class road are about 35 km away from the city hall on distance, take an hour to an hour and a half. The fourth class road are about 25 km away from the city hall on distance, take 25 min to an hour. The other class of the road take in 30 min from the city hall or aren't affected of weather and have been analyzed that a density of road is high. In A result that compare between distribution and a housing density came out different the southern and the eastern area, so this result could be suggested quantitative data for possibility of development.

Construction Claims Prediction and Decision Awareness Framework using Artificial Neural Networks and Backward Optimization

  • Hosny, Ossama A.;Elbarkouky, Mohamed M.G.;Elhakeem, Ahmed
    • Journal of Construction Engineering and Project Management
    • /
    • v.5 no.1
    • /
    • pp.11-19
    • /
    • 2015
  • This paper presents optimized artificial neural networks (ANNs) claims prediction and decision awareness framework that guides owner organizations in their pre-bid construction project decisions to minimize claims. The framework is composed of two genetic optimization ANNs models: a Claims Impact Prediction Model (CIPM), and a Decision Awareness Model (DAM). The CIPM is composed of three separate ANNs that predict the cost and time impacts of the possible claims that may arise in a project. The models also predict the expected types of relationship between the owner and the contractor based on their behavioral and technical decisions during the bidding phase of the project. The framework is implemented using actual data from international projects in the Middle East and Egypt (projects owned by either public or private local organizations who hired international prime contractors to deliver the projects). Literature review, interviews with pertinent experts in the Middle East, and lessons learned from several international construction projects in Egypt determined the input decision variables of the CIPM. The ANNs training, which has been implemented in a spreadsheet environment, was optimized using genetic algorithm (GA). Different weights were assigned as variables to the different layers of each ANN and the total square error was used as the objective function to be minimized. Data was collected from thirty-two international construction projects in order to train and test the ANNs of the CIPM, which predicted cost overruns, schedule delays, and relationships between contracting parties. A genetic optimization backward analysis technique was then applied to develop the Decision Awareness Model (DAM). The DAM combined the three artificial neural networks of the CIPM to assist project owners in setting optimum values for their behavioral and technical decision variables. It implements an intelligent user-friendly input interface which helps project owners in visualizing the impact of their decisions on the project's total cost, original duration, and expected owner-contractor relationship. The framework presents a unique and transparent hybrid genetic algorithm-ANNs training and testing method. It has been implemented in a spreadsheet environment using MS Excel$^{(R)}$ and EVOLVERTM V.5.5. It provides projects' owners of a decision-support tool that raises their awareness regarding their pre-bid decisions for a construction project.

A Preliminary Discussion on Policy Decision Making of AI in The Fourth Industrial Revolution (4차 산업혁명시대 인공지능 정책의사결정에 대한 탐색적 논의)

  • Seo, Hyung-Jun
    • Informatization Policy
    • /
    • v.26 no.3
    • /
    • pp.3-35
    • /
    • 2019
  • In the fourth industrial revolution age, because of advance in the intelligence information technologies, the various roles of AI have attracted public attention. Starting with Google's Alphago, AI is now no longer a fantasized technology but a real one that can bring ripple effect in entire society. Already, AI has performed well in the medical service, legal service, and the private sector's business decision making. This study conducted an exploratory analysis on the possibilities and issues of AI-driven policy decision making in the public sector. The three research purposes are i) could AI make a policy decision in public sector?; ii) how different is AI-driven policy decision making compared to the existing methods of decision making?; and iii) what issues would be revealed by AI's policy decision making? AI-driven policy decision making is differentiated from the traditional ways of decision making in that the former is represented by rationality based on sufficient amount of information and alternatives, increased transparency and trust, more objective views for policy issues, and faster decision making process. However, there are several controversial issues regarding superiority of AI, ethics, accountability, changes in democracy, substitution of human labor in the public sector, and data usage problems for AI. Since the adoption of AI for policy decision making will be soon realized, it is necessary to take an integrative approach, considering both the positive and adverse effects, to minimize social impact.

Extended Adaptation Database Construction for Oriental Medicine Prescriptions Based on Academic Information (학술 정보 기반 한의학 처방을 위한 확장 적응증 데이터베이스 구축)

  • Lee, So-Min;Baek, Yeon-Hee;Song, Sang-Ho;CHRISTOPHER, RETITI DIOP EMANE;Han, Xuan-Zhong;Hong, Seong-Yeon;Kim, Ik-Su;Lim, Jong-Tea;Bok, Kyoung-Soo;TRAN, MINH NHAT;NGUYEN, QUYNH HOANG NGAN;Kim, So-Young;Kim, An-Na;Lee, Sang-Hun;Yoo, Jae-Soo
    • The Journal of the Korea Contents Association
    • /
    • v.21 no.8
    • /
    • pp.367-375
    • /
    • 2021
  • The quality of medical care can be defined as four types such as effectiveness, efficiency, adequacy, and scientific-technical quality. For the management of scientific-technical aspects, medical institutions annually disseminate the latest knowledge in the form of conservative education. However, there is an obvious limit to the fact that the latest knowledge is distributed quickly enough to the clinical site with only one-time conservative education. If intelligent information processing technologies such as big data and artificial intelligence are applied to the medical field, they can overcome the limitations of having to conduct research with only a small amount of information. In this paper, we construct databases on which the existing medicine prescription adaptations can be extended. To do this, we collect, store, manage, and analyze information related to oriental medicine at domestic and abroad Journals. We design a processing and analysis technique for oriental medicine evidence research data for the construction of a database of oriental medicine prescription extended adaption. Results can be used as a basic content of evidence-based medicine prescription information in the oriental medicine-related decision support services.

A Study on How to Set up a Standard Framework for AI Ethics and Regulation (AI 윤리와 규제에 관한 표준 프레임워크 설정 방안 연구)

  • Nam, Mun-Hee
    • Journal of the Korea Convergence Society
    • /
    • v.13 no.4
    • /
    • pp.7-15
    • /
    • 2022
  • With the aim of an intelligent world in the age of individual customization through decentralization of information and technology, sharing/opening, and connection, we often see a tendency to cross expectations and concerns in the technological discourse and interest in artificial intelligence more than ever. Recently, it is easy to find claims by futurists that AI singularity will appear before and after 2045. Now, as part of preparations to create a paradigm of coexistence that coexists and prosper with AI in the coming age of artificial intelligence, a standard framework for setting up more correct AI ethics and regulations is required. This is because excluding the risk of omission of setting major guidelines and methods for evaluating reasonable and more reasonable guideline items and evaluation standards are increasingly becoming major research issues. In order to solve these research problems and at the same time to develop continuous experiences and learning effects on AI ethics and regulation setting, we collect guideline data on AI ethics and regulation of international organizations / countries / companies, and research and suggest ways to set up a standard framework (SF: Standard Framework) through a setting research model and text mining exploratory analysis. The results of this study can be contributed as basic prior research data for more advanced AI ethics and regulatory guidelines item setting and evaluation methods in the future.