• Title/Summary/Keyword: 연관 마이닝

Search Result 486, Processing Time 0.025 seconds

Stock Price Prediction by Utilizing Category Neutral Terms: Text Mining Approach (카테고리 중립 단어 활용을 통한 주가 예측 방안: 텍스트 마이닝 활용)

  • Lee, Minsik;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.123-138
    • /
    • 2017
  • Since the stock market is driven by the expectation of traders, studies have been conducted to predict stock price movements through analysis of various sources of text data. In order to predict stock price movements, research has been conducted not only on the relationship between text data and fluctuations in stock prices, but also on the trading stocks based on news articles and social media responses. Studies that predict the movements of stock prices have also applied classification algorithms with constructing term-document matrix in the same way as other text mining approaches. Because the document contains a lot of words, it is better to select words that contribute more for building a term-document matrix. Based on the frequency of words, words that show too little frequency or importance are removed. It also selects words according to their contribution by measuring the degree to which a word contributes to correctly classifying a document. The basic idea of constructing a term-document matrix was to collect all the documents to be analyzed and to select and use the words that have an influence on the classification. In this study, we analyze the documents for each individual item and select the words that are irrelevant for all categories as neutral words. We extract the words around the selected neutral word and use it to generate the term-document matrix. The neutral word itself starts with the idea that the stock movement is less related to the existence of the neutral words, and that the surrounding words of the neutral word are more likely to affect the stock price movements. And apply it to the algorithm that classifies the stock price fluctuations with the generated term-document matrix. In this study, we firstly removed stop words and selected neutral words for each stock. And we used a method to exclude words that are included in news articles for other stocks among the selected words. Through the online news portal, we collected four months of news articles on the top 10 market cap stocks. We split the news articles into 3 month news data as training data and apply the remaining one month news articles to the model to predict the stock price movements of the next day. We used SVM, Boosting and Random Forest for building models and predicting the movements of stock prices. The stock market opened for four months (2016/02/01 ~ 2016/05/31) for a total of 80 days, using the initial 60 days as a training set and the remaining 20 days as a test set. The proposed word - based algorithm in this study showed better classification performance than the word selection method based on sparsity. This study predicted stock price volatility by collecting and analyzing news articles of the top 10 stocks in market cap. We used the term - document matrix based classification model to estimate the stock price fluctuations and compared the performance of the existing sparse - based word extraction method and the suggested method of removing words from the term - document matrix. The suggested method differs from the word extraction method in that it uses not only the news articles for the corresponding stock but also other news items to determine the words to extract. In other words, it removed not only the words that appeared in all the increase and decrease but also the words that appeared common in the news for other stocks. When the prediction accuracy was compared, the suggested method showed higher accuracy. The limitation of this study is that the stock price prediction was set up to classify the rise and fall, and the experiment was conducted only for the top ten stocks. The 10 stocks used in the experiment do not represent the entire stock market. In addition, it is difficult to show the investment performance because stock price fluctuation and profit rate may be different. Therefore, it is necessary to study the research using more stocks and the yield prediction through trading simulation.

A Study of 'Emotion Trigger' by Text Mining Techniques (텍스트 마이닝을 이용한 감정 유발 요인 'Emotion Trigger'에 관한 연구)

  • An, Juyoung;Bae, Junghwan;Han, Namgi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.2
    • /
    • pp.69-92
    • /
    • 2015
  • The explosion of social media data has led to apply text-mining techniques to analyze big social media data in a more rigorous manner. Even if social media text analysis algorithms were improved, previous approaches to social media text analysis have some limitations. In the field of sentiment analysis of social media written in Korean, there are two typical approaches. One is the linguistic approach using machine learning, which is the most common approach. Some studies have been conducted by adding grammatical factors to feature sets for training classification model. The other approach adopts the semantic analysis method to sentiment analysis, but this approach is mainly applied to English texts. To overcome these limitations, this study applies the Word2Vec algorithm which is an extension of the neural network algorithms to deal with more extensive semantic features that were underestimated in existing sentiment analysis. The result from adopting the Word2Vec algorithm is compared to the result from co-occurrence analysis to identify the difference between two approaches. The results show that the distribution related word extracted by Word2Vec algorithm in that the words represent some emotion about the keyword used are three times more than extracted by co-occurrence analysis. The reason of the difference between two results comes from Word2Vec's semantic features vectorization. Therefore, it is possible to say that Word2Vec algorithm is able to catch the hidden related words which have not been found in traditional analysis. In addition, Part Of Speech (POS) tagging for Korean is used to detect adjective as "emotional word" in Korean. In addition, the emotion words extracted from the text are converted into word vector by the Word2Vec algorithm to find related words. Among these related words, noun words are selected because each word of them would have causal relationship with "emotional word" in the sentence. The process of extracting these trigger factor of emotional word is named "Emotion Trigger" in this study. As a case study, the datasets used in the study are collected by searching using three keywords: professor, prosecutor, and doctor in that these keywords contain rich public emotion and opinion. Advanced data collecting was conducted to select secondary keywords for data gathering. The secondary keywords for each keyword used to gather the data to be used in actual analysis are followed: Professor (sexual assault, misappropriation of research money, recruitment irregularities, polifessor), Doctor (Shin hae-chul sky hospital, drinking and plastic surgery, rebate) Prosecutor (lewd behavior, sponsor). The size of the text data is about to 100,000(Professor: 25720, Doctor: 35110, Prosecutor: 43225) and the data are gathered from news, blog, and twitter to reflect various level of public emotion into text data analysis. As a visualization method, Gephi (http://gephi.github.io) was used and every program used in text processing and analysis are java coding. The contributions of this study are as follows: First, different approaches for sentiment analysis are integrated to overcome the limitations of existing approaches. Secondly, finding Emotion Trigger can detect the hidden connections to public emotion which existing method cannot detect. Finally, the approach used in this study could be generalized regardless of types of text data. The limitation of this study is that it is hard to say the word extracted by Emotion Trigger processing has significantly causal relationship with emotional word in a sentence. The future study will be conducted to clarify the causal relationship between emotional words and the words extracted by Emotion Trigger by comparing with the relationships manually tagged. Furthermore, the text data used in Emotion Trigger are twitter, so the data have a number of distinct features which we did not deal with in this study. These features will be considered in further study.

Construction of Event Networks from Large News Data Using Text Mining Techniques (텍스트 마이닝 기법을 적용한 뉴스 데이터에서의 사건 네트워크 구축)

  • Lee, Minchul;Kim, Hea-Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.183-203
    • /
    • 2018
  • News articles are the most suitable medium for examining the events occurring at home and abroad. Especially, as the development of information and communication technology has brought various kinds of online news media, the news about the events occurring in society has increased greatly. So automatically summarizing key events from massive amounts of news data will help users to look at many of the events at a glance. In addition, if we build and provide an event network based on the relevance of events, it will be able to greatly help the reader in understanding the current events. In this study, we propose a method for extracting event networks from large news text data. To this end, we first collected Korean political and social articles from March 2016 to March 2017, and integrated the synonyms by leaving only meaningful words through preprocessing using NPMI and Word2Vec. Latent Dirichlet allocation (LDA) topic modeling was used to calculate the subject distribution by date and to find the peak of the subject distribution and to detect the event. A total of 32 topics were extracted from the topic modeling, and the point of occurrence of the event was deduced by looking at the point at which each subject distribution surged. As a result, a total of 85 events were detected, but the final 16 events were filtered and presented using the Gaussian smoothing technique. We also calculated the relevance score between events detected to construct the event network. Using the cosine coefficient between the co-occurred events, we calculated the relevance between the events and connected the events to construct the event network. Finally, we set up the event network by setting each event to each vertex and the relevance score between events to the vertices connecting the vertices. The event network constructed in our methods helped us to sort out major events in the political and social fields in Korea that occurred in the last one year in chronological order and at the same time identify which events are related to certain events. Our approach differs from existing event detection methods in that LDA topic modeling makes it possible to easily analyze large amounts of data and to identify the relevance of events that were difficult to detect in existing event detection. We applied various text mining techniques and Word2vec technique in the text preprocessing to improve the accuracy of the extraction of proper nouns and synthetic nouns, which have been difficult in analyzing existing Korean texts, can be found. In this study, the detection and network configuration techniques of the event have the following advantages in practical application. First, LDA topic modeling, which is unsupervised learning, can easily analyze subject and topic words and distribution from huge amount of data. Also, by using the date information of the collected news articles, it is possible to express the distribution by topic in a time series. Second, we can find out the connection of events in the form of present and summarized form by calculating relevance score and constructing event network by using simultaneous occurrence of topics that are difficult to grasp in existing event detection. It can be seen from the fact that the inter-event relevance-based event network proposed in this study was actually constructed in order of occurrence time. It is also possible to identify what happened as a starting point for a series of events through the event network. The limitation of this study is that the characteristics of LDA topic modeling have different results according to the initial parameters and the number of subjects, and the subject and event name of the analysis result should be given by the subjective judgment of the researcher. Also, since each topic is assumed to be exclusive and independent, it does not take into account the relevance between themes. Subsequent studies need to calculate the relevance between events that are not covered in this study or those that belong to the same subject.

Issue tracking and voting rate prediction for 19th Korean president election candidates (댓글 분석을 통한 19대 한국 대선 후보 이슈 파악 및 득표율 예측)

  • Seo, Dae-Ho;Kim, Ji-Ho;Kim, Chang-Ki
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.199-219
    • /
    • 2018
  • With the everyday use of the Internet and the spread of various smart devices, users have been able to communicate in real time and the existing communication style has changed. Due to the change of the information subject by the Internet, data became more massive and caused the very large information called big data. These Big Data are seen as a new opportunity to understand social issues. In particular, text mining explores patterns using unstructured text data to find meaningful information. Since text data exists in various places such as newspaper, book, and web, the amount of data is very diverse and large, so it is suitable for understanding social reality. In recent years, there has been an increasing number of attempts to analyze texts from web such as SNS and blogs where the public can communicate freely. It is recognized as a useful method to grasp public opinion immediately so it can be used for political, social and cultural issue research. Text mining has received much attention in order to investigate the public's reputation for candidates, and to predict the voting rate instead of the polling. This is because many people question the credibility of the survey. Also, People tend to refuse or reveal their real intention when they are asked to respond to the poll. This study collected comments from the largest Internet portal site in Korea and conducted research on the 19th Korean presidential election in 2017. We collected 226,447 comments from April 29, 2017 to May 7, 2017, which includes the prohibition period of public opinion polls just prior to the presidential election day. We analyzed frequencies, associative emotional words, topic emotions, and candidate voting rates. By frequency analysis, we identified the words that are the most important issues per day. Particularly, according to the result of the presidential debate, it was seen that the candidate who became an issue was located at the top of the frequency analysis. By the analysis of associative emotional words, we were able to identify issues most relevant to each candidate. The topic emotion analysis was used to identify each candidate's topic and to express the emotions of the public on the topics. Finally, we estimated the voting rate by combining the volume of comments and sentiment score. By doing above, we explored the issues for each candidate and predicted the voting rate. The analysis showed that news comments is an effective tool for tracking the issue of presidential candidates and for predicting the voting rate. Particularly, this study showed issues per day and quantitative index for sentiment. Also it predicted voting rate for each candidate and precisely matched the ranking of the top five candidates. Each candidate will be able to objectively grasp public opinion and reflect it to the election strategy. Candidates can use positive issues more actively on election strategies, and try to correct negative issues. Particularly, candidates should be aware that they can get severe damage to their reputation if they face a moral problem. Voters can objectively look at issues and public opinion about each candidate and make more informed decisions when voting. If they refer to the results of this study before voting, they will be able to see the opinions of the public from the Big Data, and vote for a candidate with a more objective perspective. If the candidates have a campaign with reference to Big Data Analysis, the public will be more active on the web, recognizing that their wants are being reflected. The way of expressing their political views can be done in various web places. This can contribute to the act of political participation by the people.

Analysis of Research Trends on Mountain Streams in the Republic of Korea: Comparison to International Research Trends (산지하천을 대상으로 한 국내 연구동향 분석: 국제 연구동향과의 비교)

  • Lee, Sang In;Seo, Jung Il;Lee, Yohan;Kim, Suk Woo;Chun, Kun Woo
    • Korean Journal of Environment and Ecology
    • /
    • v.33 no.2
    • /
    • pp.216-227
    • /
    • 2019
  • The purpose of this study is to propose the rational mountain stream management strategy considering the natural conditions and social needs of the Republic of Korea. We reviewed domestic and overseas studies related to mountain streams, identified the study areas by text mining and co-word analysis using the VOSviewer program, and then analyzed the spatial and temporal study trends and topics of each study area. The results showed that domestic studies on mountain streams are still in an initial stage compared to overseas studies. Overseas studies on mountain streams can be classified into four groups: (i) habitat and species composition of fish and invertebrates, (ii) hydrological phenomena and nutrient migration, (iii) transport of sediment and organic materials and the relevant morphological changes by runoff flows, and (iv) plant species composition in mountain streams. Of these study subjects, domestic studies belonging to the (i) group mainly focused on macroinvertebrates while domestic studies belonging to the (iii) group regarded transport of sediment and organic materials as not the ecological disturbance but the source of sediment-related disasters. We then analyzed the rate of each research group to all papers by period and country. The results showed that the overseas studies belonging to (iii) and (iv) groups have increased with time, and the increase was mostly due to the studies in the United States, Brazil, Canada, and China. On the other hand, domestic studies belonging to (i) and (iii) groups increased somewhat with time, but there was a slight lack of correlation between the two subjects. Therefore, the hybridity studies to complement the shortage is necessary for the future.

KISS Korea Computer Congress 2007 (이동 객체의 패턴 탐사를 위한 시공간 데이터 일반화 기법)

  • Ko, Hyun;Kim, Kwang-Jong;Lee, Yon-Sik
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2007.06c
    • /
    • pp.153-158
    • /
    • 2007
  • 사용자들의 특성에 맞게 개인화되고 세분화된 위치 기반 서비스를 제공하기 위해서는 방대한 이동 객체의 위치 이력 데이터 집합으로부터 유용한 패턴을 추출하여 의미 있는 지식을 탐사하기 위한 시공간 패턴 탐사가 필요하다. 현재까지 다양한 패턴 탐사 기법들이 제안되었으나 이동 패턴들 중 단순히 시공간 제약이 없는 빈발 패턴만을 추출하기 때문에 한정된 시간 범위와 제한적인 영역 범위 내에서의 빈발 패턴을 탐사하는 문제에는 적용하기 어렵다. 또한 패턴 탐사 수행 시 데이터베이스를 반복 스캔하여 탐사 수행시간이 많이 소요되는 문제를 포함하거나 메모리상에 탐사 대상인 후보 패턴 트리를 생성하는 방법을 통해 탐사 시간을 줄일 수는 있으나 이동 객체 수나 최소지지도 등에 따라 트리를 구성하고 유지하는데 드는 비용이 커질 수 있다. 따라서 이러한 문제를 해결하기 위한 효율적인 패턴 탐사 기법의 개발이 요구됨으로써 선행 작업으로 본 논문에서는 상세 수준의 객체 이력 데이터들의 시간 및 공간 속성을 의미 있는 시간영역과 공간영역 정보로 변환하는 시공간 데이터 일반화 방법을 제안한다. 제안된 방법은 공간 개념 계층에 대한 영역 정보들을 영역 Grid 해쉬 테이블(AGHT:Area Grid Hash Table)로 생성하여 공간 인덱스트리인 R*-Tree의 검색 방법을 이용해 이동 객체의 위치 속성을 2차원 공간영역으로 일반화하고, 시간 개념 계층을 생성하여 이동 객체의 시간적인 속성을 시간 영역으로 일반화함으로써 일반화된 데이터 집합을 형성하여 효율적인 이동 객체의 시간 패턴 마이닝을 유도할 수 있다.의 성능을 기대할 수 있을 것이다.onium sulfate첨가배지(添加培地)에서 가장 저조(低調)하였다. vitamin중(中)에서는 niacin과 thiamine첨가배지(添加培地)에서 근소(僅少)한 증가(增加)를 나타내었다.소시켜 항이뇨 및 Na 배설 감소를 초래하는 작용과, 둘째는 신경 경로를 통하지 않고, 아마도 humoral factor를 통하여 신세뇨관에서 Na 재흡수를 억제하는 작용이 복합적으로 나타내는 것을 알 수 있었다.으로 초래되는 복합적인 기전으로 추정되었다., 소형과와 기형과는 S-3에서 많이 나왔다. 이상 연구결과에서 입도분포가 1.2-5mm인 것이 바람직한 것으로 나타났다.omopolysaccharides로 확인되었다. EPS 생성량이 가장 좋은 Leu. kimchii GJ2의 평균 분자량은 360,606 Da이었으며, 나머지 두 균주에 대해서는 생성 EPS 형태와 점도의 차이로 미루어 보아 생성 EPS의 분자구조와 분자량이 서로 다른 것으로 판단하였다.TEX>개로 통계학적으로 유의한 차이가 없었다. Heat shock protein-70 (HSP70)과 neuronal nitric oxide synthase (nNOS)에 대한 면역조직화학검사에서 실험군 Cs2군의 신경세포가 대조군 12군에 비해 HSP70과 nNOS의 과발현을 보였으며, 이는 통계학적으로 유의한 차이를 보였다(p<0.05). nNOS와 HSP70의 발현은 강한 연관성을 보였고(상관계수 0.91, p=0.000), nNOS를 발현하는 세포가 동시에 HSP70도 발현함을 확인할 수 있었다. 결론: 우리는

  • PDF

Pattern Analysis for Civil Complaints of Local Governments Using a Text Mining (텍스트마이닝에 의한 지자체 민원청구 패턴 분석)

  • Won, Tae Hong;Yoo, Hwan Hee
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.34 no.3
    • /
    • pp.319-327
    • /
    • 2016
  • Korea faces a wide range of problems in areas such as safety, environment, and traffic due to the rapid economic development and urbanization process. Despite the local governments’ efforts to deal with electronic civil complaints and solve urban problems, civil complaints have been on the increase year by year. In this study, we collected civil complaint data over the last six years from a small and medium-sized city, Jinju-si. In order to conduct a spatial distribution pattern analysis, we indicated the location data on the area through Geocoding after classifying the reasons for civil complaints and then extracted the location data of the civil complaint occurrence spots in order to analyze the correlation between electronic civil complaints and land use. Results demonstrated that electronic civil complaints in Jinju-si were clustered in residential, central commercial, and residential-industrial mixed-use areas—areas where land development had been completed within the city center. After analyzing the civil complaints according to the land use, results revealed that complaints about illegal parking were the highest. Regarding the analysis results of facility distribution within a 50m radius from the civil complaint areas, civil complaints occurred a lot in detached housing areas located within the commercial and residential-industrial mixed-use areas. In the case of residential areas(old downtown), civil complaints were condensed in the areas with many ordinary restaurants. This research explored civil complaints in terms of the urban space and can be expected to be effectively utilized in finding solutions to the civil complaints

The Research Trend and Social Perceptions Related with the Tap Water in South Korea (수돗물 이용에 대한 국내 연구동향과 사회적 인식)

  • Kim, Ji Yoon;Do, Yuno;Joo, Gea-Jae;Kim, Eunhee;Park, Eun-Young;Lee, Sang-Hyup;Baek, Myeong Su
    • Korean Journal of Ecology and Environment
    • /
    • v.49 no.3
    • /
    • pp.208-214
    • /
    • 2016
  • We analyzed research trend and public perception related with tap water to identify major factors affecting low consumption of tap water. 805 research articles were collected for text mining analysis and 1,000 on-line questionnaires were surveyed to find social variables influencing tap water intake. Based on the word network analysis, research topics were divided into 4 major categories, 1) drinking water quality, 2) water fluoridation, 3) residual chlorine, and 4) micro-organism management. Compared with these major research topics, scientific studies of drinking behavior, or social perception were rather limited. 22.4% of total respondents used tap water as drinking water source, and only 1% drank tap water without further treatments (i.e. boiling, filtering). Experience of quality control report (B=0.392, p=0.046) and level of policy trust (B=1.002, p<0.0001) were influential factors on tap water drinking behavior. Age (B=0.020, p=0.002) and gender (B= - 1.843, p<0.0001) also showed significant difference. To increase the frequency of drinking the tap water by social members, the more scientific information of tap water quality and the water policy management should be clearly shared with social members.

VRIFA: A Prediction and Nonlinear SVM Visualization Tool using LRBF kernel and Nomogram (VRIFA: LRBF 커널과 Nomogram을 이용한 예측 및 비선형 SVM 시각화도구)

  • Kim, Sung-Chul;Yu, Hwan-Jo
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.5
    • /
    • pp.722-729
    • /
    • 2010
  • Prediction problems are widely used in medical domains. For example, computer aided diagnosis or prognosis is a key component in a CDSS (Clinical Decision Support System). SVMs with nonlinear kernels like RBF kernels, have shown superior accuracy in prediction problems. However, they are not preferred by physicians for medical prediction problems because nonlinear SVMs are difficult to visualize, thus it is hard to provide intuitive interpretation of prediction results to physicians. Nomogram was proposed to visualize SVM classification models. However, it cannot visualize nonlinear SVM models. Localized Radial Basis Function (LRBF) was proposed which shows comparable accuracy as the RBF kernel while the LRBF kernel is easier to interpret since it can be linearly decomposed. This paper presents a new tool named VRIFA, which integrates the nomogram and LRBF kernel to provide users with an interactive visualization of nonlinear SVM models, VRIFA visualizes the internal structure of nonlinear SVM models showing the effect of each feature, the magnitude of the effect, and the change at the prediction output. VRIFA also performs nomogram-based feature selection while training a model in order to remove noise or redundant features and improve the prediction accuracy. The area under the ROC curve (AUC) can be used to evaluate the prediction result when the data set is highly imbalanced. The tool can be used by biomedical researchers for computer-aided diagnosis and risk factor analysis for diseases.

Research Suggestion for Disaster Prediction using Safety Report of Korea Government (안전신문고를 이용한 재난 예측 방법론 제안)

  • Lee, Jun;Shin, Jindong;Cho, Sangmyeong;Lee, Sanghwa
    • Journal of Korean Society of Disaster and Security
    • /
    • v.12 no.4
    • /
    • pp.15-26
    • /
    • 2019
  • Anjunshinmungo (The safety e-report) has been in operation since 2014, and there are about 1 million cumulative reports by June 2019. This study analyzes the contents of more than 1 million safety newspapers reported at the present time of information age to determine how powerful and meaningful the people's voice and interest are. In particular, we are interested in forecasting ability. We wanted to check whether the report of the safety newspaper was related to possible disasters. To this end, the researchers received data reported in the safety newspaper as text and analyzed it by natural language analysis methodology. Based on this, the newspaper articles during the analysis of the safety newspaper were analyzed, and the correlation between the contents of the newspaper and the newspaper was analyzed. As a result, accidents occurred within a few months as the number of reports related to response and confirmation increased, and analyzing the contents of safety reports previously reported on social instability can be used to predict future disasters.