• Title/Summary/Keyword: 비정형분석

Search Result 484, Processing Time 0.027 seconds

Framework for Measuring Dynamic Influence Index & Influence Factors using Social Data on Facebook (페이스북 소셜 데이터를 이용한 동적 영향 요인 및 영향력 측정 방법에 관한 프레임워크)

  • Koh, Seoung-hyun;You, Yen-yoo
    • Journal of Digital Convergence
    • /
    • v.14 no.10
    • /
    • pp.137-145
    • /
    • 2016
  • The explosive growth of social networking services based on smart devices popularize these relationships and activities online in accordance with the far larger impact of this on the real life offline, the interest and importance for the online activity is increasing. In this study, factors affecting the SNS activity are defined by object, user, influence direction, influence distance and proposed a method to measure organic terms in effect between the SNS users. Influence Direction and Influence Strength (or Distance) are elaborated by using the existing influence measurement element such as structured data - the number of friends, the difference between the number of contacts - and the new influence measurement element such as unstructured data - gap between the former time and the latter time, preference and type of response behavior - that occur in social network service. In addition, the system for collecting and analysing data for measuring influence from social network service and the process model on the method for measuring influence is tested by using sample data on Facebook and explained the implementation probability.

Informal Quality Data Analysis via Sentimental analysis and Word2vec method (감성분석과 Word2vec을 이용한 비정형 품질 데이터 분석)

  • Lee, Chinuk;Yoo, Kook Hyun;Mun, Byeong Min;Bae, Suk Joo
    • Journal of Korean Society for Quality Management
    • /
    • v.45 no.1
    • /
    • pp.117-128
    • /
    • 2017
  • Purpose: This study analyzes automobile quality review data to develop alternative analytical method of informal data. Existing methods to analyze informal data are based mainly on the frequency of informal data, however, this research tries to use correlation information of each informal data. Method: After sentimental analysis to acquire the user information for automobile products, three classification methods, that is, $na{\ddot{i}}ve$ Bayes, random forest, and support vector machine, were employed to accurately classify the informal user opinions with respect to automobile qualities. Additionally, Word2vec was applied to discover correlated information about informal data. Result: As applicative results of three classification methods, random forest method shows most effective results compared to the other classification methods. Word2vec method manages to discover closest relevant data with automobile components. Conclusion: The proposed method shows its effectiveness in terms of accuracy and sensitivity on the analysis of informal quality data, however, only two sentiments (positive or negative) can be categorized due to human errors. Further studies are required to derive more sentiments to accurately classify informal quality data. Word2vec method also shows comparative results to discover the relevance of components precisely.

Rapid Management Mechanism Against Harmful Materials of Agri-Food Based on Big Data Analysis (빅 데이터 분석 기반 농 식품 위해인자 신속관리 방법)

  • Park, Hyeon;Kang, Sung-soo;Jeong, Hoon;Kim, Se-Han
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.6
    • /
    • pp.1166-1174
    • /
    • 2015
  • There were the attempts to prevent the spread of harmful materials of the agri-food through the record tracking of the products with the bar code, the partial information tracking of the agri-food storage and the delivery vehicle, or the control of the temperature by intuition. However, there were many problems in the attempts because of the insufficient information, the information distortion and the independent information network of each distribution company. As a result, it is difficult to prevent the spread over the life-cycle of the agri-food using the attempts. To solve the problems, we propose the mechanism mainly to do context awareness, predict, and track the harmful materials of agri-food using big data processing.

Context Dependent Fusion with Support Vector Machines (Support Vector Machine을 이용한 문맥 민감형 융합)

  • Heo, Gyeongyong
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.7
    • /
    • pp.37-45
    • /
    • 2013
  • Context dependent fusion (CDF) is a fusion algorithm that combines multiple outputs from different classifiers to achieve better performance. CDF tries to divide the problem context into several homogeneous sub-contexts and to fuse data locally with respect to each sub-context. CDF showed better performance than existing methods, however, it is sensitive to noise due to the large number of parameters optimized and the innate linearity limits the application of CDF. In this paper, a variant of CDF using support vector machines (SVMs) for fusion and kernel principal component analysis (K-PCA) for context extraction is proposed to solve the problems in CDF, named CDF-SVM. Kernel PCA can shape irregular clusters including elliptical ones through the non-linear kernel transformation and SVM can draw a non-linear decision boundary. Regularization terms is also included in the objective function of CDF-SVM to mitigate the noise sensitivity in CDF. CDF-SVM showed better performance than CDF and its variants, which is demonstrated through the experiments with a landmine data set.

The Study on Positioning of Giant Characters of Sci-Fi Movies & Games in Media Convergence Ages (미디어융복합 시대에서 SF영화와 게임에 등장하는 거대캐릭터 포지셔닝 연구)

  • Joo, Jin-Su;Oh, Seung-Hwan
    • Journal of Digital Convergence
    • /
    • v.13 no.7
    • /
    • pp.349-357
    • /
    • 2015
  • Giant character used various SF movies and games in media convergence ages, and is essential for giant character in success contents. It This study defined giant character of SF movies and games, it analysed the eight of external characteristics and internal characteristics of giant character in SF movies and games. The external characteristics defined shape, silhouette, size and color, the internal characteristics defined fear, satanism, image and story focus, playfulness. Above, it was structured positioning model of giant character based eight characteristics and analyzation of example of SF movies and games. The elements of positioning model of giant characters are darkness, huge, abnormal, human, animal, fear, satanism, story focused, image focused and playfulness, and this study was proposed these model of elements of eight in SF movies and games.

A Study on the Introduction of Library Services Based on Cloud Computing (클라우드 컴퓨팅 기반의 도서관 서비스 도입방안에 관한 연구)

  • Kim, Yong
    • Journal of the Korean BIBLIA Society for library and Information Science
    • /
    • v.23 no.3
    • /
    • pp.57-84
    • /
    • 2012
  • With the advent of Big data era unleashed by tremendous increase of structured and unstructured data, a library needs an effective method to store, manage and preserve various information resources. Also, needs of collaboration of libraries are continuously increased in digital environment. As an effective method to handle the changes and challenges in libraries, interest on cloud computing is growing more and more. This study aims to propose a method to introduce cloud computing in libraries. To achieve the goals, this study performed the literature review to analyze problems of existing library systems. Also, this study proposed considerations, expectations, service scenario, phased strategy to introduce cloud computing in libraries. Based on the results extracted from cases that libraries have introduced cloud computing-based systems, this study proposed introduction strategy and specific applying areas in library works as considered features of cloud computing models. The proposed phased strategy and service scenario may reduce time and effort in the process of introduction of cloud computing and maximize the effect of cloud computing.

A Study on Implementation of Fraud Detection System (FDS) Applying BigData Platform (빅데이터 기술을 활용한 이상금융거래 탐지시스템 구축 연구)

  • Kang, Jae-Goo;Lee, Ji-Yean;You, Yen-Yoo
    • Journal of the Korea Convergence Society
    • /
    • v.8 no.4
    • /
    • pp.19-24
    • /
    • 2017
  • The growing number of electronic financial transactions (e-banking) has entailed the rapid increase in security threats such as extortion and falsification of financial transaction data. Against such background, rigid security and countermeasures to hedge against such problems have risen as urgent tasks. Thus, this study aims to implement an improved case model by applying the Fraud Detection System (hereinafter, FDS) in a financial corporation 'A' using big data technique (e.g. the function to collect/store various types of typical/atypical financial transaction event data in real time regarding the external intrusion, outflow of internal data, and fraud financial transactions). As a result, There was reduction effect in terms of previous scenario detection target by minimizing false alarm via advanced scenario analysis. And further suggest the future direction of the enhanced FDS.

A Study on the Integration of Recognition Technology for Scientific Core Entities (과학기술 핵심개체 인식기술 통합에 관한 연구)

  • Choi, Yun-Soo;Jeong, Chang-Hoo;Cho, Hyun-Yang
    • Journal of the Korean Society for information Management
    • /
    • v.28 no.1
    • /
    • pp.89-104
    • /
    • 2011
  • Large-scaled information extraction plays an important role in advanced information retrieval as well as question answering and summarization. Information extraction can be defined as a process of converting unstructured documents into formalized, tabular information, which consists of named-entity recognition, terminology extraction, coreference resolution and relation extraction. Since all the elementary technologies have been studied independently so far, it is not trivial to integrate all the necessary processes of information extraction due to the diversity of their input/output formation approaches and operating environments. As a result, it is difficult to handle scientific documents to extract both named-entities and technical terms at once. In order to extract these entities automatically from scientific documents at once, we developed a framework for scientific core entity extraction which embraces all the pivotal language processors, named-entity recognizer and terminology extractor.

Consideration of the Direction for Improving RI-Biomics Information System for Using Big Data in Radiation Field (방사선 빅데이터 활용을 위한 RI-Biomics 기술정보시스템 개선 방향성에 관한 고찰)

  • Lee, Seung Hyun;Kim, Joo Yeon;Lim, Young-Khi;Park, Tai-Jin
    • Journal of Radiation Industry
    • /
    • v.11 no.1
    • /
    • pp.7-11
    • /
    • 2017
  • RI-Biomics is a fusion technology in radiation fields for evaluating in-vivo dynamics such as absorption, distribution, metabolism and excretion (RI-ADME) of new drugs and materials using radioisotopes and quantitative evaluation of their efficacy. RI-Biomics information is being provided by RIBio-Info developed as information system for distributing its information and three requirements for improving RIBio-Info system have been derived through reviewing recent big data trends in this study. Three requirements are defined as resource, technology and manpower, and some reviews for applying big data in RIBio-In system are suggested. Fist, applicable external big data have to be obtained, second, some infrastructures for realizing applying big data to be expanded, and finally, data scientists able to analyze large scale of information to be trained. Therefore, an original technology driven to analyze for atypical and large scale of data can be created and this stated technology can contribute to obtain a basis to create a new value in RI-Biomics field.

Analysis of the National Police Agency business trends using text mining (텍스트 마이닝 기법을 이용한 경찰청 업무 트렌드 분석)

  • Sun, Hyunseok;Lim, Changwon
    • The Korean Journal of Applied Statistics
    • /
    • v.32 no.2
    • /
    • pp.301-317
    • /
    • 2019
  • There has been significant research conducted on how to discover various insights through text data using statistical techniques. In this study we analyzed text data produced by the Korean National Police Agency to identify trends in the work by year and compare work characteristics among local authorities by identifying distinctive keywords in documents produced by each local authority. A preprocessing according to the characteristics of each data was conducted and the frequency of words for each document was calculated in order to draw a meaningful conclusion. The simple term frequency shown in the document is difficult to describe the characteristics of the keywords; therefore, the frequency for each term was newly calculated using the term frequency-inverse document frequency weights. The L2 norm normalization technique was used to compare the frequency of words. The analysis can be used as basic data that can be newly for future police work improvement policies and as a method to improve the efficiency of the police service that also help identify a demand for improvements in indoor work.