• Title/Summary/Keyword: KeyWord-Based System

Search Result 68, Processing Time 0.026 seconds

Research Trend Analysis on Living Lab Using Text Mining (텍스트 마이닝을 이용한 리빙랩 연구동향 분석)

  • Kim, SeongMook;Kim, YoungJun
    • Journal of Digital Convergence
    • /
    • v.18 no.8
    • /
    • pp.37-48
    • /
    • 2020
  • This study aimed at understanding trends of living lab studies and deriving implications for directions of the studies by utilizing text mining. The study included network analysis and topic modelling based on keywords and abstracts from total 166 thesis published between 2011 and November 2019. Centrality analysis showed that living lab studies had been conducted focusing on keywords like innovation, society, technology, development, user and so on. From the topic modelling, 5 topics such as "regional innovation and user support", "social policy program of government", "smart city platform building", "technology innovation model of company" and "participation in system transformation" were extracted. Since the foundation of KNoLL in 2017, the diversification of living lab study subjects has been made. Quantitative analysis using text mining provides useful results for development of living lab studies.

Analysis of Pressure Ulcer Nursing Records with Artificial Intelligence-based Natural Language Processing (인공지능 기반 자연어처리를 적용한 욕창간호기록 분석)

  • Kim, Myoung Soo;Ryu, Jung-Mi
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.10
    • /
    • pp.365-372
    • /
    • 2021
  • The purpose of this study was to examine the statements characteristics of the pressure ulcer nursing record by natural langage processing and assess the prediction accuracy for each pressure ulcer stage. Nursing records related to pressure ulcer were analyzed using descriptive statistics, and word cloud generators (http://wordcloud.kr) were used to examine the characteristics of words in the pressure ulcer prevention nursing records. The accuracy ratio for the pressure ulcer stage was calculated using deep learning. As a result of the study, the second stage and the deep tissue injury suspected were 23.1% and 23.0%, respectively, and the most frequent key words were erythema, blisters, bark, area, and size. The stages with high prediction accuracy were in the order of stage 0, deep tissue injury suspected, and stage 2. These results suggest that it can be developed as a clinical decision support system available to practice for nurses at the pressure ulcer prevention care.

A Review of the Study Trends on the Relationship between Primary Dysmenorrhea and Doppler Indicies of Uterine Artery (일차성 월경통과 자궁동맥의 혈류역학적 측정값의 관련성에 대한 최신 연구 동향)

  • Kim, Hyo-Jung;Hwang, Deok-Sang;Lee, Jin-Moo;Lee, Chang-Hoon;Jang, Jun-Bock
    • The Journal of Korean Obstetrics and Gynecology
    • /
    • v.34 no.4
    • /
    • pp.97-110
    • /
    • 2021
  • Objectives: This study was performed to review the research trends in the relationship between primary dysmenorrhea and doppler indicies of uterine artery. Methods: The search for related papers used 'Pubmed', a reserch engine in the America National Library of Medicine and Korean studies Information Service System (KISS). Used searching terms were 'primary dysmenorrhea', 'menstrual pain' in all cases. And among these studies, we searched by using key word 'uterine artery', 'doppler indices', 'doppler parameters', 'pulsatile index', 'resistance index'. Results: Overall 49 studies, 8 studies were finally selected to this study. There were 6 controlled studies and 2 randomised-controlled studies. In all 8 studies, transvaginal ultrasound was used to measure the resistance of uterine blood vessels. All of these studies reported that in patients with primary dysmenorrhea, hemodynamic values of uterine arteries measured by Doppler ultrasound were significantly higher than in normal subjects. Conclusions: According to the results, it was found that there was a positive correlation between the pain level of primary dysmenorrhea and the pulsation index and resistance index of the uterine artery. Based on these results, it can be seen that the doppler indicies of uterine artery have the potential to be used as an evaluation scale for Korean traditional medicine for primary dysmenorrhea.

Term Mapping Methodology between Everyday Words and Legal Terms for Law Information Search System (법령정보 검색을 위한 생활용어와 법률용어 간의 대응관계 탐색 방법론)

  • Kim, Ji Hyun;Lee, Jong-Seo;Lee, Myungjin;Kim, Wooju;Hong, June Seok
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.137-152
    • /
    • 2012
  • In the generation of Web 2.0, as many users start to make lots of web contents called user created contents by themselves, the World Wide Web is overflowing by countless information. Therefore, it becomes the key to find out meaningful information among lots of resources. Nowadays, the information retrieval is the most important thing throughout the whole field and several types of search services are developed and widely used in various fields to retrieve information that user really wants. Especially, the legal information search is one of the indispensable services in order to provide people with their convenience through searching the law necessary to their present situation as a channel getting knowledge about it. The Office of Legislation in Korea provides the Korean Law Information portal service to search the law information such as legislation, administrative rule, and judicial precedent from 2009, so people can conveniently find information related to the law. However, this service has limitation because the recent technology for search engine basically returns documents depending on whether the query is included in it or not as a search result. Therefore, it is really difficult to retrieve information related the law for general users who are not familiar with legal terms in the search engine using simple matching of keywords in spite of those kinds of efforts of the Office of Legislation in Korea, because there is a huge divergence between everyday words and legal terms which are especially from Chinese words. Generally, people try to access the law information using everyday words, so they have a difficulty to get the result that they exactly want. In this paper, we propose a term mapping methodology between everyday words and legal terms for general users who don't have sufficient background about legal terms, and we develop a search service that can provide the search results of law information from everyday words. This will be able to search the law information accurately without the knowledge of legal terminology. In other words, our research goal is to make a law information search system that general users are able to retrieval the law information with everyday words. First, this paper takes advantage of tags of internet blogs using the concept for collective intelligence to find out the term mapping relationship between everyday words and legal terms. In order to achieve our goal, we collect tags related to an everyday word from web blog posts. Generally, people add a non-hierarchical keyword or term like a synonym, especially called tag, in order to describe, classify, and manage their posts when they make any post in the internet blog. Second, the collected tags are clustered through the cluster analysis method, K-means. Then, we find a mapping relationship between an everyday word and a legal term using our estimation measure to select the fittest one that can match with an everyday word. Selected legal terms are given the definite relationship, and the relations between everyday words and legal terms are described using SKOS that is an ontology to describe the knowledge related to thesauri, classification schemes, taxonomies, and subject-heading. Thus, based on proposed mapping and searching methodologies, our legal information search system finds out a legal term mapped with user query and retrieves law information using a matched legal term, if users try to retrieve law information using an everyday word. Therefore, from our research, users can get exact results even if they do not have the knowledge related to legal terms. As a result of our research, we expect that general users who don't have professional legal background can conveniently and efficiently retrieve the legal information using everyday words.

Construction of Consumer Confidence index based on Sentiment analysis using News articles (뉴스기사를 이용한 소비자의 경기심리지수 생성)

  • Song, Minchae;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.3
    • /
    • pp.1-27
    • /
    • 2017
  • It is known that the economic sentiment index and macroeconomic indicators are closely related because economic agent's judgment and forecast of the business conditions affect economic fluctuations. For this reason, consumer sentiment or confidence provides steady fodder for business and is treated as an important piece of economic information. In Korea, private consumption accounts and consumer sentiment index highly relevant for both, which is a very important economic indicator for evaluating and forecasting the domestic economic situation. However, despite offering relevant insights into private consumption and GDP, the traditional approach to measuring the consumer confidence based on the survey has several limits. One possible weakness is that it takes considerable time to research, collect, and aggregate the data. If certain urgent issues arise, timely information will not be announced until the end of each month. In addition, the survey only contains information derived from questionnaire items, which means it can be difficult to catch up to the direct effects of newly arising issues. The survey also faces potential declines in response rates and erroneous responses. Therefore, it is necessary to find a way to complement it. For this purpose, we construct and assess an index designed to measure consumer economic sentiment index using sentiment analysis. Unlike the survey-based measures, our index relies on textual analysis to extract sentiment from economic and financial news articles. In particular, text data such as news articles and SNS are timely and cover a wide range of issues; because such sources can quickly capture the economic impact of specific economic issues, they have great potential as economic indicators. There exist two main approaches to the automatic extraction of sentiment from a text, we apply the lexicon-based approach, using sentiment lexicon dictionaries of words annotated with the semantic orientations. In creating the sentiment lexicon dictionaries, we enter the semantic orientation of individual words manually, though we do not attempt a full linguistic analysis (one that involves analysis of word senses or argument structure); this is the limitation of our research and further work in that direction remains possible. In this study, we generate a time series index of economic sentiment in the news. The construction of the index consists of three broad steps: (1) Collecting a large corpus of economic news articles on the web, (2) Applying lexicon-based methods for sentiment analysis of each article to score the article in terms of sentiment orientation (positive, negative and neutral), and (3) Constructing an economic sentiment index of consumers by aggregating monthly time series for each sentiment word. In line with existing scholarly assessments of the relationship between the consumer confidence index and macroeconomic indicators, any new index should be assessed for its usefulness. We examine the new index's usefulness by comparing other economic indicators to the CSI. To check the usefulness of the newly index based on sentiment analysis, trend and cross - correlation analysis are carried out to analyze the relations and lagged structure. Finally, we analyze the forecasting power using the one step ahead of out of sample prediction. As a result, the news sentiment index correlates strongly with related contemporaneous key indicators in almost all experiments. We also find that news sentiment shocks predict future economic activity in most cases. In almost all experiments, the news sentiment index strongly correlates with related contemporaneous key indicators. Furthermore, in most cases, news sentiment shocks predict future economic activity; in head-to-head comparisons, the news sentiment measures outperform survey-based sentiment index as CSI. Policy makers want to understand consumer or public opinions about existing or proposed policies. Such opinions enable relevant government decision-makers to respond quickly to monitor various web media, SNS, or news articles. Textual data, such as news articles and social networks (Twitter, Facebook and blogs) are generated at high-speeds and cover a wide range of issues; because such sources can quickly capture the economic impact of specific economic issues, they have great potential as economic indicators. Although research using unstructured data in economic analysis is in its early stages, but the utilization of data is expected to greatly increase once its usefulness is confirmed.

The Method for Real-time Complex Event Detection of Unstructured Big data (비정형 빅데이터의 실시간 복합 이벤트 탐지를 위한 기법)

  • Lee, Jun Heui;Baek, Sung Ha;Lee, Soon Jo;Bae, Hae Young
    • Spatial Information Research
    • /
    • v.20 no.5
    • /
    • pp.99-109
    • /
    • 2012
  • Recently, due to the growth of social media and spread of smart-phone, the amount of data has considerably increased by full use of SNS (Social Network Service). According to it, the Big Data concept is come up and many researchers are seeking solutions to make the best use of big data. To maximize the creative value of the big data held by many companies, it is required to combine them with existing data. The physical and theoretical storage structures of data sources are so different that a system which can integrate and manage them is needed. In order to process big data, MapReduce is developed as a system which has advantages over processing data fast by distributed processing. However, it is difficult to construct and store a system for all key words. Due to the process of storage and search, it is to some extent difficult to do real-time processing. And it makes extra expenses to process complex event without structure of processing different data. In order to solve this problem, the existing Complex Event Processing System is supposed to be used. When it comes to complex event processing system, it gets data from different sources and combines them with each other to make it possible to do complex event processing that is useful for real-time processing specially in stream data. Nevertheless, unstructured data based on text of SNS and internet articles is managed as text type and there is a need to compare strings every time the query processing should be done. And it results in poor performance. Therefore, we try to make it possible to manage unstructured data and do query process fast in complex event processing system. And we extend the data complex function for giving theoretical schema of string. It is completed by changing the string key word into integer type with filtering which uses keyword set. In addition, by using the Complex Event Processing System and processing stream data at real-time of in-memory, we try to reduce the time of reading the query processing after it is stored in the disk.

Subject-Balanced Intelligent Text Summarization Scheme (주제 균형 지능형 텍스트 요약 기법)

  • Yun, Yeoil;Ko, Eunjung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.141-166
    • /
    • 2019
  • Recently, channels like social media and SNS create enormous amount of data. In all kinds of data, portions of unstructured data which represented as text data has increased geometrically. But there are some difficulties to check all text data, so it is important to access those data rapidly and grasp key points of text. Due to needs of efficient understanding, many studies about text summarization for handling and using tremendous amounts of text data have been proposed. Especially, a lot of summarization methods using machine learning and artificial intelligence algorithms have been proposed lately to generate summary objectively and effectively which called "automatic summarization". However almost text summarization methods proposed up to date construct summary focused on frequency of contents in original documents. Those summaries have a limitation for contain small-weight subjects that mentioned less in original text. If summaries include contents with only major subject, bias occurs and it causes loss of information so that it is hard to ascertain every subject documents have. To avoid those bias, it is possible to summarize in point of balance between topics document have so all subject in document can be ascertained, but still unbalance of distribution between those subjects remains. To retain balance of subjects in summary, it is necessary to consider proportion of every subject documents originally have and also allocate the portion of subjects equally so that even sentences of minor subjects can be included in summary sufficiently. In this study, we propose "subject-balanced" text summarization method that procure balance between all subjects and minimize omission of low-frequency subjects. For subject-balanced summary, we use two concept of summary evaluation metrics "completeness" and "succinctness". Completeness is the feature that summary should include contents of original documents fully and succinctness means summary has minimum duplication with contents in itself. Proposed method has 3-phases for summarization. First phase is constructing subject term dictionaries. Topic modeling is used for calculating topic-term weight which indicates degrees that each terms are related to each topic. From derived weight, it is possible to figure out highly related terms for every topic and subjects of documents can be found from various topic composed similar meaning terms. And then, few terms are selected which represent subject well. In this method, it is called "seed terms". However, those terms are too small to explain each subject enough, so sufficient similar terms with seed terms are needed for well-constructed subject dictionary. Word2Vec is used for word expansion, finds similar terms with seed terms. Word vectors are created after Word2Vec modeling, and from those vectors, similarity between all terms can be derived by using cosine-similarity. Higher cosine similarity between two terms calculated, higher relationship between two terms defined. So terms that have high similarity values with seed terms for each subjects are selected and filtering those expanded terms subject dictionary is finally constructed. Next phase is allocating subjects to every sentences which original documents have. To grasp contents of all sentences first, frequency analysis is conducted with specific terms that subject dictionaries compose. TF-IDF weight of each subjects are calculated after frequency analysis, and it is possible to figure out how much sentences are explaining about each subjects. However, TF-IDF weight has limitation that the weight can be increased infinitely, so by normalizing TF-IDF weights for every subject sentences have, all values are changed to 0 to 1 values. Then allocating subject for every sentences with maximum TF-IDF weight between all subjects, sentence group are constructed for each subjects finally. Last phase is summary generation parts. Sen2Vec is used to figure out similarity between subject-sentences, and similarity matrix can be formed. By repetitive sentences selecting, it is possible to generate summary that include contents of original documents fully and minimize duplication in summary itself. For evaluation of proposed method, 50,000 reviews of TripAdvisor are used for constructing subject dictionaries and 23,087 reviews are used for generating summary. Also comparison between proposed method summary and frequency-based summary is performed and as a result, it is verified that summary from proposed method can retain balance of all subject more which documents originally have.

Construction of Event Networks from Large News Data Using Text Mining Techniques (텍스트 마이닝 기법을 적용한 뉴스 데이터에서의 사건 네트워크 구축)

  • Lee, Minchul;Kim, Hea-Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.183-203
    • /
    • 2018
  • News articles are the most suitable medium for examining the events occurring at home and abroad. Especially, as the development of information and communication technology has brought various kinds of online news media, the news about the events occurring in society has increased greatly. So automatically summarizing key events from massive amounts of news data will help users to look at many of the events at a glance. In addition, if we build and provide an event network based on the relevance of events, it will be able to greatly help the reader in understanding the current events. In this study, we propose a method for extracting event networks from large news text data. To this end, we first collected Korean political and social articles from March 2016 to March 2017, and integrated the synonyms by leaving only meaningful words through preprocessing using NPMI and Word2Vec. Latent Dirichlet allocation (LDA) topic modeling was used to calculate the subject distribution by date and to find the peak of the subject distribution and to detect the event. A total of 32 topics were extracted from the topic modeling, and the point of occurrence of the event was deduced by looking at the point at which each subject distribution surged. As a result, a total of 85 events were detected, but the final 16 events were filtered and presented using the Gaussian smoothing technique. We also calculated the relevance score between events detected to construct the event network. Using the cosine coefficient between the co-occurred events, we calculated the relevance between the events and connected the events to construct the event network. Finally, we set up the event network by setting each event to each vertex and the relevance score between events to the vertices connecting the vertices. The event network constructed in our methods helped us to sort out major events in the political and social fields in Korea that occurred in the last one year in chronological order and at the same time identify which events are related to certain events. Our approach differs from existing event detection methods in that LDA topic modeling makes it possible to easily analyze large amounts of data and to identify the relevance of events that were difficult to detect in existing event detection. We applied various text mining techniques and Word2vec technique in the text preprocessing to improve the accuracy of the extraction of proper nouns and synthetic nouns, which have been difficult in analyzing existing Korean texts, can be found. In this study, the detection and network configuration techniques of the event have the following advantages in practical application. First, LDA topic modeling, which is unsupervised learning, can easily analyze subject and topic words and distribution from huge amount of data. Also, by using the date information of the collected news articles, it is possible to express the distribution by topic in a time series. Second, we can find out the connection of events in the form of present and summarized form by calculating relevance score and constructing event network by using simultaneous occurrence of topics that are difficult to grasp in existing event detection. It can be seen from the fact that the inter-event relevance-based event network proposed in this study was actually constructed in order of occurrence time. It is also possible to identify what happened as a starting point for a series of events through the event network. The limitation of this study is that the characteristics of LDA topic modeling have different results according to the initial parameters and the number of subjects, and the subject and event name of the analysis result should be given by the subjective judgment of the researcher. Also, since each topic is assumed to be exclusive and independent, it does not take into account the relevance between themes. Subsequent studies need to calculate the relevance between events that are not covered in this study or those that belong to the same subject.

Evaluations of Chinese Brand Name by Different Translation Types: Focusing on The Moderating Role of Brand Concept (영문 브랜드네임의 중문 브랜드네임 전환 방식에 대한 중화권 소비자들의 브랜드 평가에 관한 연구 -브랜드컨셉의 조절효과를 중심으로-)

  • Lee, Jieun;Jeon, Jooeon;Hsiao, Chen Fei
    • Asia Marketing Journal
    • /
    • v.12 no.4
    • /
    • pp.1-25
    • /
    • 2011
  • Brand names are often considered as a part of product and important extrinsic cues of product evaluation, when consumers make purchasing decisions. For a company, brand names are also important assets. Building a strong brand name in the Chinese commonwealth is a main challenge for many global companies. One of the first problem global company has to face is how to translate English brand name into Chinese brand name. It is very difficult decision because of cultural and linguistic differences. Western languages are based on an alphabet phonetic system, whereas Chinese are based on ideogram. Chinese speakers are more likely to recall stimuli presented as brand names in visual rather than spoken recall, whereas English speakers are more likely to recall the names in spoken rather than in visual recall. We interpret these findings in terms of the fact that mental representations of verbal information in Chinese are coded primarily in a visual manner, whereas verbal information in English is coded by primarily in a phonological manner. A key linguistic differences that would affect the decision to standardize or localize when transferring English brand name to Chinese brand name is the writing system. Prior Chinese brand naming research suggests that popular Chinese naming translations foreign companies adopt are phonetic, semantic, and phonosemantic translation. The phonetic translation refers to the speech sound that is produced, such as the pronunciation of the brand name. The semantic translation involves the actual meaning of and association made with the brand name. The phonosemantic translation preserves the sound of the brand name and brand meaning. Prior brand naming research has dealt with word-level analysis in examining English brand name that are desirable for improving memorability. We predict Chinese brand name suggestiveness with different translation methods lead to different levels of consumers' evaluations. This research investigates the structural linguistic characteristics of the Chinese language and its impact on the brand name evaluation. Otherwise purpose of this study is to examine the effect of brand concept on the evaluation of brand name. We also want to examine whether the evaluation is moderated by Chinese translation types. 178 Taiwanese participants were recruited for the research. The following findings are from the empirical analysis on the hypotheses established in this study. In the functional brand concept, participants in Chinese translation by semantic were likely to evaluate positively than Chinese translation by phonetic. On the contrary, in the symbolic brand concept condition, participants in Chinese translation by phonetic evaluated positively than by semantic. And then, we found Chinese translation by phonosemantic was most favorable evaluations regardless of brand concept. The implications of these findings are discussed for Chinese commonwealth marketers with respect to brand name strategies. The proposed model helps companies to effectively select brand name, making it highly applicable for academia and practitioner. name and brand meaning. Prior brand naming research has dealt with word-level analysis in examining English brand name that are desirable for improving memorability. We predict Chinese brand name suggestiveness with different translation methods lead to different levels of consumers' evaluations. This research investigates the structural linguistic characteristics of the Chinese language and its impact on the brand name evaluation. Otherwise purpose of this study is to examine the effect of brand concept on the evaluation of brand name. We also want to examine whether the evaluation is moderated by Chinese translation types. 178 Taiwanese participants were recruited for the research. The following findings are from the empirical analysis on the hypotheses established in this study. In the functional brand concept, participants in Chinese translation by semantic were likely to evaluate positively than Chinese translation by phonetic. On the contrary, in the symbolic brand concept condition, participants in Chinese translation by phonetic evaluated positively than by semantic. And then, we found Chinese translation by phonosemantic was most favorable evaluations regardless of brand concept. The implications of these findings are discussed for Chinese commonwealth marketers with respect to brand name strategies. The proposed model helps companies to effectively select brand name, making it highly applicable for academia and practitioner.

  • PDF

Design of Serendipity Service Based on Near Field Communication Technology (NFC 기반 세렌디피티 시스템 설계)

  • Lee, Kyoung-Jun;Hong, Sung-Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.293-304
    • /
    • 2011
  • The world of ubiquitous computing is one in which we will be surrounded by an ever-richer set of networked devices and services. Especially, mobile phone now becomes one of the key issues in ubiquitous computing environments. Mobile phones have been infecting our normal lives more thoroughly, and are the fastest technology in human history that has been adapted to people. In Korea, the number of mobile phones registered to the telecom company, is more than the population of the country. Last year, the numbers of mobile phone sold are many times more than the number of personal computer sold. The new advanced technology of mobile phone is now becoming the most concern on every field of technologies. The mix of wireless communication technology (wifi) and mobile phone (smart phone) has made a new world of ubiquitous computing and people can always access to the network anywhere, in high speed, and easily. In such a world, people cannot expect to have available to us specific applications that allow them to accomplish every conceivable combination of information that they might wish. They are willing to have information they want at easy way, and fast way, compared to the world we had before, where we had to have a desktop, cable connection, limited application, and limited speed to achieve what they want. Instead, now people can believe that many of their interactions will be through highly generic tools that allow end-user discovery, configuration, interconnection, and control of the devices around them. Serendipity is an application of the architecture that will help people to solve a concern of achieving their information. The word 'serendipity', introduced to scientific fields in eighteenth century, is the meaning of making new discoveries by accidents and sagacity. By combining to the field of ubiquitous computing and smart phone, it will change the way of achieving the information. Serendipity may enable professional practitioners to function more effectively in the unpredictable, dynamic environment that informs the reality of information seeking. This paper designs the Serendipity Service based on NFC (Near Field Communication) technology. When users of NFC smart phone get information and services by touching the NFC tags, serendipity service will be core services which will give an unexpected but valuable finding. This paper proposes the architecture, scenario and the interface of serendipity service using tag touch data, serendipity cases, serendipity rule base and user profile.