• Title/Summary/Keyword: semantic classification

Search Result 329, Processing Time 0.026 seconds

Identification of Demand Type Differences and Their Impact on Consumer Behavior: A Case Study Based on Smart Wearable Product Design

  • Jialei Ye;Xiaoyou He;Ziyang Liu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.1101-1121
    • /
    • 2024
  • Thorough understanding of user demands and formulation of product development strategies are crucial in product design, and can effectively stimulate consumer behavior. Scientific categorization and classification of demands contribute to accurate design development, design efficiency, and success rates. In recent years, e-commerce has become important consumption platforms for smart wearable products. However, there are few studies on product design and development among those related to promoting platform product services and sales. Meanwhile, design strategies focusing on real consumer needs are scarce among smart wearable product design studies. Therefore, an empirical consumer demand analysis method is proposed and design development strategies are formulated based on a categorized interpretation of demands. Using representative smart bracelets from wearable smart products as a case, this paper classifies consumer demands with three methods: big data semantic analysis, KANO model analysis, and satisfaction analysis. The results reveal that analysis methods proposed herein can effectively classify consumer demands and confirm that differences in consumer demand categories have varying impacts on consumer behavior. On this basis, corresponding design strategies are proposed based on four categories of consumer demands, aiming to make product design the leading factor and promote consumer behavior on e-commerce platforms. This research further enriches demand research on smart wearable products on e-commerce platforms, and optimizes products from a design perspective, thereby promoting consumption. In future research, different data analysis methods will be tried to compare and analyze changes in consumer demands and influencing factors, thus improving research on impact factors of product design in e-commerce.

Automatic Indexing Algorithm of Golf Video Using Audio Information (오디오 정보를 이용한 골프 동영상 자동 색인 알고리즘)

  • Kim, Hyoung-Gook
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.5
    • /
    • pp.441-446
    • /
    • 2009
  • This paper proposes an automatic indexing algorithm of golf video using audio information. In the proposed algorithm, the input audio stream is demultiplexed into the stream of video and audio. By means of Adaboost-cascade classifier, the continuous audio stream is classified into announcer's speech segment recorded in studio, music segment accompanied with players' names on TV screen, reaction segment of audience according to the play, reporter's speech segment with field background, filed noise segment like wind or waves. And golf swing sound including drive shot, iron shot, and putting shot is detected by the method of impulse onset detection and modulation spectrum verification. The detected swing and applause are used effectively to index action or highlight unit. Compared with video based semantic analysis, main advantage of the proposed system is its small computation requirement so that it facilitates to apply the technology to embedded consumer electronic devices for fast browsing.

Evaluation of the Feasibility of Deep Learning for Vegetation Monitoring (딥러닝 기반의 식생 모니터링 가능성 평가)

  • Kim, Dong-woo;Son, Seung-Woo
    • Journal of the Korean Society of Environmental Restoration Technology
    • /
    • v.26 no.6
    • /
    • pp.85-96
    • /
    • 2023
  • This study proposes a method for forest vegetation monitoring using high-resolution aerial imagery captured by unmanned aerial vehicles(UAV) and deep learning technology. The research site was selected in the forested area of Mountain Dogo, Asan City, Chungcheongnam-do, and the target species for monitoring included Pinus densiflora, Quercus mongolica, and Quercus acutissima. To classify vegetation species at the pixel level in UAV imagery based on characteristics such as leaf shape, size, and color, the study employed the semantic segmentation method using the prominent U-net deep learning model. The research results indicated that it was possible to visually distinguish Pinus densiflora Siebold & Zucc, Quercus mongolica Fisch. ex Ledeb, and Quercus acutissima Carruth in 135 aerial images captured by UAV. Out of these, 104 images were used as training data for the deep learning model, while 31 images were used for inference. The optimization of the deep learning model resulted in an overall average pixel accuracy of 92.60, with mIoU at 0.80 and FIoU at 0.82, demonstrating the successful construction of a reliable deep learning model. This study is significant as a pilot case for the application of UAV and deep learning to monitor and manage representative species among climate-vulnerable vegetation, including Pinus densiflora, Quercus mongolica, and Quercus acutissima. It is expected that in the future, UAV and deep learning models can be applied to a variety of vegetation species to better address forest management.

On the Sequences of Dialogue Acts and the Dialogue Flows-w.r.t. the appointment scheduling dialogues (대화행위의 연쇄관계와 대화흐름에 대하여 -[일정협의 대화] 중심으로)

  • 박혜은;이민행
    • Korean Journal of Cognitive Science
    • /
    • v.10 no.2
    • /
    • pp.27-34
    • /
    • 1999
  • The main purpose of this paper is to propose a general dialogue flow in 'the a appointment scheduling dialogues' in German using the concept of dialogue acts. A basic a assumption of this research is that dialogue acts contribute to the improvement of a translation system. They might be very useful to solve the problems that syntactic and semantic module could not resolve using contextual knowledge. The classification of the dialogue acts was conducted as a work of VERBMOBIL project and was based on real dialogues transcribed by experts. The real dialogues were analyzed in terms of the dialogue acts. We empirically analyzed the sequences of the dialogue acts not only in a series of dialogue turns but also in one dialogue turn. We attempted to analyZe the sequences in one dialogue turn additionally because the dialogue data used in this research showed some difference from the ones in other existing researches. By examining the sequences in dialogue acts. we proposed the dialogue flowchart in 'the a appointment scheduling dialogues' 'Based on the statistical analysis of the sequences of the most frequent dialogue acts. the dialogue flowcharts seem to represent' the a appointment scheduling dialogues' in general. A further research is required on c classification of dialogue acts which was a base for the analysis of dialogues. In order to e extract the most generalized model. we did not subcategorize each dialogue acts and used a limited number of items of dialogue acts. However. generally defined dialogue acts need to be defined more concretely and new dialogue acts for specific situations should be a added.

  • PDF

A Data Taxonomy Methodology based on Their Origin (데이터 본질 기반의 데이터 분류 방법론)

  • Choi, Mi-Young;Moon, Chang-Joo;Baik, Doo-Kwon;Kwon, Ju-Hum;Lee, Young-Moo
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.2
    • /
    • pp.163-176
    • /
    • 2010
  • The representative method to efficiently manage the organization's data is to avoid data duplication through the promotion of sharing and reusing existing data. The systematic structuring of existing data and efficient searching should be supported in order to promote the sharing and reusing of data. Without regard for these points, the data for the system development would be duplicated, which would deteriorate the quality of the data. Data taxonomy provides some methods that can enable the needed data elements to be searched quickly with a systematic order of managing data. This paper proposes that the Origin data taxonomy method can best maximize data sharing, reusing, and consolidation, and it can be used for Meta Data Registry (MDR) and Semantic Web efficiently. The Origin data taxonomy method constructs the data taxonomy structure built upon the intrinsic nature of data, so it can classify the data with independence from business classification. Also, it shows a deployment method for data elements used in various areas according to the Origin data taxonomy structure with a data taxonomic procedure that supports the proposed taxonomy. Based on this case study, the proposed data taxonomy and taxonomic procedure can be applied to real world data efficiently.

Extracting Flooded Areas in Southeast Asia Using SegNet and U-Net (SegNet과 U-Net을 활용한 동남아시아 지역 홍수탐지)

  • Kim, Junwoo;Jeon, Hyungyun;Kim, Duk-jin
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1095-1107
    • /
    • 2020
  • Flood monitoring using satellite data has been constrained by obtaining satellite images for flood peak and accurately extracting flooded areas from satellite data. Deep learning is a promising method for satellite image classification, yet the potential of deep learning-based flooded area extraction using SAR data remained uncertain, which has advantages in obtaining data, comparing to optical satellite data. This research explores the performance of SegNet and U-Net on image segmentation by extracting flooded areas in the Khorat basin, Mekong river basin, and Cagayan river basin in Thailand, Laos, and the Philippines from Sentinel-1 A/B satellite data. Results show that Global Accuracy, Mean IoU, and Mean BF Score of SegNet are 0.9847, 0.6016, and 0.6467 respectively, whereas those of U-Net are 0.9937, 0.7022, 0.7125. Visual interpretation shows that the classification accuracy of U-Net is higher than SegNet, but overall processing time of SegNet is around three times faster than that of U-Net. It is anticipated that the results of this research could be used when developing deep learning-based flood monitoring models and presenting fully automated flooded area extraction models.

The Information Modeling Method based on Extended IFC for Alignment-based Objects of Railway Track (선형중심 객체 관리를 위한 확장된 IFC 기반 철도 궤도부 정보모델링 방안)

  • Kwon, Tae Ho;Park, Sang I.;Seo, Kyung-Wan;Lee, Sang-Ho
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.31 no.6
    • /
    • pp.339-346
    • /
    • 2018
  • An Industry Foundation Classes(IFC), which is a data schema developed focusing on architecture, is being expanded to civil engineering structures. However, it is difficult to create an information model based on extended IFC since the BIM software cannot provide support functions. To manage a railway track based on the extended IFC, this paper proposed a method to create an alignment-centered separated railway track model and convert it to an extended IFC-based information model. First, railway track elements have been classified into continuous and discontinuous structures. The continuous structures were created by an alignment-based software, and discontinuous structures were created as independent objects through linkage of the discretized alignment. Second, a classification system and extended IFC schema for railway track have been proposed. Finally, the semantic information was identified by using the property of classification code and user interface. The availability of the methods was verified by developing an extended IFC-based information model of the Osong railway site.

A Study on Status of Landscape Architecture Industry with National Statistics (국가통계자료를 활용한 조경산업 현황 연구)

  • Choi, Ja-Ho;Yoon, Young-Kwan;Koo, Bon-Hak
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.50 no.5
    • /
    • pp.40-53
    • /
    • 2022
  • This study carried out to provide the methodology and basic status material of using Korean national statistics needed to find the actual state of the landscape architecture industry. The landscape architecture industry was classified into 'Design', 'Construction Management', 'construction', 'Maintenance & Management', 'Materials', 'Research', 'Education', and 'Administration' areas. In each field, business types were systemized and associated in accordance with Korean standard industrial classification and legislations pertinent to construction. Among them, the business types directly defined in the construction related legislations under the Ministry of Land, Infrastructure and Transport were focused on, and the establishment, association, integration, distribution, duplication, and omission of national statistics were analyzed. As a result, the business types of statistical analysis were selected. In order for commonality of statistical items and minimized error of interpretation, semantic analysis was conducted. Finally, the number of registered business types, the number of workers, and sales were selected. Based on them, the analysis framework applicable to fundamental analysis and evaluation of the actual state of the industry was proposed. Actual national statical data were applied for analysis and evaluation. In 2019, the number of registered business types related to the landscape architecture industry was 12,160, the number of workers by business type was 106,296, and the sales by business type were 8,308.5 billion KRW. The number of registered business types and the number of workers had been on the rise from 2017, whereas the sales had been on the decrease. It is required to come up with a plan for industrial development. This study was conducted with the national statistics established by multiple public institutions, so that there are limitations in securing consistency and reliability. Therefore, it is necessary to establish systematic and consistent national statistics in accordance with 「Landscaping Promotion Act」. In the future, it will planned to research application and development plans of national statistics according to subjects including park and green.

A Study of 'Emotion Trigger' by Text Mining Techniques (텍스트 마이닝을 이용한 감정 유발 요인 'Emotion Trigger'에 관한 연구)

  • An, Juyoung;Bae, Junghwan;Han, Namgi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.2
    • /
    • pp.69-92
    • /
    • 2015
  • The explosion of social media data has led to apply text-mining techniques to analyze big social media data in a more rigorous manner. Even if social media text analysis algorithms were improved, previous approaches to social media text analysis have some limitations. In the field of sentiment analysis of social media written in Korean, there are two typical approaches. One is the linguistic approach using machine learning, which is the most common approach. Some studies have been conducted by adding grammatical factors to feature sets for training classification model. The other approach adopts the semantic analysis method to sentiment analysis, but this approach is mainly applied to English texts. To overcome these limitations, this study applies the Word2Vec algorithm which is an extension of the neural network algorithms to deal with more extensive semantic features that were underestimated in existing sentiment analysis. The result from adopting the Word2Vec algorithm is compared to the result from co-occurrence analysis to identify the difference between two approaches. The results show that the distribution related word extracted by Word2Vec algorithm in that the words represent some emotion about the keyword used are three times more than extracted by co-occurrence analysis. The reason of the difference between two results comes from Word2Vec's semantic features vectorization. Therefore, it is possible to say that Word2Vec algorithm is able to catch the hidden related words which have not been found in traditional analysis. In addition, Part Of Speech (POS) tagging for Korean is used to detect adjective as "emotional word" in Korean. In addition, the emotion words extracted from the text are converted into word vector by the Word2Vec algorithm to find related words. Among these related words, noun words are selected because each word of them would have causal relationship with "emotional word" in the sentence. The process of extracting these trigger factor of emotional word is named "Emotion Trigger" in this study. As a case study, the datasets used in the study are collected by searching using three keywords: professor, prosecutor, and doctor in that these keywords contain rich public emotion and opinion. Advanced data collecting was conducted to select secondary keywords for data gathering. The secondary keywords for each keyword used to gather the data to be used in actual analysis are followed: Professor (sexual assault, misappropriation of research money, recruitment irregularities, polifessor), Doctor (Shin hae-chul sky hospital, drinking and plastic surgery, rebate) Prosecutor (lewd behavior, sponsor). The size of the text data is about to 100,000(Professor: 25720, Doctor: 35110, Prosecutor: 43225) and the data are gathered from news, blog, and twitter to reflect various level of public emotion into text data analysis. As a visualization method, Gephi (http://gephi.github.io) was used and every program used in text processing and analysis are java coding. The contributions of this study are as follows: First, different approaches for sentiment analysis are integrated to overcome the limitations of existing approaches. Secondly, finding Emotion Trigger can detect the hidden connections to public emotion which existing method cannot detect. Finally, the approach used in this study could be generalized regardless of types of text data. The limitation of this study is that it is hard to say the word extracted by Emotion Trigger processing has significantly causal relationship with emotional word in a sentence. The future study will be conducted to clarify the causal relationship between emotional words and the words extracted by Emotion Trigger by comparing with the relationships manually tagged. Furthermore, the text data used in Emotion Trigger are twitter, so the data have a number of distinct features which we did not deal with in this study. These features will be considered in further study.

A Study on Market Size Estimation Method by Product Group Using Word2Vec Algorithm (Word2Vec을 활용한 제품군별 시장규모 추정 방법에 관한 연구)

  • Jung, Ye Lim;Kim, Ji Hui;Yoo, Hyoung Sun
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.1-21
    • /
    • 2020
  • With the rapid development of artificial intelligence technology, various techniques have been developed to extract meaningful information from unstructured text data which constitutes a large portion of big data. Over the past decades, text mining technologies have been utilized in various industries for practical applications. In the field of business intelligence, it has been employed to discover new market and/or technology opportunities and support rational decision making of business participants. The market information such as market size, market growth rate, and market share is essential for setting companies' business strategies. There has been a continuous demand in various fields for specific product level-market information. However, the information has been generally provided at industry level or broad categories based on classification standards, making it difficult to obtain specific and proper information. In this regard, we propose a new methodology that can estimate the market sizes of product groups at more detailed levels than that of previously offered. We applied Word2Vec algorithm, a neural network based semantic word embedding model, to enable automatic market size estimation from individual companies' product information in a bottom-up manner. The overall process is as follows: First, the data related to product information is collected, refined, and restructured into suitable form for applying Word2Vec model. Next, the preprocessed data is embedded into vector space by Word2Vec and then the product groups are derived by extracting similar products names based on cosine similarity calculation. Finally, the sales data on the extracted products is summated to estimate the market size of the product groups. As an experimental data, text data of product names from Statistics Korea's microdata (345,103 cases) were mapped in multidimensional vector space by Word2Vec training. We performed parameters optimization for training and then applied vector dimension of 300 and window size of 15 as optimized parameters for further experiments. We employed index words of Korean Standard Industry Classification (KSIC) as a product name dataset to more efficiently cluster product groups. The product names which are similar to KSIC indexes were extracted based on cosine similarity. The market size of extracted products as one product category was calculated from individual companies' sales data. The market sizes of 11,654 specific product lines were automatically estimated by the proposed model. For the performance verification, the results were compared with actual market size of some items. The Pearson's correlation coefficient was 0.513. Our approach has several advantages differing from the previous studies. First, text mining and machine learning techniques were applied for the first time on market size estimation, overcoming the limitations of traditional sampling based- or multiple assumption required-methods. In addition, the level of market category can be easily and efficiently adjusted according to the purpose of information use by changing cosine similarity threshold. Furthermore, it has a high potential of practical applications since it can resolve unmet needs for detailed market size information in public and private sectors. Specifically, it can be utilized in technology evaluation and technology commercialization support program conducted by governmental institutions, as well as business strategies consulting and market analysis report publishing by private firms. The limitation of our study is that the presented model needs to be improved in terms of accuracy and reliability. The semantic-based word embedding module can be advanced by giving a proper order in the preprocessed dataset or by combining another algorithm such as Jaccard similarity with Word2Vec. Also, the methods of product group clustering can be changed to other types of unsupervised machine learning algorithm. Our group is currently working on subsequent studies and we expect that it can further improve the performance of the conceptually proposed basic model in this study.