• 제목/요약/키워드: Extracting characteristics

검색결과 587건 처리시간 0.023초

Study on the Possibility of Estimating Surface Soil Moisture Using Sentinel-1 SAR Satellite Imagery Based on Google Earth Engine (Google Earth Engine 기반 Sentinel-1 SAR 위성영상을 이용한 지표 토양수분량 산정 가능성에 관한 연구)

  • Younghyun Cho
    • Korean Journal of Remote Sensing
    • /
    • 제40권2호
    • /
    • pp.229-241
    • /
    • 2024
  • With the advancement of big data processing technology using cloud platforms, access, processing, and analysis of large-volume data such as satellite imagery have recently been significantly improved. In this study, the Change Detection Method, a relatively simple technique for retrieving soil moisture, was applied to the backscattering coefficient values of pre-processed Sentinel-1 synthetic aperture radar (SAR) satellite imagery product based on Google Earth Engine (GEE), one of those platforms, to estimate the surface soil moisture for six observatories within the Yongdam Dam watershed in South Korea for the period of 2015 to 2023, as well as the watershed average. Subsequently, a correlation analysis was conducted between the estimated values and actual measurements, along with an examination of the applicability of GEE. The results revealed that the surface soil moisture estimated for small areas within the soil moisture observatories of the watershed exhibited low correlations ranging from 0.1 to 0.3 for both VH and VV polarizations, likely due to the inherent measurement accuracy of the SAR satellite imagery and variations in data characteristics. However, the surface soil moisture average, which was derived by extracting the average SAR backscattering coefficient values for the entire watershed area and applying moving averages to mitigate data uncertainties and variability, exhibited significantly improved results at the level of 0.5. The results obtained from estimating soil moisture using GEE demonstrate its utility despite limitations in directly conducting desired analyses due to preprocessed SAR data. However, the efficient processing of extensive satellite imagery data allows for the estimation and evaluation of soil moisture over broad ranges, such as long-term watershed averages. This highlights the effectiveness of GEE in handling vast satellite imagery datasets to assess soil moisture. Based on this, it is anticipated that GEE can be effectively utilized to assess long-term variations of soil moisture average in major dam watersheds, in conjunction with soil moisture observation data from various locations across the country in the future.

Analysis of Micro-Sedimentary Structure Characteristics Using Ultra-High Resolution UAV Imagery: Hwangdo Tidal Flat, South Korea (초고해상도 무인항공기 영상을 이용한 한국 황도 갯벌의 미세 퇴적 구조 특성 분석)

  • Minju Kim;Won-Kyung Baek;Hoi Soo Jung;Joo-Hyung Ryu
    • Korean Journal of Remote Sensing
    • /
    • 제40권3호
    • /
    • pp.295-305
    • /
    • 2024
  • This study aims to analyze the micro-sedimentary structures of the Hwangdo tidal flats using ultra-high resolution unmanned aerial vehicle (UAV) data. Tidal flats, located in the transitional area between land and sea, constantly change due to tidal activities and provide a unique environment important for understanding sedimentary processes and environmental conditions. Traditional field observation methods are limited in spatial and temporal coverage, and existing satellite imagery does not provide sufficient resolution to study micro-sedimentary structures. To overcome these limitations, high-resolution images of the Hwangdo tidal flats in Chungcheongnam-do were acquired using UAVs. This area has experienced significant changes in its sedimentary environment due to coastal development projects such as sea wall construction. From May 17 to 18, 2022, sediment samples were collected from 91 points during field surveys and 25 in-situ points were intensively analyzed. UAV data with a spatial resolution of approximately 0.9 mm allowed identifying and extracting parameters related to micro-sedimentary structures. For mud cracks, the length of the major axis of the polygons was extracted, and the wavelength and ripple symmetry index were extracted for ripple marks. The results of the study showed that in areas with mud content above 80%, mud cracks formed at an average major axis length of 37.3 cm. In regions with sand content above 60%, ripples with an average wavelength of 8 cm and a ripple symmetry index of 2.0 were formed. This study demonstrated that micro-sedimentary structures of tidal flats can be effectively analyzed using ultra-high resolution UAV data without field surveys. This highlights the potential of UAV technology as an important tool in environmental monitoring and coastal management and shows its usefulness in the study of sedimentary structures. In addition, the results of this study are expected to serve as baseline data for more accurate sedimentary facies classification.

Motives for Writing After-Purchase Consumer Reviews in Online Stores and Classification of Online Store Shoppers (인터넷 점포에서의 구매후기 작성 동기 및 점포 고객 유형화)

  • Hong, Hee-Sook;Ryu, Sung-Min
    • Journal of Distribution Research
    • /
    • 제17권3호
    • /
    • pp.25-57
    • /
    • 2012
  • This study identified motives for writing apparel product reviews in online stores, and determined what motives increase the behavior of writing reviews. It also classified store customers based on the type of writing motives, and clarified the characteristics of internet purchase behavior and of a demographic profile. Data were collected from 252 females aged 20s' and 30s' who have experience of reading and writing reviews on online shopping. The five types of writing motives were altruistic information sharing, remedying of a grievance and vengeance, economic incentives, helping new product development, and the expression of satisfaction feelings. Among five motives, altruistic information sharing, economic incentives, and helping new product development stimulate writing reviews. Store customers who write reviews were classified into three groups based on their writing motive types: Other consumer advocates(29.8%), self-interested shoppers(40.5%) and shoppers with moderate motives(29.8%). There were significant differences among three groups in writing behavior (the frequency of writing reviews, writing intent of reviews, duration of writing reviews, and frequency of online shopping) and age. Based on results, managerial implications were suggested. Long Abstract : The purpose of present study is to identify the types of writing motives on online shopping, and to clarify the motives affecting the behavior of writing reviews. This study also classifies online shoppers based on the motive types, and identifies the characteristics of the classified groups in terms of writing behavior, frequency of online shopping, and demographics. Use and Gratification Theory was adopted in this study. Qualitative research (focus group interview) and quantitative research were used. Korean women(20 to 39 years old) who reported experience with purchasing clothing online, and reading and writing reviews were selected as samples(n=252). Most of the respondents were relatively young (20-34yrs., 86.1%,), single (61.1%), employed(61.1%) and residents living in big cities(50.9%). About 69.8% of respondents read and 40.5% write apparel reviews frequently or very frequently. 24.6% of the respondents indicated an "average" in their writing frequency. Based on the qualitative result of focus group interviews and previous studies on motives for online community activities, measurement items of motives for writing after-purchase reviews were developed. All items were used a five-point Likert scale with endpoints 1 (strongly disagree) and 5 (strongly agree). The degree of writing behavior was measured by items concerning experience of writing reviews, frequency of writing reviews, amount of writing reviews, and intention of writing reviews. A five-point scale(strongly disagree-strongly agree) was employed. SPSS 18.0 was used for exploratory factor analysis, K-means cluster analysis, one-way ANOVA(Scheffe test) and ${\chi}^2$-test. Confirmatory factor analysis and path model analysis were conducted by AMOS 18.0. By conducting principal components factor analysis (varimax rotation, extracting factors with eigenvalues above 1.0) on the measurement items, five factors were identified: Altruistic information sharing, remedying of a grievance and vengeance, economic incentives, helping new product development, and expression of satisfaction feelings(see Table 1). The measurement model including these final items was analyzed by confirmatory factor analysis. The measurement model had good fit indices(GFI=.918, AGFI=.884, RMR=.070, RMSEA=.054, TLI=.941) except for the probability value associated with the ${\chi}^2$ test(${\chi}^2$=189.078, df=109, p=.00). Convergent validities of all variables were confirmed using composite reliability. All SMC values were found to be lower than AVEs confirming discriminant validity. The path model's goodness-of-fit was greater than the recommended limits based on several indices(GFI=.905, AGFI=.872, RMR=.070, RMSEA=.052, TLI=.935; ${\chi}^2$=260.433, df=155, p=.00). Table 2 shows that motives of altruistic information sharing, economic incentives and helping new product development significantly increased the degree of writing product reviews of online shopping. In particular, the effect of altruistic information sharing and pursuit of economic incentives on the behavior of writing reviews were larger than the effect of helping new product development. As shown in table 3, online store shoppers were classified into three groups: Other consumer advocates (29.8%), self-interested shoppers (40.5%), and moderate shoppers (29.8%). There were significant differences among the three groups in the degree of writing reviews (experience of writing reviews, frequency of writing reviews, amount of writing reviews, intention of writing reviews, and duration of writing reviews, frequency of online shopping) and age. For five aspects of writing behavior, the group of other consumer advocates who is mainly comprised of 20s had higher scores than the other two groups. There were not any significant differences between self-interested group and moderate group regarding writing behavior and demographics.

  • PDF

Construction of Event Networks from Large News Data Using Text Mining Techniques (텍스트 마이닝 기법을 적용한 뉴스 데이터에서의 사건 네트워크 구축)

  • Lee, Minchul;Kim, Hea-Jin
    • Journal of Intelligence and Information Systems
    • /
    • 제24권1호
    • /
    • pp.183-203
    • /
    • 2018
  • News articles are the most suitable medium for examining the events occurring at home and abroad. Especially, as the development of information and communication technology has brought various kinds of online news media, the news about the events occurring in society has increased greatly. So automatically summarizing key events from massive amounts of news data will help users to look at many of the events at a glance. In addition, if we build and provide an event network based on the relevance of events, it will be able to greatly help the reader in understanding the current events. In this study, we propose a method for extracting event networks from large news text data. To this end, we first collected Korean political and social articles from March 2016 to March 2017, and integrated the synonyms by leaving only meaningful words through preprocessing using NPMI and Word2Vec. Latent Dirichlet allocation (LDA) topic modeling was used to calculate the subject distribution by date and to find the peak of the subject distribution and to detect the event. A total of 32 topics were extracted from the topic modeling, and the point of occurrence of the event was deduced by looking at the point at which each subject distribution surged. As a result, a total of 85 events were detected, but the final 16 events were filtered and presented using the Gaussian smoothing technique. We also calculated the relevance score between events detected to construct the event network. Using the cosine coefficient between the co-occurred events, we calculated the relevance between the events and connected the events to construct the event network. Finally, we set up the event network by setting each event to each vertex and the relevance score between events to the vertices connecting the vertices. The event network constructed in our methods helped us to sort out major events in the political and social fields in Korea that occurred in the last one year in chronological order and at the same time identify which events are related to certain events. Our approach differs from existing event detection methods in that LDA topic modeling makes it possible to easily analyze large amounts of data and to identify the relevance of events that were difficult to detect in existing event detection. We applied various text mining techniques and Word2vec technique in the text preprocessing to improve the accuracy of the extraction of proper nouns and synthetic nouns, which have been difficult in analyzing existing Korean texts, can be found. In this study, the detection and network configuration techniques of the event have the following advantages in practical application. First, LDA topic modeling, which is unsupervised learning, can easily analyze subject and topic words and distribution from huge amount of data. Also, by using the date information of the collected news articles, it is possible to express the distribution by topic in a time series. Second, we can find out the connection of events in the form of present and summarized form by calculating relevance score and constructing event network by using simultaneous occurrence of topics that are difficult to grasp in existing event detection. It can be seen from the fact that the inter-event relevance-based event network proposed in this study was actually constructed in order of occurrence time. It is also possible to identify what happened as a starting point for a series of events through the event network. The limitation of this study is that the characteristics of LDA topic modeling have different results according to the initial parameters and the number of subjects, and the subject and event name of the analysis result should be given by the subjective judgment of the researcher. Also, since each topic is assumed to be exclusive and independent, it does not take into account the relevance between themes. Subsequent studies need to calculate the relevance between events that are not covered in this study or those that belong to the same subject.

Image Watermarking for Copyright Protection of Images on Shopping Mall (쇼핑몰 이미지 저작권보호를 위한 영상 워터마킹)

  • Bae, Kyoung-Yul
    • Journal of Intelligence and Information Systems
    • /
    • 제19권4호
    • /
    • pp.147-157
    • /
    • 2013
  • With the advent of the digital environment that can be accessed anytime, anywhere with the introduction of high-speed network, the free distribution and use of digital content were made possible. Ironically this environment is raising a variety of copyright infringement, and product images used in the online shopping mall are pirated frequently. There are many controversial issues whether shopping mall images are creative works or not. According to Supreme Court's decision in 2001, to ad pictures taken with ham products is simply a clone of the appearance of objects to deliver nothing but the decision was not only creative expression. But for the photographer's losses recognized in the advertising photo shoot takes the typical cost was estimated damages. According to Seoul District Court precedents in 2003, if there are the photographer's personality and creativity in the selection of the subject, the composition of the set, the direction and amount of light control, set the angle of the camera, shutter speed, shutter chance, other shooting methods for capturing, developing and printing process, the works should be protected by copyright law by the Court's sentence. In order to receive copyright protection of the shopping mall images by the law, it is simply not to convey the status of the product, the photographer's personality and creativity can be recognized that it requires effort. Accordingly, the cost of making the mall image increases, and the necessity for copyright protection becomes higher. The product images of the online shopping mall have a very unique configuration unlike the general pictures such as portraits and landscape photos and, therefore, the general image watermarking technique can not satisfy the requirements of the image watermarking. Because background of product images commonly used in shopping malls is white or black, or gray scale (gradient) color, it is difficult to utilize the space to embed a watermark and the area is very sensitive even a slight change. In this paper, the characteristics of images used in shopping malls are analyzed and a watermarking technology which is suitable to the shopping mall images is proposed. The proposed image watermarking technology divide a product image into smaller blocks, and the corresponding blocks are transformed by DCT (Discrete Cosine Transform), and then the watermark information was inserted into images using quantization of DCT coefficients. Because uniform treatment of the DCT coefficients for quantization cause visual blocking artifacts, the proposed algorithm used weighted mask which quantizes finely the coefficients located block boundaries and coarsely the coefficients located center area of the block. This mask improves subjective visual quality as well as the objective quality of the images. In addition, in order to improve the safety of the algorithm, the blocks which is embedded the watermark are randomly selected and the turbo code is used to reduce the BER when extracting the watermark. The PSNR(Peak Signal to Noise Ratio) of the shopping mall image watermarked by the proposed algorithm is 40.7~48.5[dB] and BER(Bit Error Rate) after JPEG with QF = 70 is 0. This means the watermarked image is high quality and the algorithm is robust to JPEG compression that is used generally at the online shopping malls. Also, for 40% change in size and 40 degrees of rotation, the BER is 0. In general, the shopping malls are used compressed images with QF which is higher than 90. Because the pirated image is used to replicate from original image, the proposed algorithm can identify the copyright infringement in the most cases. As shown the experimental results, the proposed algorithm is suitable to the shopping mall images with simple background. However, the future study should be carried out to enhance the robustness of the proposed algorithm because the robustness loss is occurred after mask process.

The Framework of Research Network and Performance Evaluation on Personal Information Security: Social Network Analysis Perspective (개인정보보호 분야의 연구자 네트워크와 성과 평가 프레임워크: 소셜 네트워크 분석을 중심으로)

  • Kim, Minsu;Choi, Jaewon;Kim, Hyun Jin
    • Journal of Intelligence and Information Systems
    • /
    • 제20권1호
    • /
    • pp.177-193
    • /
    • 2014
  • Over the past decade, there has been a rapid diffusion of electronic commerce and a rising number of interconnected networks, resulting in an escalation of security threats and privacy concerns. Electronic commerce has a built-in trade-off between the necessity of providing at least some personal information to consummate an online transaction, and the risk of negative consequences from providing such information. More recently, the frequent disclosure of private information has raised concerns about privacy and its impacts. This has motivated researchers in various fields to explore information privacy issues to address these concerns. Accordingly, the necessity for information privacy policies and technologies for collecting and storing data, and information privacy research in various fields such as medicine, computer science, business, and statistics has increased. The occurrence of various information security accidents have made finding experts in the information security field an important issue. Objective measures for finding such experts are required, as it is currently rather subjective. Based on social network analysis, this paper focused on a framework to evaluate the process of finding experts in the information security field. We collected data from the National Discovery for Science Leaders (NDSL) database, initially collecting about 2000 papers covering the period between 2005 and 2013. Outliers and the data of irrelevant papers were dropped, leaving 784 papers to test the suggested hypotheses. The co-authorship network data for co-author relationship, publisher, affiliation, and so on were analyzed using social network measures including centrality and structural hole. The results of our model estimation are as follows. With the exception of Hypothesis 3, which deals with the relationship between eigenvector centrality and performance, all of our hypotheses were supported. In line with our hypothesis, degree centrality (H1) was supported with its positive influence on the researchers' publishing performance (p<0.001). This finding indicates that as the degree of cooperation increased, the more the publishing performance of researchers increased. In addition, closeness centrality (H2) was also positively associated with researchers' publishing performance (p<0.001), suggesting that, as the efficiency of information acquisition increased, the more the researchers' publishing performance increased. This paper identified the difference in publishing performance among researchers. The analysis can be used to identify core experts and evaluate their performance in the information privacy research field. The co-authorship network for information privacy can aid in understanding the deep relationships among researchers. In addition, extracting characteristics of publishers and affiliations, this paper suggested an understanding of the social network measures and their potential for finding experts in the information privacy field. Social concerns about securing the objectivity of experts have increased, because experts in the information privacy field frequently participate in political consultation, and business education support and evaluation. In terms of practical implications, this research suggests an objective framework for experts in the information privacy field, and is useful for people who are in charge of managing research human resources. This study has some limitations, providing opportunities and suggestions for future research. Presenting the difference in information diffusion according to media and proximity presents difficulties for the generalization of the theory due to the small sample size. Therefore, further studies could consider an increased sample size and media diversity, the difference in information diffusion according to the media type, and information proximity could be explored in more detail. Moreover, previous network research has commonly observed a causal relationship between the independent and dependent variable (Kadushin, 2012). In this study, degree centrality as an independent variable might have causal relationship with performance as a dependent variable. However, in the case of network analysis research, network indices could be computed after the network relationship is created. An annual analysis could help mitigate this limitation.

Query-based Answer Extraction using Korean Dependency Parsing (의존 구문 분석을 이용한 질의 기반 정답 추출)

  • Lee, Dokyoung;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • 제25권3호
    • /
    • pp.161-177
    • /
    • 2019
  • In this paper, we study the performance improvement of the answer extraction in Question-Answering system by using sentence dependency parsing result. The Question-Answering (QA) system consists of query analysis, which is a method of analyzing the user's query, and answer extraction, which is a method to extract appropriate answers in the document. And various studies have been conducted on two methods. In order to improve the performance of answer extraction, it is necessary to accurately reflect the grammatical information of sentences. In Korean, because word order structure is free and omission of sentence components is frequent, dependency parsing is a good way to analyze Korean syntax. Therefore, in this study, we improved the performance of the answer extraction by adding the features generated by dependency parsing analysis to the inputs of the answer extraction model (Bidirectional LSTM-CRF). The process of generating the dependency graph embedding consists of the steps of generating the dependency graph from the dependency parsing result and learning the embedding of the graph. In this study, we compared the performance of the answer extraction model when inputting basic word features generated without the dependency parsing and the performance of the model when inputting the addition of the Eojeol tag feature and dependency graph embedding feature. Since dependency parsing is performed on a basic unit of an Eojeol, which is a component of sentences separated by a space, the tag information of the Eojeol can be obtained as a result of the dependency parsing. The Eojeol tag feature means the tag information of the Eojeol. The process of generating the dependency graph embedding consists of the steps of generating the dependency graph from the dependency parsing result and learning the embedding of the graph. From the dependency parsing result, a graph is generated from the Eojeol to the node, the dependency between the Eojeol to the edge, and the Eojeol tag to the node label. In this process, an undirected graph is generated or a directed graph is generated according to whether or not the dependency relation direction is considered. To obtain the embedding of the graph, we used Graph2Vec, which is a method of finding the embedding of the graph by the subgraphs constituting a graph. We can specify the maximum path length between nodes in the process of finding subgraphs of a graph. If the maximum path length between nodes is 1, graph embedding is generated only by direct dependency between Eojeol, and graph embedding is generated including indirect dependencies as the maximum path length between nodes becomes larger. In the experiment, the maximum path length between nodes is adjusted differently from 1 to 3 depending on whether direction of dependency is considered or not, and the performance of answer extraction is measured. Experimental results show that both Eojeol tag feature and dependency graph embedding feature improve the performance of answer extraction. In particular, considering the direction of the dependency relation and extracting the dependency graph generated with the maximum path length of 1 in the subgraph extraction process in Graph2Vec as the input of the model, the highest answer extraction performance was shown. As a result of these experiments, we concluded that it is better to take into account the direction of dependence and to consider only the direct connection rather than the indirect dependence between the words. The significance of this study is as follows. First, we improved the performance of answer extraction by adding features using dependency parsing results, taking into account the characteristics of Korean, which is free of word order structure and omission of sentence components. Second, we generated feature of dependency parsing result by learning - based graph embedding method without defining the pattern of dependency between Eojeol. Future research directions are as follows. In this study, the features generated as a result of the dependency parsing are applied only to the answer extraction model in order to grasp the meaning. However, in the future, if the performance is confirmed by applying the features to various natural language processing models such as sentiment analysis or name entity recognition, the validity of the features can be verified more accurately.