• 제목/요약/키워드: euclidean distance

Search Result 557, Processing Time 0.023 seconds

Implementation of DTW-kNN-based Decision Support System for Discriminating Emerging Technologies (DTW-kNN 기반의 유망 기술 식별을 위한 의사결정 지원 시스템 구현 방안)

  • Jeong, Do-Heon;Park, Ju-Yeon
    • Journal of Industrial Convergence
    • /
    • v.20 no.8
    • /
    • pp.77-84
    • /
    • 2022
  • This study aims to present a method for implementing a decision support system that can be used for selecting emerging technologies by applying a machine learning-based automatic classification technique. To conduct the research, the architecture of the entire system was built and detailed research steps were conducted. First, emerging technology candidate items were selected and trend data was automatically generated using a big data system. After defining the conceptual model and pattern classification structure of technological development, an efficient machine learning method was presented through an automatic classification experiment. Finally, the analysis results of the system were interpreted and methods for utilization were derived. In a DTW-kNN-based classification experiment that combines the Dynamic Time Warping(DTW) method and the k-Nearest Neighbors(kNN) classification model proposed in this study, the identification performance was up to 87.7%, and particularly in the 'eventual' section where the trend highly fluctuates, the maximum performance difference was 39.4% points compared to the Euclidean Distance(ED) algorithm. In addition, through the analysis results presented by the system, it was confirmed that this decision support system can be effectively utilized in the process of automatically classifying and filtering by type with a large amount of trend data.

A Study on the Method of Scholarly Paper Recommendation Using Multidimensional Metadata Space (다차원 메타데이터 공간을 활용한 학술 문헌 추천기법 연구)

  • Miah Kam;Jee Yeon Lee
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.1
    • /
    • pp.121-148
    • /
    • 2023
  • The purpose of this study is to propose a scholarly paper recommendation system based on metadata attribute similarity with excellent performance. This study suggests a scholarly paper recommendation method that combines techniques from two sub-fields of Library and Information Science, namely metadata use in Information Organization and co-citation analysis, author bibliographic coupling, co-occurrence frequency, and cosine similarity in Bibliometrics. To conduct experiments, a total of 9,643 paper metadata related to "inequality" and "divide" were collected and refined to derive relative coordinate values between author, keyword, and title attributes using cosine similarity. The study then conducted experiments to select weight conditions and dimension numbers that resulted in a good performance. The results were presented and evaluated by users, and based on this, the study conducted discussions centered on the research questions through reference node and recommendation combination characteristic analysis, conjoint analysis, and results from comparative analysis. Overall, the study showed that the performance was excellent when author-related attributes were used alone or in combination with title-related attributes. If the technique proposed in this study is utilized and a wide range of samples are secured, it could help improve the performance of recommendation techniques not only in the field of literature recommendation in information services but also in various other fields in society.

Development for rainfall classification based on local flood vulnerability using entropy weight in Seoul metropolitan area (엔트로피 가중치를 활용한 지역별 홍수취약도 기반의 서울지역 강우기준 산정기법)

  • Lee, Seonmi;Choi, Youngje;Lee, Eunkyung;Ji, Jungwon;Yi, Jaeeung
    • Journal of Korea Water Resources Association
    • /
    • v.55 no.4
    • /
    • pp.267-278
    • /
    • 2022
  • Recently Flood damage volume has increased as heavy rain has frequently occurred. Especially urban areas are a vulnerability to flooding damage because of densely concentrated population and property. A local government is preparing to mitigate flood damage through the heavy rain warning issued by Korea Meteorological Administration. This warning classification is identical for a national scale. However, Seoul has 25 administrative districts with different regional characteristics such as climate, topography, disaster prevention state, and flood damage severity. This study considered the regional characteristics of 25 administrative districts to analyze the flood vulnerability using entropy weight and Euclidean distance. The rainfall classification was derived based on probability rainfall and flood damage rainfall that occurred in the past. The result shows the step 2 and step 4 of rainfall classification was not significantly different from the heavy rain classification of the Korea Meteorological Administration. The flood vulnerability is high with high climate exposure and low adaptability to climate change, and the rainfall classification is low in the northern region of Seoul. It is possible to preemptively respond to floods in the northern region of Seoul based on relatively low rainfall classification. In the future, we plan to review the applicability of rainfall forecast data using the rainfall classification of results from this study. These results will contribute to research for preemptive flood response measures.

Managing the Reverse Extrapolation Model of Radar Threats Based Upon an Incremental Machine Learning Technique (점진적 기계학습 기반의 레이더 위협체 역추정 모델 생성 및 갱신)

  • Kim, Chulpyo;Noh, Sanguk
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.13 no.4
    • /
    • pp.29-39
    • /
    • 2017
  • Various electronic warfare situations drive the need to develop an integrated electronic warfare simulator that can perform electronic warfare modeling and simulation on radar threats. In this paper, we analyze the components of a simulation system to reversely model the radar threats that emit electromagnetic signals based on the parameters of the electronic information, and propose a method to gradually maintain the reverse extrapolation model of RF threats. In the experiment, we will evaluate the effectiveness of the incremental model update and also assess the integration method of reverse extrapolation models. The individual model of RF threats are constructed by using decision tree, naive Bayesian classifier, artificial neural network, and clustering algorithms through Euclidean distance and cosine similarity measurement, respectively. Experimental results show that the accuracy of reverse extrapolation models improves, while the size of the threat sample increases. In addition, we use voting, weighted voting, and the Dempster-Shafer algorithm to integrate the results of the five different models of RF threats. As a result, the final decision of reverse extrapolation through the Dempster-Shafer algorithm shows the best performance in its accuracy.

Contents and Sequences for Line Segments, Straight Lines, and Rays in Elementary Mathematics Curricula and Textbooks (선분, 직선, 반직선의 학습 내용과 학습 계열 분석)

  • Kim, Sangmee
    • Communications of Mathematical Education
    • /
    • v.37 no.4
    • /
    • pp.635-652
    • /
    • 2023
  • This study conducts a comprehensive analysis of the curricular progression of the concepts and learning sequences of 'lines', specifically, 'line segments', 'straight lines', and 'rays', at the elementary school level. By examining mathematics curricula and textbooks, spanning from 2nd to 7th and 2007, 2009, 2015, and up to 2022 revised version, the study investigates the timing and methods of introducing these essential geometric concepts. It also explores the sequential delivery of instruction and the key focal points of pedagogy. Through the analysis of shifts in the timing and definitions, it becomes evident that these concepts of lines have predominantly been integrated as integral components of two-dimensional plane figures. This includes their role in defining the sides of polygons and the angles formed by lines. This perspective underscores the importance of providing ample opportunities for students to explore these basic geometric entities. Furthermore, the definitions of line segments, straight lines, and rays, their interrelations with points, and the relationships established between different types of lines significantly influence the development of these core concepts. Lastly, the study emphasizes the significance of introducing fundamental mathematical concepts, such as the notion of straight lines as the shortest distance in line segments and the concept of lines extending infinitely (infiniteness) in straight lines and rays. These ideas serve as foundational elements of mathematical thinking, emphasizing the necessity for students to grasp concretely these concepts through visualization and experiences in their daily surroundings. This progression aligns with a shift towards the comprehension of Euclidean geometry. This research suggests a comprehensive reassessment of how line concepts are introduced and taught, with a particular focus on connecting real-life exploratory experiences to the foundational principles of geometry, thereby enhancing the quality of mathematics education.

Histogram-Based Singular Value Decomposition for Object Identification and Tracking (객체 식별 및 추적을 위한 히스토그램 기반 특이값 분해)

  • Ye-yeon Kang;Jeong-Min Park;HoonJoon Kouh;Kyungyong Chung
    • Journal of Internet Computing and Services
    • /
    • v.24 no.5
    • /
    • pp.29-35
    • /
    • 2023
  • CCTV is used for various purposes such as crime prevention, public safety reinforcement, and traffic management. However, as the range and resolution of the camera improve, there is a risk of exposing personal information in the video. Therefore, there is a need for new technologies that can identify individuals while protecting personal information in images. In this paper, we propose histogram-based singular value decomposition for object identification and tracking. The proposed method distinguishes different objects present in the image using color information of the object. For object recognition, YOLO and DeepSORT are used to detect and extract people present in the image. Color values are extracted with a black-and-white histogram using location information of the detected person. Singular value decomposition is used to extract and use only meaningful information among the extracted color values. When using singular value decomposition, the accuracy of object color extraction is increased by using the average of the upper singular value in the result. Color information extracted using singular value decomposition is compared with colors present in other images, and the same person present in different images is detected. Euclidean distance is used for color information comparison, and Top-N is used for accuracy evaluation. As a result of the evaluation, when detecting the same person using a black-and-white histogram and singular value decomposition, it recorded a maximum of 100% to a minimum of 74%.

A study on the Analysis of Locational Characteristics of REITs Assets (운영부동산 유형별 리츠자산의 입지특성 분석에 관한 연구)

  • Jung Jaeyeon;Lee Changsoo
    • Journal of the Korean Regional Science Association
    • /
    • v.40 no.1
    • /
    • pp.89-110
    • /
    • 2024
  • REITs are very closely related to real estate management, but there have been no prior studies analyzing the location of REITs assets. Therefore, this study analyzed the location characteristics of REITs assets in two aspects to clarify the location characteristics by using spatial information of REITs assets. First, the characteristics of the type of city where REITs assets are distributed were analyzed, and second, the characteristics of the zoning where REITs assets are distributed were analyzed. As a result of analyzing the characteristics of the city where REITs assets are distributed by type, it was analyzed that in the case of the capital area, both the ratio of cities with REITs assets location and the intensity of REITs assets location (number of REITs assets per city) have location characteristics by city hierarchy in the order of metropolitan city > big city > small and medium-sized city. In the case of non-capital area's metropolitan and large cities, the ratio of REITs assets location cities is similar to that of the capital area, but the location intensity of REITs assets was analyzed to be significantly lower than that of the capital area. As a result of the analysis of REITs assets by type, housing REITs assets tend to be located in the old downtown commercial zoning and the new downtown residential zoning, office REITs assets are characterized by concentration of location in specific commercial zoning of Seoul, and retail REITs assets are located mainly in the old downtown station area. In addition, it was found that logistics REITs assets tend to be located in management zoning, centering on key logistics hub cities in the region.

Development of an Automatic 3D Coregistration Technique of Brain PET and MR Images (뇌 PET과 MR 영상의 자동화된 3차원적 합성기법 개발)

  • Lee, Jae-Sung;Kwark, Cheol-Eun;Lee, Dong-Soo;Chung, June-Key;Lee, Myung-Chul;Park, Kwang-Suk
    • The Korean Journal of Nuclear Medicine
    • /
    • v.32 no.5
    • /
    • pp.414-424
    • /
    • 1998
  • Purpose: Cross-modality coregistration of positron emission tomography (PET) and magnetic resonance imaging (MR) could enhance the clinical information. In this study we propose a refined technique to improve the robustness of registration, and to implement more realistic visualization of the coregistered images. Materials and Methods: Using the sinogram of PET emission scan, we extracted the robust head boundary and used boundary-enhanced PET to coregister PET with MR. The pixels having 10% of maximum pixel value were considered as the boundary of sinogram. Boundary pixel values were exchanged with maximum value of sinogram. One hundred eighty boundary points were extracted at intervals of about 2 degree using simple threshold method from each slice of MR images. Best affined transformation between the two point sets was performed using least square fitting which should minimize the sum of Euclidean distance between the point sets. We reduced calculation time using pre-defined distance map. Finally we developed an automatic coregistration program using this boundary detection and surface matching technique. We designed a new weighted normalization technique to display the coregistered PET and MR images simultaneously. Results: Using our newly developed method, robust extraction of head boundary was possible and spatial registration was successfully performed. Mean displacement error was less than 2.0 mm. In visualization of coregistered images using weighted normalization method, structures shown in MR image could be realistically represented. Conclusion: Our refined technique could practically enhance the performance of automated three dimensional coregistration.

  • PDF

Hierarchical Overlapping Clustering to Detect Complex Concepts (중복을 허용한 계층적 클러스터링에 의한 복합 개념 탐지 방법)

  • Hong, Su-Jeong;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.1
    • /
    • pp.111-125
    • /
    • 2011
  • Clustering is a process of grouping similar or relevant documents into a cluster and assigning a meaningful concept to the cluster. By this process, clustering facilitates fast and correct search for the relevant documents by narrowing down the range of searching only to the collection of documents belonging to related clusters. For effective clustering, techniques are required for identifying similar documents and grouping them into a cluster, and discovering a concept that is most relevant to the cluster. One of the problems often appearing in this context is the detection of a complex concept that overlaps with several simple concepts at the same hierarchical level. Previous clustering methods were unable to identify and represent a complex concept that belongs to several different clusters at the same level in the concept hierarchy, and also could not validate the semantic hierarchical relationship between a complex concept and each of simple concepts. In order to solve these problems, this paper proposes a new clustering method that identifies and represents complex concepts efficiently. We developed the Hierarchical Overlapping Clustering (HOC) algorithm that modified the traditional Agglomerative Hierarchical Clustering algorithm to allow overlapped clusters at the same level in the concept hierarchy. The HOC algorithm represents the clustering result not by a tree but by a lattice to detect complex concepts. We developed a system that employs the HOC algorithm to carry out the goal of complex concept detection. This system operates in three phases; 1) the preprocessing of documents, 2) the clustering using the HOC algorithm, and 3) the validation of semantic hierarchical relationships among the concepts in the lattice obtained as a result of clustering. The preprocessing phase represents the documents as x-y coordinate values in a 2-dimensional space by considering the weights of terms appearing in the documents. First, it goes through some refinement process by applying stopwords removal and stemming to extract index terms. Then, each index term is assigned a TF-IDF weight value and the x-y coordinate value for each document is determined by combining the TF-IDF values of the terms in it. The clustering phase uses the HOC algorithm in which the similarity between the documents is calculated by applying the Euclidean distance method. Initially, a cluster is generated for each document by grouping those documents that are closest to it. Then, the distance between any two clusters is measured, grouping the closest clusters as a new cluster. This process is repeated until the root cluster is generated. In the validation phase, the feature selection method is applied to validate the appropriateness of the cluster concepts built by the HOC algorithm to see if they have meaningful hierarchical relationships. Feature selection is a method of extracting key features from a document by identifying and assigning weight values to important and representative terms in the document. In order to correctly select key features, a method is needed to determine how each term contributes to the class of the document. Among several methods achieving this goal, this paper adopted the $x^2$�� statistics, which measures the dependency degree of a term t to a class c, and represents the relationship between t and c by a numerical value. To demonstrate the effectiveness of the HOC algorithm, a series of performance evaluation is carried out by using a well-known Reuter-21578 news collection. The result of performance evaluation showed that the HOC algorithm greatly contributes to detecting and producing complex concepts by generating the concept hierarchy in a lattice structure.

Video Scene Detection using Shot Clustering based on Visual Features (시각적 특징을 기반한 샷 클러스터링을 통한 비디오 씬 탐지 기법)

  • Shin, Dong-Wook;Kim, Tae-Hwan;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.47-60
    • /
    • 2012
  • Video data comes in the form of the unstructured and the complex structure. As the importance of efficient management and retrieval for video data increases, studies on the video parsing based on the visual features contained in the video contents are researched to reconstruct video data as the meaningful structure. The early studies on video parsing are focused on splitting video data into shots, but detecting the shot boundary defined with the physical boundary does not cosider the semantic association of video data. Recently, studies on structuralizing video shots having the semantic association to the video scene defined with the semantic boundary by utilizing clustering methods are actively progressed. Previous studies on detecting the video scene try to detect video scenes by utilizing clustering algorithms based on the similarity measure between video shots mainly depended on color features. However, the correct identification of a video shot or scene and the detection of the gradual transitions such as dissolve, fade and wipe are difficult because color features of video data contain a noise and are abruptly changed due to the intervention of an unexpected object. In this paper, to solve these problems, we propose the Scene Detector by using Color histogram, corner Edge and Object color histogram (SDCEO) that clusters similar shots organizing same event based on visual features including the color histogram, the corner edge and the object color histogram to detect video scenes. The SDCEO is worthy of notice in a sense that it uses the edge feature with the color feature, and as a result, it effectively detects the gradual transitions as well as the abrupt transitions. The SDCEO consists of the Shot Bound Identifier and the Video Scene Detector. The Shot Bound Identifier is comprised of the Color Histogram Analysis step and the Corner Edge Analysis step. In the Color Histogram Analysis step, SDCEO uses the color histogram feature to organizing shot boundaries. The color histogram, recording the percentage of each quantized color among all pixels in a frame, are chosen for their good performance, as also reported in other work of content-based image and video analysis. To organize shot boundaries, SDCEO joins associated sequential frames into shot boundaries by measuring the similarity of the color histogram between frames. In the Corner Edge Analysis step, SDCEO identifies the final shot boundaries by using the corner edge feature. SDCEO detect associated shot boundaries comparing the corner edge feature between the last frame of previous shot boundary and the first frame of next shot boundary. In the Key-frame Extraction step, SDCEO compares each frame with all frames and measures the similarity by using histogram euclidean distance, and then select the frame the most similar with all frames contained in same shot boundary as the key-frame. Video Scene Detector clusters associated shots organizing same event by utilizing the hierarchical agglomerative clustering method based on the visual features including the color histogram and the object color histogram. After detecting video scenes, SDCEO organizes final video scene by repetitive clustering until the simiarity distance between shot boundaries less than the threshold h. In this paper, we construct the prototype of SDCEO and experiments are carried out with the baseline data that are manually constructed, and the experimental results that the precision of shot boundary detection is 93.3% and the precision of video scene detection is 83.3% are satisfactory.