• Title/Summary/Keyword: Similarity Matching

Search Result 412, Processing Time 0.026 seconds

A Method to Solve the Entity Linking Ambiguity and NIL Entity Recognition for efficient Entity Linking based on Wikipedia (위키피디아 기반의 효과적인 개체 링킹을 위한 NIL 개체 인식과 개체 연결 중의성 해소 방법)

  • Lee, Hokyung;An, Jaehyun;Yoon, Jeongmin;Bae, Kyoungman;Ko, Youngjoong
    • Journal of KIISE
    • /
    • v.44 no.8
    • /
    • pp.813-821
    • /
    • 2017
  • Entity Linking find the meaning of an entity mention, which indicate the entity using different expressions, in a user's query by linking the entity mention and the entity in the knowledge base. This task has four challenges, including the difficult knowledge base construction problem, multiple presentation of the entity mention, ambiguity of entity linking, and NIL entity recognition. In this paper, we first construct the entity name dictionary based on Wikipedia to build a knowledge base and solve the multiple presentation problem. We then propose various methods for NIL entity recognition and solve the ambiguity of entity linking by training the support vector machine based on several features, including the similarity of the context, semantic relevance, clue word score, named entity type similarity of the mansion, entity name matching score, and object popularity score. We sequentially use the proposed two methods based on the constructed knowledge base, to obtain the good performance in the entity linking. In the result of the experiment, our system achieved 83.66% and 90.81% F1 score, which is the performance of the NIL entity recognition to solve the ambiguity of the entity linking.

Variation Analysis of Long-term in vitro Cultured Cymbidium goeringii Lindley and Cymbidium kanran Makino (장기간 기내 배양한 춘란(Cymbidium goeringii Lindley) 및 한란(Cymbidium kanran Makino)의 변이 비교)

  • Ryu, Jai-Hyunk;Lee, Hyo-Yeon;Bae, Chang-Hyu
    • Korean Journal of Plant Resources
    • /
    • v.24 no.2
    • /
    • pp.139-149
    • /
    • 2011
  • RAPD (random amplified polymorphic DNA) analysis was examined to detect variation of in vitro cultured 30 rhizomes of Cymbidium goeringii Lindley and Cymbidium kanran Makino, with long-term (8 years) subculture, respectively. Out of 151 DNA bands detected, the 40 were polymorphic with a polymorphic rate 26.4% in the C. goeringii. Out of 155 DNA bands detected, the 56 were polymorphic with a polymorphic rate 36.1% in the C. kanran. Genetic similarity matrix (GSM) shows from 0.825 to 1.00 with an average of 0.944 in the rhizomes of C. goeringii and 0.812 to 1.00 with an average of 0.913 in the C. kanran. According to the clustering analysis, C. goeringii was divided into 1 group and 2 independent individuals and its structure of clustering was simple than that of C. kanran. The higher polymorphism and the decreased GSM were showed in the long-term in vitro cultured C. goeringii and C. kanran supplemented with growth regulators. The results provide as fundamental data to develop a new materials for plant breeding and resources plant.

Generation and Detection of Cranial Landmark

  • Heo, Suwoong;Kang, Jiwoo;Kim, Yong Oock;Lee, Sanghoon
    • Journal of International Society for Simulation Surgery
    • /
    • v.2 no.1
    • /
    • pp.26-32
    • /
    • 2015
  • Purpose When a surgeon examines the morphology of skull of patient, locations of craniometric landmarks of 3D computed tomography(CT) volume are one of the most important information for surgical purpose. The locations of craniometric landmarks can be found manually by surgeon from the 3D rendered volume or 2D sagittal, axial, and coronal slices which are taken by CT. Since there are many landmarks on the skull, finding these manually is time-consuming, exhaustive, and occasionally inexact. These inefficiencies raise a demand for a automatic localization technique for craniometric landmark points. So in this paper, we propose a novel method through which we can automatically find these landmark points, which are useful for surgical purpose. Materials and Methods At first, we align the experimental data (CT volumes) using Frankfurt Horizontal Plane (FHP) and Mid Sagittal Plane(MSP) which are defined by 3 and 2 cranial landmark points each. The target landmark of our experiment is the anterior nasal spine. Prior to constructing a statistical cubic model which would be used for detecting the location of the landmark from a given CT volume, reference points for the anterior nasal spine were manually chosen by a surgeon from several CT volume sets. The statistical cubic model is constructed by calculating weighted intensity means of these CT sets around the reference points. By finding the location where similarity function (squared difference function) has the minimal value with this model, the location of the landmark can be found from any given CT volume. Results In this paper, we used 5 CT volumes to construct the statistical cubic model. The 20 CT volumes including the volumes, which were used to construct the model, were used for testing. The range of age of subjects is up to 2 years (24 months) old. The found points of each data are almost close to the reference point which were manually chosen by surgeon. Also it has been seen that the similarity function always has the global minimum at the detection point. Conclusion Through the experiment, we have seen the proposed method shows the outstanding performance in searching the landmark point. This algorithm would make surgeons efficiently work with morphological informations of skull. We also expect the potential of our algorithm for searching the anatomic landmarks not only cranial landmarks.

Study of a Brain Tumor and Blood Vessel Detection System Using Multiple Fluorescence Imaging by a Surgical Microscope (수술현미경에서의 다중형광영상을 이용한 뇌종양과 혈관영상 검출 시스템 연구)

  • Lee, Hyun Min;Kim, Hong Rae;Yoon, Woong Bae;Kim, Young Jae;Kim, Kwang Gi;Kim, Seok Ki;Yoo, Heon;Lee, Seung Hoon;Shin, Min Sun;Kwon, Ki Chul
    • Korean Journal of Optics and Photonics
    • /
    • v.26 no.1
    • /
    • pp.23-29
    • /
    • 2015
  • In this paper, we propose a microscope system for detecting both a tumor and blood vessels in brain tumor surgery as fluorescence images by using multiple light sources and a beam-splitter module. The proposed method displays fluorescent images of the tumor and blood vessels on the same display device and also provides accurate information about them to the operator. To acquire a fluorescence image, we utilized 5-ALA (5-aminolevulinic acid) for the tumor and ICG (Indocyanine green) for blood vessels, and we used a beam-splitter module combined with a microscope for simultaneous detection of both. The beam-splitter module showed the best performance at 600 nm for 5-ALA and above 800 nm for ICG. The beam-splitter is flexible to enable diverse objective setups and designed to mount a filter easily, so beam-splitter and filter can be changed as needed, and other fluorescent dyes besides 5-ALA and ICG are available. The fluorescent images of the tumor and the blood vessels can be displayed on the same monitor through the beam-splitter module with a CCD camera. For ICG, a CCD that can detect the near-infrared region is needed. This system provides the acquired fluorescent image to an operator in real time, matching it to the original image through a similarity transform.

Feasibility Study on FSIM Index to Evaluate SAR Image Co-registration Accuracy (SAR 영상 정합 정확도 평가를 위한 FSIM 인자 활용 가능성)

  • Kim, Sang-Wan;Lee, Dongjun
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.5_1
    • /
    • pp.847-859
    • /
    • 2021
  • Recently, as the number of high-resolution satellite SAR images increases, the demand for precise matching of SAR imagesin change detection and image fusion is consistently increasing. RMSE (Root Mean Square Error) values using GCPs (Ground Control Points) selected by analysts have been widely used for quantitative evaluation of image registration results, while it is difficult to find an approach for automatically measuring the registration accuracy. In this study, a feasibility analysis was conducted on using the FSIM (Feature Similarity) index as a measure to evaluate the registration accuracy. TerraSAR-X (TSX) staring spotlight data collected from various incidence angles and orbit directions were used for the analysis. FSIM was almost independent on the spatial resolution of the SAR image. Using a single SAR image, the FSIM with respect to registration errors was analyzed, then use it to compare with the value estimated from TSX data with different imaging geometry. FSIM index slightly decreased due to the differencesin imaging geometry such as different look angles, different orbit tracks. As the result of analyzing the FSIM value by land cover type, the change in the FSIM index according to the co-registration error was most evident in the urban area. Therefore, the FSIM index calculated in the urban was mostsuitable for determining the accuracy of image registration. It islikely that the FSIM index has sufficient potential to be used as an index for the co-registration accuracy of SAR image.

A Combined Hough Transform based Edge Detection and Region Growing Method for Region Extraction (영역 추출을 위한 Hough 변환 기반 에지 검출과 영역 확장을 통합한 방법)

  • N.T.B., Nguyen;Kim, Yong-Kwon;Chung, Chin-Wan;Lee, Seok-Lyong;Kim, Deok-Hwan
    • Journal of KIISE:Databases
    • /
    • v.36 no.4
    • /
    • pp.263-279
    • /
    • 2009
  • Shape features in a content-based image retrieval (CBIR) system are divided into two classes: contour-based and region-based. Contour-based shape features are simple but they are not as efficient as region-based shape features. Most systems using the region-based shape feature have to extract the region firs t. The prior works on region-based systems still have shortcomings. They are complex to implement, particularly with respect to region extraction, and do not sufficiently use the spatial relationship between regions in the distance model In this paper, a region extraction method that is the combination of an edge-based method and a region growing method is proposed to accurately extract regions inside an object. Edges inside an object are accurately detected based on the Canny edge detector and the Hough transform. And the modified Integrated Region Matching (IRM) scheme which includes the adjacency relationship of regions is also proposed. It is used to compute the distance between images for the similarity search using shape features. The experimental results show the effectiveness of our region extraction method as well as the modified IRM. In comparison with other works, it is shown that the new region extraction method outperforms others.

An Efficient Frequent Melody Indexing Method to Improve Performance of Query-By-Humming System (허밍 질의 처리 시스템의 성능 향상을 위한 효율적인 빈번 멜로디 인덱싱 방법)

  • You, Jin-Hee;Park, Sang-Hyun
    • Journal of KIISE:Databases
    • /
    • v.34 no.4
    • /
    • pp.283-303
    • /
    • 2007
  • Recently, the study of efficient way to store and retrieve enormous music data is becoming the one of important issues in the multimedia database. Most general method of MIR (Music Information Retrieval) includes a text-based approach using text information to search a desired music. However, if users did not remember the keyword about the music, it can not give them correct answers. Moreover, since these types of systems are implemented only for exact matching between the query and music data, it can not mine any information on similar music data. Thus, these systems are inappropriate to achieve similarity matching of music data. In order to solve the problem, we propose an Efficient Query-By-Humming System (EQBHS) with a content-based indexing method that efficiently retrieve and store music when a user inquires with his incorrect humming. For the purpose of accelerating query processing in EQBHS, we design indices for significant melodies, which are 1) frequent melodies occurring many times in a single music, on the assumption that users are to hum what they can easily remember and 2) melodies partitioned by rests. In addition, we propose an error tolerated mapping method from a note to a character to make searching efficient, and the frequent melody extraction algorithm. We verified the assumption for frequent melodies by making up questions and compared the performance of the proposed EQBHS with N-gram by executing various experiments with a number of music data.

The Recognition of Occluded 2-D Objects Using the String Matching and Hash Retrieval Algorithm (스트링 매칭과 해시 검색을 이용한 겹쳐진 이차원 물체의 인식)

  • Kim, Kwan-Dong;Lee, Ji-Yong;Lee, Byeong-Gon;Ahn, Jae-Hyeong
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.7
    • /
    • pp.1923-1932
    • /
    • 1998
  • This paper deals with a 2-D objects recognition algorithm. And in this paper, we present an algorithm which can reduce the computation time in model retrieval by means of hashing technique instead of using the binary~tree method. In this paper, we treat an object boundary as a string of structural units and use an attributed string matching algorithm to compute similarity measure between two strings. We select from the privileged strings a privileged string wIth mmimal eccentricity. This privileged string is treated as the reference string. And thell we wllstructed hash table using the distance between privileged string and the reference string as a key value. Once the database of all model strings is built, the recognition proceeds by segmenting the scene into a polygonal approximation. The distance between privileged string extracted from the scene and the reference string is used for model hypothesis rerieval from the table. As a result of the computer simulation, the proposed method can recognize objects only computing, the distance 2-3tiems, while previous method should compute the distance 8-10 times for model retrieval.

  • PDF

Detection of M:N corresponding class group pairs between two spatial datasets with agglomerative hierarchical clustering (응집 계층 군집화 기법을 이용한 이종 공간정보의 M:N 대응 클래스 군집 쌍 탐색)

  • Huh, Yong;Kim, Jung-Ok;Yu, Ki-Yun
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.30 no.2
    • /
    • pp.125-134
    • /
    • 2012
  • In this paper, we propose a method to analyze M:N corresponding relations in semantic matching, especially focusing on feature class matching. Similarities between any class pairs are measured by spatial objects which coexist in the class pairs, and corresponding classes are obtained by clustering with these pairwise similarities. We applied a graph embedding method, which constructs a global configuration of each class in a low-dimensional Euclidean space while preserving the above pairwise similarities, so that the distances between the embedded classes are proportional to the overall degree of similarity on the edge paths in the graph. Thus, the clustering problem could be solved by employing a general clustering algorithm with the embedded coordinates. We applied the proposed method to polygon object layers in a topographic map and land parcel categories in a cadastral map of Suwon area and evaluated the results. F-measures of the detected class pairs were analyzed to validate the results. And some class pairs which would not detected by analysis on nominal class names were detected by the proposed method.

An Effcient Two-Level Hybrid Signature File Method for Large Text Databases (대용량 텍스트 데이터베이스를 위한 효율적인 2단계 합성 요약 화일 방법)

  • Yoo, Jae-Soo;Gang, Hyeong-Il
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.4
    • /
    • pp.923-932
    • /
    • 1997
  • In this paper, we propose a two-level hybrid signature file method(THM) to dffciently deal with large txt databases that use a term discrimination concept.In addition, we apply Yoo's clustering scheme to the two-level hybeid signature file method. The clustering schme groups similar signatures together according to the similarity of the highly discriminatiory tems so that we may achive better performance on retrival. The space-time ana-lyhtical model of the proposed two-level hybrid method is provided. Based on the analytical model and experiments, we compare it with the exsting methods, i.e. the bit-sliced method(BM), the-level method(TM), and the hybrid method(HM). As a result, we show that THM achives the best retrival performance in a large database with 100,000 records when the mumber fo matching records is less than 160.

  • PDF