• Title/Summary/Keyword: information retrieval.

Search Result 3,667, Processing Time 0.034 seconds

A Korean Community-based Question Answering System Using Multiple Machine Learning Methods (다중 기계학습 방법을 이용한 한국어 커뮤니티 기반 질의-응답 시스템)

  • Kwon, Sunjae;Kim, Juae;Kang, Sangwoo;Seo, Jungyun
    • Journal of KIISE
    • /
    • v.43 no.10
    • /
    • pp.1085-1093
    • /
    • 2016
  • Community-based Question Answering system is a system which provides answers for each question from the documents uploaded on web communities. In order to enhance the capacity of question analysis, former methods have developed specific rules suitable for a target region or have applied machine learning to partial processes. However, these methods incur an excessive cost for expanding fields or lead to cases in which system is overfitted for a specific field. This paper proposes a multiple machine learning method which automates the overall process by adapting appropriate machine learning in each procedure for efficient processing of community-based Question Answering system. This system can be divided into question analysis part and answer selection part. The question analysis part consists of the question focus extractor, which analyzes the focused phrases in questions and uses conditional random fields, and the question type classifier, which classifies topics of questions and uses support vector machine. In the answer selection part, the we trains weights that are used by the similarity estimation models through an artificial neural network. Also these are a number of cases in which the results of morphological analysis are not reliable for the data uploaded on web communities. Therefore, we suggest a method that minimizes the impact of morphological analysis by using character features in the stage of question analysis. The proposed system outperforms the former system by showing a Mean Average Precision criteria of 0.765 and R-Precision criteria of 0.872.

Improvement of Cloud-data Filtering Method Using Spectrum of AERI (AERI 스펙트럼 분석을 통한 구름에 영향을 받은 스펙트럼 자료 제거 방법 개선)

  • Cho, Joon-Sik;Goo, Tae-Young;Shin, Jinho
    • Korean Journal of Remote Sensing
    • /
    • v.31 no.2
    • /
    • pp.137-148
    • /
    • 2015
  • The National Institute of Meteorological Research (NIMR) has operated the Fourier Transform InfraRed (FTIR) spectrometer which is the Atmospheric Emitted Radiance Interferometer (AERI) in Anmyeon island, Korea since June 2010. The ground-based AERI with similar hyper-spectral infrared sensor to satellite could be an alternative way to validate satellite-based remote sensing. In this regard, the NIMR has focused on the improvement of retrieval quality from the AERI, particularly cloud-data filtering method. The AERI spectrum which is measured on a typical clear day is selected reference spectrum and we used region of atmospheric window. We performed test of threshold in order to select valid threshold. We retrieved methane using new method which is used reference spectrum, and the other method which is used KLAPS cloud cover information, each retrieved methane was compared with that of ground-based in-situ measurements. The quality of AERI methane retrievals of new method was significantly more improved than method of used KLAPS. In addition, the comparison of vertical total column of methane from AERI and GOSAT shows good result.

Text Filtering using Iterative Boosting Algorithms (반복적 부스팅 학습을 이용한 문서 여과)

  • Hahn, Sang-Youn;Zang, Byoung-Tak
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.4
    • /
    • pp.270-277
    • /
    • 2002
  • Text filtering is a task of deciding whether a document has relevance to a specified topic. As Internet and Web becomes wide-spread and the number of documents delivered by e-mail explosively grows the importance of text filtering increases as well. The aim of this paper is to improve the accuracy of text filtering systems by using machine learning techniques. We apply AdaBoost algorithms to the filtering task. An AdaBoost algorithm generates and combines a series of simple hypotheses. Each of the hypotheses decides the relevance of a document to a topic on the basis of whether or not the document includes a certain word. We begin with an existing AdaBoost algorithm which uses weak hypotheses with their output of 1 or -1. Then we extend the algorithm to use weak hypotheses with real-valued outputs which was proposed recently to improve error reduction rates and final filtering performance. Next, we attempt to achieve further improvement in the AdaBoost's performance by first setting weights randomly according to the continuous Poisson distribution, executing AdaBoost, repeating these steps several times, and then combining all the hypotheses learned. This has the effect of mitigating the ovefitting problem which may occur when learning from a small number of data. Experiments have been performed on the real document collections used in TREC-8, a well-established text retrieval contest. This dataset includes Financial Times articles from 1992 to 1994. The experimental results show that AdaBoost with real-valued hypotheses outperforms AdaBoost with binary-valued hypotheses, and that AdaBoost iterated with random weights further improves filtering accuracy. Comparison results of all the participants of the TREC-8 filtering task are also provided.

The effect of Meister high school students' career maturity with respect to the impact on school maladjustment (마이스터고등학교 학생들의 진로성숙도가 학교 부적응에 미치는 영향)

  • Yoo, Jae-Man;Lee, Byung-Wook
    • 대한공업교육학회지
    • /
    • v.41 no.2
    • /
    • pp.1-23
    • /
    • 2016
  • This study was conducted to analyze the effect Meister high school students' career maturity with respect to the impact on school maladjustment. Also, this study clarify the relationship. This study purpose is to permanently provide Meister as the basis for the vocational education sector career education needed to faithfully serve as a special purpose high schools. Tools used for the survey is maladaptive measurement tools developed by Leegyumi (2004) and Career maturity measurement tools developed at Korea Research Institute for Vocational Education and Training (2012). Using these tools, a reliability test was conducted. Meister students' career maturity was conducted correlation analysis and multiple regression analysis to analyze the impact of school maladjustment. Independent variables are consisted of career maturity and independence, attitude toward the job, planning, self-understanding, rational decision-making, information retrieval, knowledge of the desired job, career exploration and ready for action. Meister high school student's career maturity according to the students' background variables are little girls was higher than boys, but it was not statistically significant. T-test was conducted to ascertain the career maturity and school maladjustment differences of adaptation groups and maladaptive group in meister school students in background variables. A career maturity and school maladjustment between adaptive and maladaptive population groups showed a statistically significant difference in background variables.

Optimizing Similarity Threshold and Coverage of CBR (사례기반추론의 유사 임계치 및 커버리지 최적화)

  • Ahn, Hyunchul
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.8
    • /
    • pp.535-542
    • /
    • 2013
  • Since case-based reasoning(CBR) has many advantages, it has been used for supporting decision making in various areas including medical checkup, production planning, customer classification, and so on. However, there are several factors to be set by heuristics when designing effective CBR systems. Among these factors, this study addresses the issue of selecting appropriate neighbors in case retrieval step. As the criterion for selecting appropriate neighbors, conventional studies have used the preset number of neighbors to combine(i.e. k of k-nearest neighbor), or the relative portion of the maximum similarity. However, this study proposes to use the absolute similarity threshold varying from 0 to 1, as the criterion for selecting appropriate neighbors to combine. In this case, too small similarity threshold value may make the model rarely produce the solution. To avoid this, we propose to adopt the coverage, which implies the ratio of the cases in which solutions are produced over the total number of the training cases, and to set it as the constraint when optimizing the similarity threshold. To validate the usefulness of the proposed model, we applied it to a real-world target marketing case of an online shopping mall in Korea. As a result, we found that the proposed model might significantly improve the performance of CBR.

Grieving among Adolescent Survivors of Childhood Cancer: A Situational Analysis (청소년 소아암 생존자의 슬픔: 상황분석)

  • Jin, Juhye
    • Child Health Nursing Research
    • /
    • v.20 no.1
    • /
    • pp.49-57
    • /
    • 2014
  • Purpose: The purpose of this qualitative study was to explore how adolescent survivors of childhood cancer grieve the death of cancer peers. Methods: Data were obtained from Korean adolescents with cancer between the ages of 13 and 18 (N=12) through semi-structured interviews (face-to-face, telephone, and Internet chatting), observations of the social dynamics of participants in self-help groups, and retrieval of personal Web journals. Based on the grounded theory methodology, data collection and analysis were conducted simultaneously, and constant comparative methods were used. Clarke's situational analysis was adopted, and this paper focused on presenting "how to" and "what we can learn" from this analytic strategy. Results: Mapping examples were visualized using of three modes of maps. Adolescent cancer survivors coped with reminders of the "darkness" that ultimately featured their overall grief. Additionally, adolescents' encounters and avoidance of grief were triggered by introspection and interactions with family and friends. Conclusion: Situational analysis provided an efficient way to analyze the experiences of adolescent survivors of childhood cancer by systematizing possible information within the relational social contexts of the research phenomenon.

A Reduction Method of Over-Segmented Regions at Image Segmentation based on Homogeneity Threshold (동질성 문턱 값 기반 영상분할에서 과분할 영역 축소 방법)

  • Han, Gi-Tae
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.1
    • /
    • pp.55-68
    • /
    • 2012
  • In this paper, we propose a novel method to solve the problem of excessive segmentation out of the method of segmenting regions from an image using Homogeneity Threshold($H_T$). The algorithm of the previous image segmentation based on $H_T$ was carried out region growth by using only the center pixel of selected window. Therefore it was caused resulting in excessive segmented regions. However, before carrying region growth, the proposed method first of all finds out whether the selected window is homogeneity or not. Subsequently, if the selected window is homogeneity it carries out region growth using the total pixels of selected window. But if the selected window is not homogeneity, it carries out region growth using only the center pixel of selected window. So, the method can reduce remarkably the number of excessive segmented regions of image segmentation based on $H_T$. In order to show the validity of the proposed method, we carried out multiple experiments to compare the proposed method with previous method in same environment and conditions. As the results, the proposed method can reduce the number of segmented regions above 40% and doesn't make any difference in the quality of visual image when we compare with previous method. Especially, when we compare the image united with regions of descending order by size of segmented regions in experimentation with the previous method, even though the united image has regions more than 1,000, we can't recognize what the image means. However, in the proposed method, even though image is united by segmented regions less than 10, we can recognize what the image is. For these reason, we expect that the proposed method will be utilized in various fields, such as the extraction of objects, the retrieval of informations from the image, research for anatomy, biology, image visualization, and animation and so on.

SOSiM: Shape-based Object Similarity Matching using Shape Feature Descriptors (SOSiM: 형태 특징 기술자를 사용한 형태 기반 객체 유사성 매칭)

  • Noh, Chung-Ho;Lee, Seok-Lyong;Chung, Chin-Wan;Kim, Sang-Hee;Kim, Deok-Hwan
    • Journal of KIISE:Databases
    • /
    • v.36 no.2
    • /
    • pp.73-83
    • /
    • 2009
  • In this paper we propose an object similarity matching method based on shape characteristics of an object in an image. The proposed method extracts edge points from edges of objects and generates a log polar histogram with respect to each edge point to represent the relative placement of extracted points. It performs the matching in such a way that it compares polar histograms of two edge points sequentially along with edges of objects, and uses a well-known k-NN(nearest neighbor) approach to retrieve similar objects from a database. To verify the proposed method, we've compared it to an existing Shape-Context method. Experimental results reveal that our method is more accurate in object matching than the existing method, showing that when k=5, the precision of our method is 0.75-0.90 while that of the existing one is 0.37, and when k=10, the precision of our method is 0.61-0.80 while that of the existing one is 0.31. In the experiment of rotational transformation, our method is also more robust compared to the existing one, showing that the precision of our method is 0.69 while that of the existing one is 0.30.

Mining Maximal Frequent Contiguous Sequences in Biological Data Sequences (생물학적 데이터 서열들에서 빈번한 최대길이 연속 서열 마이닝)

  • Kang, Tae-Ho;Yoo, Jae-Soo
    • The KIPS Transactions:PartD
    • /
    • v.15D no.2
    • /
    • pp.155-162
    • /
    • 2008
  • Biological sequences such as DNA sequences and amino acid sequences typically contain a large number of items. They have contiguous sequences that ordinarily consist of hundreds of frequent items. In biological sequences analysis(BSA), a frequent contiguous sequence search is one of the most important operations. Many studies have been done for mining sequential patterns efficiently. Most of the existing methods for mining sequential patterns are based on the Apriori algorithm. In particular, the prefixSpan algorithm is one of the most efficient sequential pattern mining schemes based on the Apriori algorithm. However, since the algorithm expands the sequential patterns from frequent patterns with length-1, it is not suitable for biological dataset with long frequent contiguous sequences. In recent years, the MacosVSpan algorithm was proposed based on the idea of the prefixSpan algorithm to significantly reduce its recursive process. However, the algorithm is still inefficient for mining frequent contiguous sequences from long biological data sequences. In this paper, we propose an efficient method to mine maximal frequent contiguous sequences in large biological data sequences by constructing the spanning tree with the fixed length. To verify the superiority of the proposed method, we perform experiments in various environments. As the result, the experiments show that the proposed method is much more efficient than MacosVSpan in terms of retrieval performance.

H*-tree/H*-cubing-cubing: Improved Data Cube Structure and Cubing Method for OLAP on Data Stream (H*-tree/H*-cubing: 데이터 스트림의 OLAP를 위한 향상된 데이터 큐브 구조 및 큐빙 기법)

  • Chen, Xiangrui;Li, Yan;Lee, Dong-Wook;Kim, Gyoung-Bae;Bae, Hae-Young
    • The KIPS Transactions:PartD
    • /
    • v.16D no.4
    • /
    • pp.475-486
    • /
    • 2009
  • Data cube plays an important role in multi-dimensional, multi-level data analysis. Meeting on-line analysis requirements of data stream, several cube structures have been proposed for OLAP on data stream, such as stream cube, flowcube, S-cube. Since it is costly to construct data cube and execute ad-hoc OLAP queries, more research works should be done considering efficient data structure, query method and algorithms. Stream cube uses H-cubing to compute selected cuboids and store the computed cells in an H-tree, which form the cuboids along popular-path. However, the H-tree layoutis disorderly and H-cubing method relies too much on popular path.In this paper, first, we propose $H^*$-tree, an improved data structure, which makes the retrieval operation in tree structure more efficient. Second, we propose an improved cubing method, $H^*$-cubing, with respect to computing the cuboids that cannot be retrieved along popular-path when an ad-hoc OLAP query is executed. $H^*$-tree construction and $H^*$-cubing algorithms are given. Performance study turns out that during the construction step, $H^*$-tree outperforms H-tree with a more desirable trade-off between time and memory usage, and $H^*$-cubing is better adapted to ad-hoc OLAP querieswith respect to the factors such as time and memory space.