• Title/Summary/Keyword: NLP

Search Result 351, Processing Time 0.023 seconds

NIF Application for Korean Natural Language Processing (한국어 자연언어처리의 NIF 적용에 관한 연구)

  • Seo, Jiwoo;Won, Yousung;Kim, Jeongwook;Hahm, YoungGyun;Choi, Key-Sun
    • Annual Conference on Human and Language Technology
    • /
    • 2014.10a
    • /
    • pp.167-172
    • /
    • 2014
  • 본 논문에서는 한국어 자연언어처리 결과물들을 통일된 형식으로 표준화하기 위해서 NIF를 적용한 내용을 다룬다. 한국어 자연언어처리에 NIF 온톨로지를 적용한 이유와 적용과정에서 야기된 문제점들을 논의한다. 한국어 NLP2RDF 구축과정에서 한국어 자연언어처리에 필요한 새로운 클래스와 프로퍼티들을 추가로 정의하여 NIF 온톨로지를 변형 적용하였다.

  • PDF

Summarization and Evaluation; Where are we today?!

  • Shamsfard, Mehrnoush;Saffarian, Amir;Ghodratnama, Samaneh
    • Proceedings of the Korean Society for Language and Information Conference
    • /
    • 2007.11a
    • /
    • pp.422-429
    • /
    • 2007
  • The rapid growth of the online information services causes the problem of information explosion. Automatic text summarization techniques are essential for dealing with this problem. There are different approaches to text summarization and different systems have used one or a combination of them. Considering the wide variety of summarization techniques there should be an evaluation mechanism to assess the process of summarization. The evaluation of automatic summarization is important and challenging, since in general it is difficult to agree on an ideal summary of a text. Currently evaluating summaries is a laborious task that could not be done simply by human so automatic evaluation techniques are appearing to help this matter. In this paper, we will take a look at summarization approaches and examine summarizers' general architecture. The importance of evaluation methods is discussed and the need to find better automatic systems to evaluate summaries is studied.

  • PDF

Bio-NER using LSTM-CRF (LSTM-CRF를 이용한 생명과학분야 개체명 인식)

  • Choi, Kyoungho;Hwang, Hyunsun;Lee, Changki
    • Annual Conference on Human and Language Technology
    • /
    • 2015.10a
    • /
    • pp.85-89
    • /
    • 2015
  • 본 논문에서는 시퀀스 레이블링 문제에 적합하다고 알려진 Long Short Term Memory Recurrent Neural Network에 아웃풋간의 의존관계를 추가한 LSTM-CRF(Conditional Random Field)를 이용하여 생명과학분야 개체명 인식 시스템을 구축하였다. 학습 및 평가를 위해 BioNLP 2011-st REL data를 개체명 인식 실험에 사용하였으며, 실험결과 LSTM-CRF를 사용한 시스템은 81.83의 F1-score를 기록해, 기존의 시스템인 "BANNER"의 F1-score 81.96과 비슷한 성능을 보였다.

  • PDF

Neural Model for Named Entity Recognition Considering Aligned Representation

  • Sun, Hongyang;Kim, Taewhan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2018.10a
    • /
    • pp.613-616
    • /
    • 2018
  • Sequence tagging is an important task in Natural Language Processing (NLP), in which the Named Entity Recognition (NER) is the key issue. So far the most widely adopted model for NER in NLP is that of combining the neural network of bidirectional long short-term memory (BiLSTM) and the statistical sequence prediction method of Conditional Random Field (CRF). In this work, we improve the prediction accuracy of the BiLSTM by supporting an aligned word representation mechanism. We have performed experiments on multilingual (English, Spanish and Dutch) datasets and confirmed that our proposed model outperformed the existing state-of-the-art models.

Development and Evaluation of a Korean Treebank and its Application to NLP

  • Han, Chung-Hye;Han, Na-Rae;Ko, Eon-Suk;Martha Palmer
    • Language and Information
    • /
    • v.6 no.1
    • /
    • pp.123-138
    • /
    • 2002
  • This paper discusses issues in building a 54-thousand-word Korean Treebank using a phrase structure annotation, along with developing annotation guidelines based on the morpho-syntactic phenomena represented in the corpus. Various methods that were employed for quality control are presented. The evaluation on the quality of the Treebank and some of the NLP applications under development using the Treebank are also pre-sented.

  • PDF

Towards cross-platform interoperability for machine-assisted text annotation

  • de Castilho, Richard Eckart;Ide, Nancy;Kim, Jin-Dong;Klie, Jan-Christoph;Suderman, Keith
    • Genomics & Informatics
    • /
    • v.17 no.2
    • /
    • pp.19.1-19.10
    • /
    • 2019
  • In this paper, we investigate cross-platform interoperability for natural language processing (NLP) and, in particular, annotation of textual resources, with an eye toward identifying the design elements of annotation models and processes that are particularly problematic for, or amenable to, enabling seamless communication across different platforms. The study is conducted in the context of a specific annotation methodology, namely machine-assisted interactive annotation (also known as human-in-the-loop annotation). This methodology requires the ability to freely combine resources from different document repositories, access a wide array of NLP tools that automatically annotate corpora for various linguistic phenomena, and use a sophisticated annotation editor that enables interactive manual annotation coupled with on-the-fly machine learning. We consider three independently developed platforms, each of which utilizes a different model for representing annotations over text, and each of which performs a different role in the process.

Intelligent Wordcloud Using Text Mining (텍스트 마이닝을 이용한 지능적 워드클라우드)

  • Kim, Yeongchang;Ji, Sangsu;Park, Dongseo;Lee, Choong Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.325-326
    • /
    • 2019
  • This paper proposes an intelligent word cloud by improving the existing method of representing word cloud by examining the frequency of nouns with text mining technique. In this paper, we propose a method to visually show word clouds focused on other parts, such as verbs, by effectively adding newly-coined words and the like to a dictionary that extracts noun words in text mining. In the experiment, the KoNLP package was used for extracting the frequency of existing nouns, and 80 new words that were not supported were added manually by examining frequency.

  • PDF

Explaining the Translation Error Factors of Machine Translation Services Using Self-Attention Visualization (Self-Attention 시각화를 사용한 기계번역 서비스의 번역 오류 요인 설명)

  • Zhang, Chenglong;Ahn, Hyunchul
    • Journal of Information Technology Services
    • /
    • v.21 no.2
    • /
    • pp.85-95
    • /
    • 2022
  • This study analyzed the translation error factors of machine translation services such as Naver Papago and Google Translate through Self-Attention path visualization. Self-Attention is a key method of the Transformer and BERT NLP models and recently widely used in machine translation. We propose a method to explain translation error factors of machine translation algorithms by comparison the Self-Attention paths between ST(source text) and ST'(transformed ST) of which meaning is not changed, but the translation output is more accurate. Through this method, it is possible to gain explainability to analyze a machine translation algorithm's inside process, which is invisible like a black box. In our experiment, it was possible to explore the factors that caused translation errors by analyzing the difference in key word's attention path. The study used the XLM-RoBERTa multilingual NLP model provided by exBERT for Self-Attention visualization, and it was applied to two examples of Korean-Chinese and Korean-English translations.

Generative Linguistic Steganography: A Comprehensive Review

  • Xiang, Lingyun;Wang, Rong;Yang, Zhongliang;Liu, Yuling
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.3
    • /
    • pp.986-1005
    • /
    • 2022
  • Text steganography is one of the most imminent and promising research interests in the information security field. With the unprecedented success of the neural network and natural language processing (NLP), the last years have seen a surge of research on generative linguistic steganography (GLS). This paper provides a thorough and comprehensive review to summarize the existing key contributions, and creates a novel taxonomy for GLS according to NLP techniques and steganographic encoding algorithm, then summarizes the characteristics of generative linguistic steganographic methods properly to analyze the relationship and difference between each type of them. Meanwhile, this paper also comprehensively introduces and analyzes several evaluation metrics to evaluate the performance of GLS from diverse perspective. Finally, this paper concludes the future research work, which is more conducive to the follow-up research and innovation of researchers.

Fake News Checking Tool Based on Siamese Neural Networks and NLP (NLP와 Siamese Neural Networks를 이용한 뉴스 사실 확인 인공지능 연구)

  • Vadim, Saprunov;Kang, Sung-Won;Rhee, Kyung-hyune
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.05a
    • /
    • pp.627-630
    • /
    • 2022
  • Over the past few years, fake news has become one of the most significant problems. Since it is impossible to prevent people from spreading misinformation, people should analyze the news themselves. However, this process takes some time and effort, so the routine part of this analysis should be automated. There are many different approaches to this problem, but they only analyze the text and messages, ignoring the images. The fake news problem should be solved using a complex analysis tool to reach better performance. In this paper, we propose the approach of training an Artificial Intelligence using an unsupervised learning algorithm, combined with online data parsing tools, providing independence from subjective data set. Therefore it will be more difficult to spread fake news since people could quickly check if the news or article is trustworthy.