• Title/Summary/Keyword: 검색질의어

Search Result 468, Processing Time 0.031 seconds

A fishway information system of the east coast at Kangwondo (강원도 동해안 지역의 어도 정보시스템 구축)

  • 이창수;최상수;박상덕;박기영
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 1997.10a
    • /
    • pp.21-24
    • /
    • 1997
  • 강원도 동해안 지역 하천은 유로연장과 유역면적이 작고 하천 상류부는 급경사이며 하류부는 완경사의 특징을 보여주고 있다. 이 지역 대부분의 하천이 바다로 직접 유입되고 있어 바다와 하천을 옮겨 다니는 회유성 어류가 서식하고 있으며 이들은 소하성, 강하성 및 양측성 어류로 구분된다. 이러한 회유성 어류의 이동은 동해안 지역 하천 하류부 완경사지에서 농업용수를 취수하기 위한 보의 설치로 인하여 어려워졌으며, 이러한 문제에 대처하기 위하여 수산자원 보호령(대통령령 제 5027호) 제 21조에서 어도설치를 규정하고 있다. 강원도 동해안 지역 하천에 설치되어 있는 어도의 실태를 보면 총 19개 하천에 있는 83개소의 농업용 취수보에 약 130개의 어도가 있다. 기존에 설치되어 있는 전어도에 대한 현장조사결과 어도의 기능에 장애를 일으킬 만한 원인은 어도시설의 구조적인 문제와 관리문제로 대별할 수 있다. 본 연구는 어도의 기능이 원활하게 유지되도록 어도시설물에 대한 각종 자료를 체계적으로 집적한 어도정보시스템을 구축하여 어도시설에 대한 현황파악을 용이하게 하고, 지방자치단체에서 개별적으로 수행하던 어도관리를 통합관리할수 있게 하며, 어도시설 표준모형 설계의 기초자료 및 하천 정보, 어류생태정보와 연결된 정보시스템으로의 확장을 가능하게 한다. 본 어도 정보시스템은 기본적으로 인터넷상에서 서비스가 이루어진다. 사용자는 웹 브라우저를 이용하여 어도 정보시스템이 탑재된 웹 서버에 접속하여 서비스를 요청한다. 사용자의 요청을 받은 웹 서버는 게시판과 DBMS를 제어하여 그 결과를 사용자에게 건네준다. 사용된 기술은 ISAPI 규격의 확장인 IDC, HTX이다. 관리자와 사용자의 입력, 수정, 검색, 삭제등에 관한 읽고 쓰기 권한을 구분한다.이 경과함에 따라 소멸하며 약산성에서 높은 생성능을 보임을 알 수 있었다, 강별로는 낙동강에서 THMs파 HANs의 생성능이 비교적 높았고, 영산강에서는 HAAs의 생성능이 높았다. 각 원수의 특성에 따른 생성능을 파악함으로써 생성능 저감방안을 마련하며, 소독 부산물질의 인체노출평가에 따른 위해성평가를 통해 관리기준을 설정해야 할 것이다.는 것으로 나타났다. 본 연구는 한국 문화의 특징이라 할 수 있는 교수-학생간의 인간적인 관계가 중시되는 교육환경하에서, 정보화 사회의 한 수업형태인 재택수업을 외국(특히 미국)의 방법을 그대로 따라 실시할 경우, 많은 부작용이 있을 수 있다는 것을 암시하며, 따라서 우리 교육문화에 맞는 재택수업 형태의 개발이 시급함을 제시한다고 하겠다.column density of HCaN is (1-3):n1014cm-2. Column density at distant position from MD5 is larger than that in the (:entral region. We have deduced that this hot-core has a mass of 10sR1 which i:s about an order of magnitude larger those obtained by previous studies.previous studies.업순서들의 상관관계를 고려하여 보다 개선된 해를 구하기 위한 연구가 요구된다. 또한, 준비작업비용을 발생시키는 작업장의 작업순서결정에 대해서도 연구를 행하여, 보완작업비용과 준비비용을 고려한 GMMAL 작업순서문제를 해결하기 위한 연구가 수행되어야 할 것이다.로 이루어 져야 할 것이다.태를 보다 효율적으로 증진시킬 수 있는 대안이 마련되어져야 한다고 사료된다.$\ulco

  • PDF

A Study on the Integration of Information Extraction Technology for Detecting Scientific Core Entities based on Large Resources (대용량 자원 기반 과학기술 핵심개체 탐지를 위한 정보추출기술 통합에 관한 연구)

  • Choi, Yun-Soo;Cheong, Chang-Hoo;Choi, Sung-Pil;You, Beom-Jong;Kim, Jae-Hoon
    • Journal of Information Management
    • /
    • v.40 no.4
    • /
    • pp.1-22
    • /
    • 2009
  • Large-scaled information extraction plays an important role in advanced information retrieval as well as question answering and summarization. Information extraction can be defined as a process of converting unstructured documents into formalized, tabular information, which consists of named-entity recognition, terminology extraction, coreference resolution and relation extraction. Since all the elementary technologies have been studied independently so far, it is not trivial to integrate all the necessary processes of information extraction due to the diversity of their input/output formation approaches and operating environments. As a result, it is difficult to handle scientific documents to extract both named-entities and technical terms at once. In this study, we define scientific as a set of 10 types of named entities and technical terminologies in a biomedical domain. in order to automatically extract these entities from scientific documents at once, we develop a framework for scientific core entity extraction which embraces all the pivotal language processors, named-entity recognizer, co-reference resolver and terminology extractor. Each module of the integrated system has been evaluated with various corpus as well as KEEC 2009. The system will be utilized for various information service areas such as information retrieval, question-answering(Q&A), document indexing, dictionary construction, and so on.

Metadata Management System for XML-based Digital Broadcasting (XML 기반 디지털 방송용 메타데이타 관리시스템)

  • Park Jong-Hyun;Kim Byung-Kyu;Lee Young-Hee;Lee Min-Woo;Jung Min-Ok;Kang Ji-Hoon
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.11 no.4
    • /
    • pp.334-348
    • /
    • 2005
  • The goal of next generation digital broadcasting is offering the interaction among consumers and providers as well as variety services. One of the important factors for this new broadcasting environment keeps the interoperability among providers and consumers since the environment is distributed. Therefore a standard metadata for digital broadcasting is required and TV-Anytime metadata is one of the metadata standards for digital broadcasting. The terminal nodes of TV-Anytime metadata are defined by using MPEG-7 metadata. MPEG-7 metadata is standard metadata for describing multimedia content. Therefore, if we use the MPEG-7 metadata for describing broadcasting content can offer multimedia search services like content-based search by the extension of metadata. The efficient management system for these metadata is important for offering the services with high Duality on real broadcasting environment TV-Anytime metadata and MPEG-7 metadata are technically defined using a single XML schema, so its instances are XML data. Currently, a lot of systemsfor managing XML data are proposed in many researchers and we can expect to adapt these systems for managing broadcasting metadata. But the methods used in these systems are not specific methods for managing broadcasting metadata because of methods for general-purpose. In this paper, we find the properties of broadcasting metadata and develop an efficient metadata management system that is based on the found properties. Since our systemis implemented on real broadcasting environment, we expect that the system is most efficient and suitable. Also our system is interoperable since we use XQuery as query language for querying broadcasting metadata.

A Study on Processing XML Documents (XML 문서 처리에 관한 연구)

  • Kim, Tae Gwon
    • Journal of KIISE
    • /
    • v.43 no.4
    • /
    • pp.489-496
    • /
    • 2016
  • XML can effectively express structured or semi-structured data as well as relational databases. XQuery is a query language for retrieving information for such an XML document. In this paper, an XQuery composer is designed and implemented, with an API provided for XQuery processors, and a proper processor is registered. This composer shows query results immediately processed by the processor. As this composer contains a parser for XQuery, it can compose XQuery effectively using a diverse dialog box designed for XQuery grammar. A dialog box is affiliated with a clause region, which is a region that algebra operates from the parsing tree. It can compose path expressions for an XML document easily as it shows an element tree from DTD graphically. Path expressions are composed automatically by marking elements in the structural hierarchy and by specifying the predicate of an element partially.

Improving Bidirectional LSTM-CRF model Of Sequence Tagging by using Ontology knowledge based feature (온톨로지 지식 기반 특성치를 활용한 Bidirectional LSTM-CRF 모델의 시퀀스 태깅 성능 향상에 관한 연구)

  • Jin, Seunghee;Jang, Heewon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.253-266
    • /
    • 2018
  • This paper proposes a methodology applying sequence tagging methodology to improve the performance of NER(Named Entity Recognition) used in QA system. In order to retrieve the correct answers stored in the database, it is necessary to switch the user's query into a language of the database such as SQL(Structured Query Language). Then, the computer can recognize the language of the user. This is the process of identifying the class or data name contained in the database. The method of retrieving the words contained in the query in the existing database and recognizing the object does not identify the homophone and the word phrases because it does not consider the context of the user's query. If there are multiple search results, all of them are returned as a result, so there can be many interpretations on the query and the time complexity for the calculation becomes large. To overcome these, this study aims to solve this problem by reflecting the contextual meaning of the query using Bidirectional LSTM-CRF. Also we tried to solve the disadvantages of the neural network model which can't identify the untrained words by using ontology knowledge based feature. Experiments were conducted on the ontology knowledge base of music domain and the performance was evaluated. In order to accurately evaluate the performance of the L-Bidirectional LSTM-CRF proposed in this study, we experimented with converting the words included in the learned query into untrained words in order to test whether the words were included in the database but correctly identified the untrained words. As a result, it was possible to recognize objects considering the context and can recognize the untrained words without re-training the L-Bidirectional LSTM-CRF mode, and it is confirmed that the performance of the object recognition as a whole is improved.

Restoring Omitted Sentence Constituents in Encyclopedia Documents Using Structural SVM (Structural SVM을 이용한 백과사전 문서 내 생략 문장성분 복원)

  • Hwang, Min-Kook;Kim, Youngtae;Ra, Dongyul;Lim, Soojong;Kim, Hyunki
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.2
    • /
    • pp.131-150
    • /
    • 2015
  • Omission of noun phrases for obligatory cases is a common phenomenon in sentences of Korean and Japanese, which is not observed in English. When an argument of a predicate can be filled with a noun phrase co-referential with the title, the argument is more easily omitted in Encyclopedia texts. The omitted noun phrase is called a zero anaphor or zero pronoun. Encyclopedias like Wikipedia are major source for information extraction by intelligent application systems such as information retrieval and question answering systems. However, omission of noun phrases makes the quality of information extraction poor. This paper deals with the problem of developing a system that can restore omitted noun phrases in encyclopedia documents. The problem that our system deals with is almost similar to zero anaphora resolution which is one of the important problems in natural language processing. A noun phrase existing in the text that can be used for restoration is called an antecedent. An antecedent must be co-referential with the zero anaphor. While the candidates for the antecedent are only noun phrases in the same text in case of zero anaphora resolution, the title is also a candidate in our problem. In our system, the first stage is in charge of detecting the zero anaphor. In the second stage, antecedent search is carried out by considering the candidates. If antecedent search fails, an attempt made, in the third stage, to use the title as the antecedent. The main characteristic of our system is to make use of a structural SVM for finding the antecedent. The noun phrases in the text that appear before the position of zero anaphor comprise the search space. The main technique used in the methods proposed in previous research works is to perform binary classification for all the noun phrases in the search space. The noun phrase classified to be an antecedent with highest confidence is selected as the antecedent. However, we propose in this paper that antecedent search is viewed as the problem of assigning the antecedent indicator labels to a sequence of noun phrases. In other words, sequence labeling is employed in antecedent search in the text. We are the first to suggest this idea. To perform sequence labeling, we suggest to use a structural SVM which receives a sequence of noun phrases as input and returns the sequence of labels as output. An output label takes one of two values: one indicating that the corresponding noun phrase is the antecedent and the other indicating that it is not. The structural SVM we used is based on the modified Pegasos algorithm which exploits a subgradient descent methodology used for optimization problems. To train and test our system we selected a set of Wikipedia texts and constructed the annotated corpus in which gold-standard answers are provided such as zero anaphors and their possible antecedents. Training examples are prepared using the annotated corpus and used to train the SVMs and test the system. For zero anaphor detection, sentences are parsed by a syntactic analyzer and subject or object cases omitted are identified. Thus performance of our system is dependent on that of the syntactic analyzer, which is a limitation of our system. When an antecedent is not found in the text, our system tries to use the title to restore the zero anaphor. This is based on binary classification using the regular SVM. The experiment showed that our system's performance is F1 = 68.58%. This means that state-of-the-art system can be developed with our technique. It is expected that future work that enables the system to utilize semantic information can lead to a significant performance improvement.

Design and Performance Analysis of Storage Schema for Efficient Storing of GML Documents (효율적인 GML 문서 저장을 위한 저장 스키마의 설계 및 성능평가)

  • Chang, Jae-Woo;Wang, Tae-Woong;Lee, Hyun-Jo
    • Journal of Korea Spatial Information System Society
    • /
    • v.9 no.2
    • /
    • pp.35-53
    • /
    • 2007
  • GML is a mark-up language which is suggested by OGC(Open GIS Consortium) for use of encoding standard concerning with storing and transferring Geographic information. For general spatial network database, researches supporting GML documents can be classified into three categories : parsing, storing, and retrieval of GML documents. Among them, the 'Storing of GML document' is essential for efficient GML document retrieval. However, There is little research on schema for storing GML documents. In addition, the existing schema for storing XML documents can't be used for GML documents due to geographic information. Therefore, In this paper, we propose efficient schema for storing GML documents In addition, we do performance evaluation of the GML schema.

  • PDF

Development of the Tool for Software Re-engineering and Maintenance (소프트웨어 재공학과 유지보수 지원을 위한 툴의 개발)

  • Kim, Haeng-Gon;Hwang, Seon-Myeong
    • The Transactions of the Korea Information Processing Society
    • /
    • v.1 no.3
    • /
    • pp.299-310
    • /
    • 1994
  • Re-engineering tools can substantially increase software maintenance productivity and the quality of maintenance work. Re-engineering usually involves changing the form(e.g.changining objects names and definitions, restructuring process logic) of a program. In this paper, we describe the design and implementation of InMaC++ that is a software tool oriented towards maintenance of C++ object oriented programs. With InMaC++, programms can be displayed and edited in two forms : as the code and as the diagram InMaC++ also contains transformations in both directions, i,e. from code to diagram and from diagram to skeletons of code. Hence, it is suitable for re-engineering and maintenance of existing code. Specially designed browsers implement the graphical interface. InMaC++ contains a database that is based on a simple but effective data model of InMaC++ programs. The model contains only four object classes and three relations, which makes the tool small, and easy to implement and use. A simple query language allows browsing through the database.

  • PDF

Development of Database System for T-50 Flutter/Vibroacoustic Flight Test Data (T-50 플러터/진동소음 비행시험 데이터베이스 시스템 개발)

  • Gwak, Dong-Il;Baek, Seung-Gil;Park, Geum-Dang;Kim, Yeong-Ik
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.34 no.2
    • /
    • pp.82-89
    • /
    • 2006
  • The flutter/vibroacoustic characteristics can be acquired by conducting flight tests with various conditions for a long period of test. Accordingly it is indispensable to build a specially designed database system to efficiently accumulate the enormous data obtained from flight tests. Hence, T-50 Flight Test Database System(FTDS) based on MS-Access is developed to handle the flutter/vibroacoustic environment data obtained from flight tests. The developed system is structured with the items related to aircraft flight test, the tables composed of the relevant items and a relational database logically connecting the tables. The T-50 FTDS is implemented with data searching GUI(Graphic User Interface) programed with Visual Basic and Structured Query Language to make intuitive searches over the stored data. The developed system has been used for accumulating the flutter/vibroacoustic data and verifying vibroacoustic Specifications.

Crawling algorithm design and experiment for automatic deep web document collection (심층 웹 문서 자동 수집을 위한 크롤링 알고리즘 설계 및 실험)

  • Yun-Jeong, Kang;Min-Hye, Lee;Dong-Hyun, Won
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.27 no.1
    • /
    • pp.1-7
    • /
    • 2023
  • Deep web collection means entering a query in a search form and collecting response results. It is estimated that the information possessed by the deep web has about 450 to 550 times more information than the statically constructed surface web. The static method does not show the changed information until the web page is refreshed, but the dynamic web page method updates the necessary information in real time and provides real-time information without reloading the web page, but crawler has difficulty accessing the updated information. Therefore, there is a need for a way to automatically collect information on these deep webs using a crawler. Therefore, this paper proposes a method of utilizing scripts as general links, and for this purpose, an algorithm that can utilize client scripts like regular URLs is proposed and experimented. The proposed algorithm focused on collecting web information by menu navigation and script execution instead of the usual method of entering data into search forms.