• Title/Summary/Keyword: Distributed Database Systems

Search Result 201, Processing Time 0.026 seconds

How librarians really use the network for advanced service (정보봉사의 증진을 위한 사서들의 네트워크 이용연구)

  • 한복희
    • Journal of Korean Library and Information Science Society
    • /
    • v.23
    • /
    • pp.1-27
    • /
    • 1995
  • The purpose of this study is twofold: to investigate into general characteristics of the networks in Korea as a new information technology and to discuss general directions of development of the use of the Internet. This study is designed to achieve the purpose by gathering and analysing data related to the use of Internet of librarians those who work in public libraries and research and development libraries and university libraries. The major conclusions made in this study is summarized as follows. (1) From this survey, received detailed response from 69 librarians, the majority (42) from research and development libraries. The majority (56) were from Library and Information Science subject area, half of them (37) hold advanced degrees. (2) Majority (40) have accessed Internet for one year or less, 9(17%) respondents for two years, 17(32%) spend every day Internet related activity. (3) 44.9% of the respondents taught themselves. 28.9% learned informally from a colleague. Formal training from a single one-hour class to more structured learning was available to 30.4%. (4) The most common reason respondents use the Internet are to access remote database searching(73.9%), to communicate with colleagues and friends and electronic mail(52.2%), to transfer files and data exchange(36.2%), to know the current research front(23.2%). They search OPACs for a variety of traditional task-related reasons(59.4%) and to see what other libraries are doing with their automated systems(31.9%). (5) Respondents for the most part use the functions : WWW (68. 1%), E-Mail(59.4%), FTP(52.2%), Gopher(34.8%), Wais(7.2%). (6) Respondents mentioned the following advantages : access to remote log-in database, an excellent and swift communications vehicle, reduced telecommunication cost, saving time. (7) Respondents mentioned the following disadvantages : low speed of communication, difficult of access to the relevant information and library materials, and shortage of database be distributed within Korea.

  • PDF

Development of Integrated Retrieval System of the Biology Sequence Database Using Web Service (웹 서비스를 이용한 바이오 서열 정보 데이터베이스 및 통합 검색 시스템 개발)

  • Lee, Su-Jung;Yong, Hwan-Seung
    • The KIPS Transactions:PartD
    • /
    • v.11D no.4
    • /
    • pp.755-764
    • /
    • 2004
  • Recently, the rapid development of biotechnology brings the explosion of biological data and biological data host. Moreover, these data are highly distributed and heterogeneous, reflecting the distribution and heterogeneity of the Molecular Biology research community. As a consequence, the integration and interoperability of molecular biology databases are issue of considerable importance. But, up to now, most of the integrated systems such as link based system, data warehouse based system have many problems which are keeping the data up to date when the schema and data of the data source are changed. For this reason, the integrated system using web service technology that allow biological data to be fully exploited have been proposed. In this paper, we built the integrated system if the bio sequence information bated on the web service technology. The developed system allows users to get data with many format such as BSML, GenBank, Fasta to traverse disparate data resources. Also, it has better retrieval performance because the retrieval modules of the external database proceed in parallel.

A Study on Distribution Query Conversion Method for Real-time Integrating Retrieval based on TMDR (TMDR 기반의 실시간 통합 검색을 위한 분산질의 변환 기법에 대한 연구)

  • Hwang, Chi-Gon;Shin, Hyo-Young;Jung, Kye-Dong;Choi, Young-Keun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.14 no.7
    • /
    • pp.1701-1707
    • /
    • 2010
  • This study is intended for implementing the system environment that can help integrate and retrieve various types of data in real-time by providing semantic interoperability among distributed heterogeneous information systems. The semantic interoperability is made possible by providing a TMDR(Topicmaps Metadata Registry), a set of ontologies. TMDR, which has been made by combining MDR(MetaData Registry) and TopicMaps and storing them in the database, is able to generate distributed query and provide efficient knowledge. MDR is a metadata management technique for distributed data management. TopicMaps is an ontology representation technique that takes into consideration the hierarchy and association for accessing knowledge data. We have created TMDR, a kind of ontology, that is fit for any system and able to detect and resolve semantic conflicts on the level of data and schema. With this system we propose a query-processing technique to integrate and access heterogeneous information sources. Unlike existing retrieval methods this makes possible efficient retrieval and reasoning by providing association focusing on subjects.

Design and Implementation of an Expert Search System Using Academic Data in Big Data Processing Platforms (빅데이터 처리 플랫폼에서 학술 데이터를 사용한 전문가 검색 시스템 설계 및 구현)

  • Choi, Dojin;Kim, Minsoo;Kim, Daeyun;Lee, Seohee;Han, Jinsu;Seo, Indeok;Lim, Jongtae;Bok, Kyoungsoo;Yoo, Jaesoo
    • The Journal of the Korea Contents Association
    • /
    • v.17 no.3
    • /
    • pp.100-114
    • /
    • 2017
  • Most of the researchers establish research directions to conduct the study of new fields by getting advice from experts or through the papers of experts. The existing academic data search services provide paper information by field but do not provide experts by field. Therefore, users should decide experts by field using the searched papers by themselves. In this paper, we design and implement an expert search system by discipline through big data processing based on papers that have been published in the academic societies. The proposed system utilizes distributed big data storage systems to store and manage large papers. We also discriminate experts and analyze data related to the experts by using distributed big data processing technologies. The processed results are provided through web pages when a user searches for experts. The user can get a lot of helps for the research of a particular field since the proposed system recommends the experts of the corresponding research field.

An Agent-based Approach for Distributed Collaborative Filtering (분산 협력 필터링에 대한 에이전트 기반 접근 방법)

  • Kim, Byeong-Man;Li, Qing;Howe Adele E.;Yeo, Dong-Gyu
    • Journal of KIISE:Software and Applications
    • /
    • v.33 no.11
    • /
    • pp.953-964
    • /
    • 2006
  • Due to the usefulness of the collaborative filtering, it has been widely used in both the research and commercial field. However, there are still some challenges for it to be more efficient, especially the scalability problem, the sparsity problem and the cold start problem. In this paper. we address these problems and provide a novel distributed approach based on agents collaboration for the problems. We have tried to solve the scalability problem by making each agent save its users ratings and broadcast them to the users friends so that only friends ratings and his own ratings are kept in an agents local database. To reduce quality degradation of recommendation caused by the lack of rating data, we introduce a method using friends opinions instead of real rating data when they are not available. We also suggest a collaborative filtering algorithm based on user profile to provide new users with recommendation service. Experiments show that our suggested approach is helpful to the new user problem as well as is more scalable than traditional centralized CF filtering systems and alleviate the sparsity problem.

A Taxonomy of Workflow Architectures

  • Kim, Kwang-Hoon;Paik, Su-Ki
    • The Journal of Information Technology and Database
    • /
    • v.5 no.1
    • /
    • pp.97-108
    • /
    • 1998
  • This paper proposes a conceptual taxonomy of architectures for workflow management systems. The systematic classification work is based on a framework for workflow architectures. The framework, consisting of generic-level, conceptual-level and implementation-level architectures, provides common architectural principles for designing a workflow management system. We define the taxonomy by considering the possibilities for centralization or distribution of data, control, and execution. That is, we take into account three criteria. How are the major components of a workflow model and system, like activities, roles, actors, and workcases, concretized in workflow architecture. Which of the components is represented as software modules of the workflow architecture\ulcorner And how are they configured and operating in the architecture\ulcorner The workflow components might be embodied, as active (processes or threads) modules or as passive (data) modules, in the software architecture of a workflow management system. One or combinations of the components might become software modules in the software architecture. Finally, they might be centralized or distributed. The distribution of the components should be broken into three: Vertically, Horizontally and Fully distributed. Through the combination of these aspects, we can conceptually generate about 64 software Architectures for a workflow management system. That is, it should be possible to comprehend and characterize all kinds of software architectures for workflow management systems including the current existing systems as well as future systems. We believe that this taxonomy is a significant contribution because it adds clarity, completeness, and global perspective to workflow architectural discussions. The vocabulary suggested here includes workflow levels and aspects, allowing very different architectures to be discussed, compared, and contrasted. Added clarity is obtained because similar architectures from different vendors that used different terminology and techniques can now be seen to be identical at the higher level. Much of the complexity can be removed by thinking of workflow systems. Therefore, it is used to categorize existing workflow architectures and suggest a plethora of new workflow architectures. Finally, the taxonomy can be used for sorting out gems and stones amongst the architectures possibly generated. Thus, it might be a guideline not only for characterizing the existing workflow management systems, but also for solving the long-term and short-term architectural research issues, such as dynamic changes in workflow, transactional workflow, dynamically evolving workflow, large-scale workflow, etc., that have been proposed in the literature.

  • PDF

Picture archiving and communications systems development and performance results

  • Nam, Ji-Seung;Ralph Martinez
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1991.10b
    • /
    • pp.1796-1800
    • /
    • 1991
  • Picture Archiving and Communication Systems(PACS) provide an integration of digital imaging information in a hospital, which encompasses various imaging equipment, viewing workstations, database archive systems, and a high speed fiber optic network. One of the most important requirements for integration is the standardization of communication protocols to connect devices from different vendors. Since 1985, the ACR-NEMA standard provides a hardware interface, a set of software commands, and a consistent set of data formats for point-to-point interconnection of medical equipment. However, it has been shown to be inadequate for PACS networking environments, because of its point-to-point nature and its inflexibility to allow other services and protocols in the future. Based on previous experience of PACS developments in The University of Arizona, a new communication protocol for PACS networks has been suggested to the ACR-NEMA Working Group VI. The defined PACS protocol is intended to facilitate the development of PACS's capable of interfacing with other hospital information systems. Also, it is intended to allow the creation of diagnostic information data bases which can be interrogated by a variety of distributed devices. A particularly important goal is to support communications in a multivendor environment. The new protocol specifications are defined primarily as a combination of the International Organization for Standardization / Open Systems Interconnection (ISO/OSI) protocols and the data format portion of ACR-NEMA standard. This paper addresses the specification and implementation of the proposed PACS protocol into network node. The protocol specification, which covers Presentation, Session, Transport, and Network layers, is summarized briefly. The implementation has natural extentions to Global PACS environments. The protocol implementation is discussed based on our implementation efforts in the UNIX Operating System Environment. At the same time, results of performance evaluation are presented to demonstrate the implementation of defined protocol. The testing of performance analysis is performed on the PACS prototype node.

  • PDF

Design and Implementation of A Distributed Information Integration System based on Metadata Registry (메타데이터 레지스트리 기반의 분산 정보 통합 시스템 설계 및 구현)

  • Kim, Jong-Hwan;Park, Hea-Sook;Moon, Chang-Joo;Baik, Doo-Kwon
    • The KIPS Transactions:PartD
    • /
    • v.10D no.2
    • /
    • pp.233-246
    • /
    • 2003
  • The mediator-based system integrates heterogeneous information systems with the flexible manner. But it does not give much attention on the query optimization issues, especially for the query reusing. The other thing is that it does not use standardized metadata for schema matching. To improve this two issues, we propose mediator-based Distributed Information Integration System (DIIS) which uses query caching regarding performance and uses ISO/IEC 11179 metadata registry in terms of standardization. The DIIS is designed to provide decision-making support, which logically integrates the distributed heterogeneous business information systems based on the Web environment. We designed the system in the aspect of three-layer expression formula architecture using the layered pattern to improve the system reusability and to facilitate the system maintenance. The functionality and flow of core components of three-layer architecture are expressed in terms of process line diagrams and assembly line diagrams of Eriksson Penker Extension Model (EPEM), a methodology of an extension of UML. For the implementation, Supply Chain Management (SCM) domain is used. And we used the Web-based environment for user interface. The DIIS supports functions of query caching and query reusability through Query Function Manager (QFM) and Query Function Repository (QFR) such that it enhances the query processing speed and query reusability by caching the frequently used queries and optimizing the query cost. The DIIS solves the diverse heterogeneity problems by mapping MetaData Registry (MDR) based on ISO/IEC 11179 and Schema Repository (SCR).

A Taxonomy of Workflow Architectures

  • Kim, Kwang-Hoon;Paik, Su-Ki
    • Proceedings of the Korea Database Society Conference
    • /
    • 1998.09a
    • /
    • pp.525-543
    • /
    • 1998
  • This paper proposes a conceptual taxonomy of architectures far workflow management systems. The systematic classification work is based on a framework for workflow architectures. The framework, consisting of generic-level, conceptual-level and implementation-level architectures, provides common architectural principles for designing a workflow management system. We define the taxonomy by considering the possibilities for centralization or distribution of data, control, and execution. That is, we take into account three criteria. How are the major components of a workflow model and system, like activities, roles, actors, and workcases, concretized in workflow architecture? Which of the components is represented as software modules of the workflow architecture? And how are they configured and operating in the architecture? The workflow components might be embodied, as active (processes or threads) modules or as passive (data) modules, in the software architecture of a workflow management system. One or combinations of the components might become software modules in the software architecture. Finally, they might be centralized or distributed. The distribution of the components should be broken into three: Vertically, Horizontally and Fully distributed. Through the combination of these aspects, we can conceptually generate about 64 software Architectures for a workflow management system. That is, it should be possible to comprehend and characterize all kinds of software architectures for workflow management systems including the current existing systems as well as future systems. We believe that this taxonomy is a significant contribution because it adds clarity, completeness, and "global perspective" to workflow architectural discussions. The vocabulary suggested here includes workflow levels and aspects, allowing very different architectures to be discussed, compared, and contrasted. Added clarity is obtained because similar architectures from different vendors that used different terminology and techniques can now be seen to be identical at the higher level. Much of the complexity can be removed by thinking of workflow systems. Therefore, it is used to categorize existing workflow architectures and suggest a plethora of new workflow architectures. Finally, the taxonomy can be used for sorting out gems and stones amongst the architectures possibly generated. Thus, it might be a guideline not only for characterizing the existing workflow management systems, but also for solving the long-term and short-term architectural research issues, such as dynamic changes in workflow, transactional workflow, dynamically evolving workflow, large-scale workflow, etc., that have been proposed in the literature.

  • PDF

Efficient Publishing Spatial Information as GML for Interoperability of Heterogeneous Spatial Database Systems (이질적인 공간정보시스템의 상호 운용성을 위한 효과적인 지리데이터의 GML 사상)

  • 정원일;배해영
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.1
    • /
    • pp.12-26
    • /
    • 2004
  • In the past, geographic data is constructed and serviced through independent formats of its own according to each GIS(Geographic Information System). Recently the provision of interoperability in GIS is important to efficiently apply the various geographic data between conventional GIS's. Whereupon OGC(Open GIS Consortium) proposed GML(Geography Markup Language) to offer the interoperability between heterogeneous GISs in distributed environments. The GML is an XML encoding for the transport and storage of geographic information, including both the spatial and non-spatial properties of geographic features. Also, the GML includes Web Map Server Implementation Specification to service the GML documents. Accordingly the prototype to provide the reciprocal interchange of geographic information between conventional GIS's and GML documents is widely studied. In this paper, we propose a mapping method of geographic in formation between spatial database and GML for the prototype to support the interoperability between heterogeneous geographic information. For this method, firstly the scheme of converting geographic in Formation of the conventional spatial database into the GML document according to the GML specification is explained, and secondly the scheme to transform geographic information of GML documents to geographic data of spatial database is showed. Consequently, the proposed method is applicable to the framework for integrated geographic information services based on Web by making an offer the interoperability between already built geographic information of conventional GIS's using a mapping method of geographic information between spatial database and GML.

  • PDF