• Title/Summary/Keyword: data-based model

Search Result 21,096, Processing Time 0.049 seconds

Comparative Study of Citizen Science and Expert Based Survey Data Using the Species Distribution Model of Rana uenoi (큰산개구리(Rana uenoi ) 종분포모형을 활용한 시민과학 및 전문가 기반 조사자료의 비교연구)

  • Woncheol Lee;Jeongwoo Yoo;Paikho Rho
    • Journal of Environmental Science International
    • /
    • v.32 no.6
    • /
    • pp.429-440
    • /
    • 2023
  • Quantitative habitat model is established with species occurrence and spatial abundance data, which were usually acquired by professional field ecologists and citizen scientists. The importance of citizen science data is increasing, but the quality of these data needs to be evaluated. This study aims to identify and compare both expert-based data and citizen science data based on the performance power of quantitative models derived from both data sets. A Maximum Entropy (MaxENT) model was developed using eight environmental variables, including climate, topography, landcover and distance to forest edge. The AUC values derived from the MaxENT model were 0.842 and 0.809, respectively, indicating a high level of explanatory power. All environmental variables has similar values for both data sets, except for the distance to forest edge and rice paddy, which was relatively higher for expert-based survey data than that of the citizen science data as the distances increased. This result suggests that habitat model derived from expert-based survey data shows more ecological niche including wider ranges from forest edges and isolated habitat patches of rice paddy. This is presumably because citizen scientists focuses on direct observation methods, whereas professional field surveys investigate a wider variety of methods.

Generating Training Dataset of Machine Learning Model for Context-Awareness in a Health Status Notification Service (사용자 건강 상태알림 서비스의 상황인지를 위한 기계학습 모델의 학습 데이터 생성 방법)

  • Mun, Jong Hyeok;Choi, Jong Sun;Choi, Jae Young
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.9 no.1
    • /
    • pp.25-32
    • /
    • 2020
  • In the context-aware system, rule-based AI technology has been used in the abstraction process for getting context information. However, the rules are complicated by the diversification of user requirements for the service and also data usage is increased. Therefore, there are some technical limitations to maintain rule-based models and to process unstructured data. To overcome these limitations, many studies have applied machine learning techniques to Context-aware systems. In order to utilize this machine learning-based model in the context-aware system, a management process of periodically injecting training data is required. In the previous study on the machine learning based context awareness system, a series of management processes such as the generation and provision of learning data for operating several machine learning models were considered, but the method was limited to the applied system. In this paper, we propose a training data generating method of a machine learning model to extend the machine learning based context-aware system. The proposed method define the training data generating model that can reflect the requirements of the machine learning models and generate the training data for each machine learning model. In the experiment, the training data generating model is defined based on the training data generating schema of the cardiac status analysis model for older in health status notification service, and the training data is generated by applying the model defined in the real environment of the software. In addition, it shows the process of comparing the accuracy by learning the training data generated in the machine learning model, and applied to verify the validity of the generated learning data.

Analysis Model Evaluation based on IoT Data and Machine Learning Algorithm for Prediction of Acer Mono Sap Liquid Water

  • Lee, Han Sung;Jung, Se Hoon
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.10
    • /
    • pp.1286-1295
    • /
    • 2020
  • It has been increasingly difficult to predict the amounts of Acer mono sap to be collected due to droughts and cold waves caused by recent climate changes with few studies conducted on the prediction of its collection volume. This study thus set out to propose a Big Data prediction system based on meteorological information for the collection of Acer mono sap. The proposed system would analyze collected data and provide managers with a statistical chart of prediction values regarding climate factors to affect the amounts of Acer mono sap to be collected, thus enabling efficient work. It was designed based on Hadoop for data collection, treatment and analysis. The study also analyzed and proposed an optimal prediction model for climate conditions to influence the volume of Acer mono sap to be collected by applying a multiple regression analysis model based on Hadoop and Mahout.

AUTOMATED INTEGRATION OF CONSTRUCTION IMAGES IN MODEL BASED SYSTEMS

  • Ioannis K. Brilakis;Lucio Soibelman
    • International conference on construction engineering and project management
    • /
    • 2005.10a
    • /
    • pp.503-508
    • /
    • 2005
  • In the modern, distributed and dynamic construction environment it is important to exchange information from different sources and in different data formats in order to improve the processes supported by these systems. Previous research has demonstrated that (i) a significant percentage of construction data is stored in semi-structured or unstructured data formats (ii) locating and identifying such data that are needed for the important decision making processes is a very hard and time-consuming task. In this paper, an automated methodology for the classification and retrieval of construction images in AEC/FM model based systems will be presented. Specifically, a combination of techniques from the areas of image processing, computer vision, and content-based image retrieval have been deployed to develop a method that can retrieve related construction site image data from components of a project model.

  • PDF

A Study of Query Processing Model to applied Meta Rule in 4-Level Layer based on Hybrid Databases (하이브리드 데이터베이스 기반의 4단계 레이어 계층구조에서 메타규칙을 적용한 질의어 수행 모델에 관한 연구)

  • Oh, Ryum-Duck
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.6
    • /
    • pp.125-134
    • /
    • 2009
  • A biological data acquisition based on web has emerged as a powerful tool for allowing scientists to interactively view entries form different databases, and to navigate from one database to another molecular-biology database links. In this paper, the biological conceptual model is constructed hybrid biological data model to represent interesting entities in the data sources to applying navigation rule property for each biological data source based on four biological data integrating layers to control biological data. When some user's requests for application service are occurred, we can get the data from database and data source via web service. In this paper, we propose a query processing model and execution structure based on integrating data layers that can search information on biological data sources.

Development and Lessons Learned of Clinical Data Warehouse based on Common Data Model for Drug Surveillance (약물부작용 감시를 위한 공통데이터모델 기반 임상데이터웨어하우스 구축)

  • Mi Jung Rho
    • Korea Journal of Hospital Management
    • /
    • v.28 no.3
    • /
    • pp.1-14
    • /
    • 2023
  • Purposes: It is very important to establish a clinical data warehouse based on a common data model to offset the different data characteristics of each medical institution and for drug surveillance. This study attempted to establish a clinical data warehouse for Dankook university hospital for drug surveillance, and to derive the main items necessary for development. Methodology/Approach: This study extracted the electronic medical record data of Dankook university hospital tracked for 9 years from 2013 (2013.01.01. to 2021.12.31) to build a clinical data warehouse. The extracted data was converted into the Observational Medical Outcomes Partnership Common Data Model (Version 5.4). Data term mapping was performed using the electronic medical record data of Dankook university hospital and the standard term mapping guide. To verify the clinical data warehouse, the use of angiotensin receptor blockers and the incidence of liver toxicity were analyzed, and the results were compared with the analysis of hospital raw data. Findings: This study used a total of 670,933 data from electronic medical records for the Dankook university clinical data warehouse. Excluding the number of overlapping cases among the total number of cases, the target data was mapped into standard terms. Diagnosis (100% of total cases), drug (92.1%), and measurement (94.5%) were standardized. For treatment and surgery, the insurance EDI (electronic data interchange) code was used as it is. Extraction, conversion and loading were completed. R language-based conversion and loading software for the process was developed, and clinical data warehouse construction was completed through data verification. Practical Implications: In this study, a clinical data warehouse for Dankook university hospitals based on a common data model supporting drug surveillance research was established and verified. The results of this study provide guidelines for institutions that want to build a clinical data warehouse in the future by deriving key points necessary for building a clinical data warehouse.

  • PDF

Knowledge Recommendation Based on Dual Channel Hypergraph Convolution

  • Yue Li
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.11
    • /
    • pp.2903-2923
    • /
    • 2023
  • Knowledge recommendation is a type of recommendation system that recommends knowledge content to users in order to satisfy their needs. Although using graph neural networks to extract data features is an effective method for solving the recommendation problem, there is information loss when modeling real-world problems because an edge in a graph structure can only be associated with two nodes. Because one super-edge in the hypergraph structure can be connected with several nodes and the effectiveness of knowledge graph for knowledge expression, a dual-channel hypergraph convolutional neural network model (DCHC) based on hypergraph structure and knowledge graph is proposed. The model divides user data and knowledge data into user subhypergraph and knowledge subhypergraph, respectively, and extracts user data features by dual-channel hypergraph convolution and knowledge data features by combining with knowledge graph technology, and finally generates recommendation results based on the obtained user embedding and knowledge embedding. The performance of DCHC model is higher than the comparative model under AUC and F1 evaluation indicators, comparative experiments with the baseline also demonstrate the validity of DCHC model.

A Study on Quantitative Modeling for EPCIS Event Data (EPCIS Event 데이터 크기의 정량적 모델링에 관한 연구)

  • Lee, Chang-Ho;Jho, Yong-Chul
    • Journal of the Korea Safety Management & Science
    • /
    • v.11 no.4
    • /
    • pp.221-228
    • /
    • 2009
  • Electronic Product Code Information Services(EPCIS) is an EPCglobal standard for sharing EPC related information between trading partners. EPCIS provides a new important capability to improve efficiency, security, and visibility in the global supply chain. EPCIS data are classified into two categories, master data (static data) and event data (dynamic data). Master data are static and constant for objects, for example, the name and code of product and the manufacturer, etc. Event data refer to things that happen dynamically with the passing of time, for example, the date of manufacture, the period and the route of circulation, the date of storage in warehouse, etc. There are four kinds of event data which are Object Event data, Aggregation Event data, Quantity Event data, and Transaction Event data. This thesis we propose an event-based data model for EPC Information Service repository in RFID based integrated logistics center. This data model can reduce the data volume and handle well all kinds of entity relationships. From the point of aspect of data quantity, we propose a formula model that can explain how many EPCIS events data are created per one business activity. Using this formula model, we can estimate the size of EPCIS events data of RFID based integrated logistics center for a one day under the assumed scenario.

Data Mining Approach for Real-Time Processing of Large Data Using Case-Based Reasoning : High-Risk Group Detection Data Warehouse for Patients with High Blood Pressure (사례기반추론을 이용한 대용량 데이터의 실시간 처리 방법론 : 고혈압 고위험군 관리를 위한 자기학습 시스템 프레임워크)

  • Park, Sung-Hyuk;Yang, Kun-Woo
    • Journal of Information Technology Services
    • /
    • v.10 no.1
    • /
    • pp.135-149
    • /
    • 2011
  • In this paper, we propose the high-risk group detection model for patients with high blood pressure using case-based reasoning. The proposed model can be applied for public health maintenance organizations to effectively manage knowledge related to high blood pressure and efficiently allocate limited health care resources. Especially, the focus is on the development of the model that can handle constraints such as managing large volume of data, enabling the automatic learning to adapt to external environmental changes and operating the system on a real-time basis. Using real data collected from local public health centers, the optimal high-risk group detection model was derived incorporating optimal parameter sets. The results of the performance test for the model using test data show that the prediction accuracy of the proposed model is two times better than the natural risk of high blood pressure.

A Standard Way of Constructing a Data Warehouse based on a Neutral Model for Sharing Product Dat of Nuclear Power Plants (원자력 발전소 제품 데이터의 공유를 위한 중립 모델 기반의 데이터 웨어하우스의 구축)

  • Mun, D.H.;Cheon, S.U.;Choi, Y.J.;Han, S.H.
    • Korean Journal of Computational Design and Engineering
    • /
    • v.12 no.1
    • /
    • pp.74-85
    • /
    • 2007
  • During the lifecycle of a nuclear power plant many organizations are involved in KOREA. Korea Plant Engineering Co. (KOPEC) participates in the design stage, Korea Hydraulic and Nuclear Power (KHNP) operates and manages all nuclear power plants in KOREA, Dusan Heavy Industries manufactures the main equipment, and a construction company constructs the plant. Even though each organization has a digital data management system inside and obtains a certain level of automation, data sharing among organizations is poor. KHNP gets drawing and technical specifications from KOPEC in the form of paper. It results in manual re-work of definition and there are potential errors in the process. A data warehouse based on a neutral model has been constructed in order to make an information bridge between design and O&M phases. GPM(generic product model), a data model from Hitachi, Japan is addressed and extended in this study. GPM has a similar architecture with ISO 15926 "life cycle data for process plant". The extension is oriented to nuclear power plants. This paper introduces some of implementation results: 1) 2D piping and instrument diagram (P&ID) and 3D CAD model exchanges and their visualization; 2) Interface between GPM-based data warehouse and KHNP ERP system.