• Title/Summary/Keyword: Meta-validation

Search Result 38, Processing Time 0.024 seconds

Topic and Survey Methodological Trends in 'The Journal of Information Systems' ('정보시스템연구'의 연구주제와 서베이 방법론 동향분석)

  • Ryoo, Sung-Yul;Park, Sang-Cheol
    • The Journal of Information Systems
    • /
    • v.27 no.4
    • /
    • pp.1-33
    • /
    • 2018
  • Purpose The purpose of this study is to review topic and survey methodological trends in 'The Journal of Information Systems' in order to present the practical guidelines for the future IS research. By attempting to conduct a meta-analysis on both topic and survey methodological trends, this study could provide researchers wishing to pursue this line of work further with what can be done to improve IS disciplines. Design/methodology/approach In this study, we have reviewed 185 papers that were published in 'The Journal of Information Systems' from 2010 to 2018 and classified them based on topics studied and survey methodologies used. The classification guidelines, which was developed by Palvia et al.(2015), has been used to capture the topic trends. We have also employed Struab et al.(2004)s' guidelines for securing rigor of validation issues. By using two guidelines, this study could also present topic and rigor trends in 'The Journal of Information Systems' and compare them to those trends in International Journals. Findings Our findings have identified dominant research topics in 'The Journal of Information Systems'; 1) social media and social computing, 2) IS usage and adoption, 3) mobile computing, 4) electronic commerce/business, 5) security and privacy, 6) supply chain management, 7) innovation, 8) knowledge management, and 9) IS management and planning. This study also could offer researchers who pursue this line of work further practical guidelines on mandatory (convergent and discriminant validity, reliability, and statistical conclusion validity), highly recommended (common method bias testing), and optional validations (measurement invariance testing for subgroup analysis, bootstrapping methods for testing mediating effects).

Analysis of Occupational Injury and Feature Importance of Fall Accidents on the Construction Sites using Adaboost (에이다 부스트를 활용한 건설현장 추락재해의 강도 예측과 영향요인 분석)

  • Choi, Jaehyun;Ryu, HanGuk
    • Journal of the Architectural Institute of Korea Structure & Construction
    • /
    • v.35 no.11
    • /
    • pp.155-162
    • /
    • 2019
  • The construction industry is the highest safety accident causing industry as 28.55% portion of all industries' accidents in Korea. In particular, falling is the highest accidents type composed of 60.16% among the construction field accidents. Therefore, we analyzed the factors of major disaster affecting the fall accident and then derived feature importances by considering various variables. We used data collected from Korea Occupational Safety & Health Agency (KOSHA) for learning and predicting in the proposed model. We have an effort to predict the degree of occupational fall accidents by using the machine learning model, i.e., Adaboost, short for Adaptive Boosting. Adaboost is a machine learning meta-algorithm which can be used in conjunction with many other types of learning algorithms to improve performance. Decision trees were combined with AdaBoost in this model to predict and classify the degree of occupational fall accidents. HyOperpt was also used to optimize hyperparameters and to combine k-fold cross validation by hierarchy. We extracted and analyzed feature importances and affecting fall disaster by permutation technique. In this study, we verified the degree of fall accidents with predictive accuracy. The machine learning model was also confirmed to be applicable to the safety accident analysis in construction site. In the future, if the safety accident data is accumulated automatically in the network system using IoT(Internet of things) technology in real time in the construction site, it will be possible to analyze the factors and types of accidents according to the site conditions from the real time data.

Monitoring Ground-level SO2 Concentrations Based on a Stacking Ensemble Approach Using Satellite Data and Numerical Models (위성 자료와 수치모델 자료를 활용한 스태킹 앙상블 기반 SO2 지상농도 추정)

  • Choi, Hyunyoung;Kang, Yoojin;Im, Jungho;Shin, Minso;Park, Seohui;Kim, Sang-Min
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1053-1066
    • /
    • 2020
  • Sulfur dioxide (SO2) is primarily released through industrial, residential, and transportation activities, and creates secondary air pollutants through chemical reactions in the atmosphere. Long-term exposure to SO2 can result in a negative effect on the human body causing respiratory or cardiovascular disease, which makes the effective and continuous monitoring of SO2 crucial. In South Korea, SO2 monitoring at ground stations has been performed, but this does not provide spatially continuous information of SO2 concentrations. Thus, this research estimated spatially continuous ground-level SO2 concentrations at 1 km resolution over South Korea through the synergistic use of satellite data and numerical models. A stacking ensemble approach, fusing multiple machine learning algorithms at two levels (i.e., base and meta), was adopted for ground-level SO2 estimation using data from January 2015 to April 2019. Random forest and extreme gradient boosting were used as based models and multiple linear regression was adopted for the meta-model. The cross-validation results showed that the meta-model produced the improved performance by 25% compared to the base models, resulting in the correlation coefficient of 0.48 and root-mean-square-error of 0.0032 ppm. In addition, the temporal transferability of the approach was evaluated for one-year data which were not used in the model development. The spatial distribution of ground-level SO2 concentrations based on the proposed model agreed with the general seasonality of SO2 and the temporal patterns of emission sources.

Validation of Korean Water Quality Standards to Hot Springs for Agreement with Legionella-Incidence Risk (레지오넬라균 출현위해도에 대한 현행 온천수 수질기준의 적합성 분석)

  • Kim, Jin-Nam;Lee, Soyoung;Zo, Young-Gun
    • Microbiology and Biotechnology Letters
    • /
    • v.43 no.3
    • /
    • pp.259-266
    • /
    • 2015
  • Observed trends in climate change, globalization and an aging population have an effect on public health conditions in Korea, prompting a reevaluation of current environmental regulations. In this study, we evaluated the performance of the total coliform (TC) standard, which is the only microbiological standard in the current regulation regime for hot spring water, by estimating correlation with the presence/absence of Legionella, a non-fecal opportunistic pathogen with heat-tolerance. Microbiological data in 7 studies that surveyed Legionella in hot spring waters were subjected to meta-analyses with the odds ratio (OR) as the effect size. The presence/absence of Legionella was significantly correlated to TC levels [OR = 3.1(1.5–6.4, 95% CI), p = 0.002]. Due to there being no direct explanation as to the reason for the occurrence of TC, mesophilic fecal bacteria, being coupled with Legionella presence, the mechanism of the correlation between the two kinds of bacteria was further investigated. Legionella presence was more prevalent with a high heterotrophic plate count [HPC; 4.0(2.2–7.2); p < 0.001] and water temperature [4.3(1.4–13.6), p = 0.011] when the temperature range was <40℃. However, it was reverse-correlated with water temperature when the temperature was >50℃ [0.2(0.1–0.4), p < 0.001]. Therefore, bacterial standing crops in hot spring waters appear to be determined by water temperature in general, and this forces TC and Legionella levels be correlated. In accordance with this relationship, HPC rather than TC reflect the levels of non-fecal contamination better. Therefore, employing HPC as the sole microbiological standard, or adding HPC into the current standard of hot spring water assessment, is suggested as a proactive measure to prevent health issues arising from contamination.

A Systematic Review of Constraint Induced Movement Therapy about Upper Extremity in Stroke (뇌졸중 환자의 상지 강제유도운동치료에 관한 체계적 고찰)

  • Park, Su-Hyang;Baek, Soon-Hyung;Shin, Joong-il
    • The Journal of the Korea Contents Association
    • /
    • v.16 no.3
    • /
    • pp.149-161
    • /
    • 2016
  • The purpose of this study is provided to useful data to establish the Constraint Induced Movement Therapy(CIMT) in clinical plan to more specific for stroke patients. Also It is provided way for further study about CIMT. Methods used a systematic review. Systematic review is a research method that can be presented to the scientific evidence. Data were organized by PICO(Patient, Intervention, Comparison, Outcome). Research using the database Embase and Medline, It was searched for CIMT and Stroke. We selected for a total of 42 studies that meet the purpose of the present study. We was selected for a total of 42 studies that meet the purpose of the present study. Results was that the quality of the study is a systematic review, meta-analyzes, randomized controlled. CIMT studies was based on a high quality level of 50% of the total. The difference between the study period was 42.8%, more research was conducted prior to 2010. CIMT has been used more than mCIMT by to differ 40.5%. It is effective in over 75% of study, regardless of the CIMT intervention. In conclusion, CIMT has an effect on the upper limbs of stroke patients damaged, results will be used as a useful material to develop a CIMT in the clinical treatment plan. In future studies will need to validate studies on the effectiveness of the mCIMT, It will require a review of the effectiveness of validation studies.

Improved Estimation of Hourly Surface Ozone Concentrations using Stacking Ensemble-based Spatial Interpolation (스태킹 앙상블 모델을 이용한 시간별 지상 오존 공간내삽 정확도 향상)

  • KIM, Ye-Jin;KANG, Eun-Jin;CHO, Dong-Jin;LEE, Si-Woo;IM, Jung-Ho
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.25 no.3
    • /
    • pp.74-99
    • /
    • 2022
  • Surface ozone is produced by photochemical reactions of nitrogen oxides(NOx) and volatile organic compounds(VOCs) emitted from vehicles and industrial sites, adversely affecting vegetation and the human body. In South Korea, ozone is monitored in real-time at stations(i.e., point measurements), but it is difficult to monitor and analyze its continuous spatial distribution. In this study, surface ozone concentrations were interpolated to have a spatial resolution of 1.5km every hour using the stacking ensemble technique, followed by a 5-fold cross-validation. Base models for the stacking ensemble were cokriging, multi-linear regression(MLR), random forest(RF), and support vector regression(SVR), while MLR was used as the meta model, having all base model results as additional input variables. The results showed that the stacking ensemble model yielded the better performance than the individual base models, resulting in an averaged R of 0.76 and RMSE of 0.0065ppm during the study period of 2020. The surface ozone concentration distribution generated by the stacking ensemble model had a wider range with a spatial pattern similar with terrain and urbanization variables, compared to those by the base models. Not only should the proposed model be capable of producing the hourly spatial distribution of ozone, but it should also be highly applicable for calculating the daily maximum 8-hour ozone concentrations.

Investigating Dynamic Mutation Process of Issues Using Unstructured Text Analysis (비정형 텍스트 분석을 활용한 이슈의 동적 변이과정 고찰)

  • Lim, Myungsu;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.1
    • /
    • pp.1-18
    • /
    • 2016
  • Owing to the extensive use of Web media and the development of the IT industry, a large amount of data has been generated, shared, and stored. Nowadays, various types of unstructured data such as image, sound, video, and text are distributed through Web media. Therefore, many attempts have been made in recent years to discover new value through an analysis of these unstructured data. Among these types of unstructured data, text is recognized as the most representative method for users to express and share their opinions on the Web. In this sense, demand for obtaining new insights through text analysis is steadily increasing. Accordingly, text mining is increasingly being used for different purposes in various fields. In particular, issue tracking is being widely studied not only in the academic world but also in industries because it can be used to extract various issues from text such as news, (SocialNetworkServices) to analyze the trends of these issues. Conventionally, issue tracking is used to identify major issues sustained over a long period of time through topic modeling and to analyze the detailed distribution of documents involved in each issue. However, because conventional issue tracking assumes that the content composing each issue does not change throughout the entire tracking period, it cannot represent the dynamic mutation process of detailed issues that can be created, merged, divided, and deleted between these periods. Moreover, because only keywords that appear consistently throughout the entire period can be derived as issue keywords, concrete issue keywords such as "nuclear test" and "separated families" may be concealed by more general issue keywords such as "North Korea" in an analysis over a long period of time. This implies that many meaningful but short-lived issues cannot be discovered by conventional issue tracking. Note that detailed keywords are preferable to general keywords because the former can be clues for providing actionable strategies. To overcome these limitations, we performed an independent analysis on the documents of each detailed period. We generated an issue flow diagram based on the similarity of each issue between two consecutive periods. The issue transition pattern among categories was analyzed by using the category information of each document. In this study, we then applied the proposed methodology to a real case of 53,739 news articles. We derived an issue flow diagram from the articles. We then proposed the following useful application scenarios for the issue flow diagram presented in the experiment section. First, we can identify an issue that actively appears during a certain period and promptly disappears in the next period. Second, the preceding and following issues of a particular issue can be easily discovered from the issue flow diagram. This implies that our methodology can be used to discover the association between inter-period issues. Finally, an interesting pattern of one-way and two-way transitions was discovered by analyzing the transition patterns of issues through category analysis. Thus, we discovered that a pair of mutually similar categories induces two-way transitions. In contrast, one-way transitions can be recognized as an indicator that issues in a certain category tend to be influenced by other issues in another category. For practical application of the proposed methodology, high-quality word and stop word dictionaries need to be constructed. In addition, not only the number of documents but also additional meta-information such as the read counts, written time, and comments of documents should be analyzed. A rigorous performance evaluation or validation of the proposed methodology should be performed in future works.

A Study on Web-based Technology Valuation System (웹기반 지능형 기술가치평가 시스템에 관한 연구)

  • Sung, Tae-Eung;Jun, Seung-Pyo;Kim, Sang-Gook;Park, Hyun-Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.23-46
    • /
    • 2017
  • Although there have been cases of evaluating the value of specific companies or projects which have centralized on developed countries in North America and Europe from the early 2000s, the system and methodology for estimating the economic value of individual technologies or patents has been activated on and on. Of course, there exist several online systems that qualitatively evaluate the technology's grade or the patent rating of the technology to be evaluated, as in 'KTRS' of the KIBO and 'SMART 3.1' of the Korea Invention Promotion Association. However, a web-based technology valuation system, referred to as 'STAR-Value system' that calculates the quantitative values of the subject technology for various purposes such as business feasibility analysis, investment attraction, tax/litigation, etc., has been officially opened and recently spreading. In this study, we introduce the type of methodology and evaluation model, reference information supporting these theories, and how database associated are utilized, focusing various modules and frameworks embedded in STAR-Value system. In particular, there are six valuation methods, including the discounted cash flow method (DCF), which is a representative one based on the income approach that anticipates future economic income to be valued at present, and the relief-from-royalty method, which calculates the present value of royalties' where we consider the contribution of the subject technology towards the business value created as the royalty rate. We look at how models and related support information (technology life, corporate (business) financial information, discount rate, industrial technology factors, etc.) can be used and linked in a intelligent manner. Based on the classification of information such as International Patent Classification (IPC) or Korea Standard Industry Classification (KSIC) for technology to be evaluated, the STAR-Value system automatically returns meta data such as technology cycle time (TCT), sales growth rate and profitability data of similar company or industry sector, weighted average cost of capital (WACC), indices of industrial technology factors, etc., and apply adjustment factors to them, so that the result of technology value calculation has high reliability and objectivity. Furthermore, if the information on the potential market size of the target technology and the market share of the commercialization subject refers to data-driven information, or if the estimated value range of similar technologies by industry sector is provided from the evaluation cases which are already completed and accumulated in database, the STAR-Value is anticipated that it will enable to present highly accurate value range in real time by intelligently linking various support modules. Including the explanation of the various valuation models and relevant primary variables as presented in this paper, the STAR-Value system intends to utilize more systematically and in a data-driven way by supporting the optimal model selection guideline module, intelligent technology value range reasoning module, and similar company selection based market share prediction module, etc. In addition, the research on the development and intelligence of the web-based STAR-Value system is significant in that it widely spread the web-based system that can be used in the validation and application to practices of the theoretical feasibility of the technology valuation field, and it is expected that it could be utilized in various fields of technology commercialization.