• Title/Summary/Keyword: complete linkage algorithm

Search Result 7, Processing Time 0.023 seconds

A Study on Maintenance Bundle Alternatives of BTL Project for Educational Facilities Using Complete Linkage Algorithm (컴플리트 링키지 알고리즘을 이용한 교육시설물 BTL사업 유지관리번들 구성방안에 관한 연구)

  • Cho, Chang-Yeon;Son, Jae-Ho
    • Journal of the Korean Institute of Educational Facilities
    • /
    • v.15 no.3
    • /
    • pp.4-16
    • /
    • 2008
  • BTL(Build-Transfer-Lease) Project for Education Facilities is contracted as a package which consists of several education facilities and its maintenance period is 20 years. Thus, total cost variation largely depends on the accuracy of the maintenance cost forecasting in the early stage in the life cycle of the BTL Projects. This research develops a method using complete linkage algorithm and branch & bound algorithm to help in finding optimal bundling combination. The result of this research suggests more reasonable and effective forecasting method for the maintenance bundle in BTL projects.

A Study on Development of Maintenance Cost Estimation System for BTL Project of Education Facilities Using Optimization Methodology (최적화기법을 활용한 교육시설물 BTL 사업 운영관리비용 비용예측 시스템 개발 기초연구)

  • Cho, Chang-Yeon;Son, Jae-Ho;Kim, Jea-On
    • Korean Journal of Construction Engineering and Management
    • /
    • v.10 no.1
    • /
    • pp.45-57
    • /
    • 2009
  • BTL (Build-Transfer-Lease) Project for Education Facilities are contracted as a package which consists of several education facilities. The general maintenance period of BTL project for education facilities is 20 years. Thus, total cost variation largely depends on the accuracy of the maintenance cost forecasting in the early stage in the life cycle of the BTL Project. This research develops a cost forecasting system using complete linkage algorithm and branch & bound algorithm to help in finding optimal bundling combination. This system helps owner's decision-making to estimate the total project cost with various constraints changing. The result of this research suggests more reasonable and effective forecasting model for the maintenance facilities package in the BTL project.

Development of a Clustering Model for Automatic Knowledge Classification (지식 분류의 자동화를 위한 클러스터링 모형 연구)

  • 정영미;이재윤
    • Journal of the Korean Society for information Management
    • /
    • v.18 no.2
    • /
    • pp.203-230
    • /
    • 2001
  • The purpose of this study is to develop a document clustering model for automatic classification of knowledge. Two test collections of newspaper article texts and journal article abstracts are built for the clustering experiment. Various feature reduction criteria as well as term weighting methods are applied to the term sets of the test collections, and cosine and Jaccard coefficients are used as similarity measures. The performances of complete linkage and K-means clustering algorithms are compared using different feature selection methods and various term weights. It was found that complete linkage clustering outperforms K-means algorithm and feature reduction up to almost 10% of the total feature sets does not lower the performance of document clustering to any significant extent.

  • PDF

Algorithm Design to Judge Fake News based on Bigdata and Artificial Intelligence

  • Kang, Jangmook;Lee, Sangwon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.11 no.2
    • /
    • pp.50-58
    • /
    • 2019
  • The clear and specific objective of this study is to design a false news discriminator algorithm for news articles transmitted on a text-based basis and an architecture that builds it into a system (H/W configuration with Hadoop-based in-memory technology, Deep Learning S/W design for bigdata and SNS linkage). Based on learning data on actual news, the government will submit advanced "fake news" test data as a result and complete theoretical research based on it. The need for research proposed by this study is social cost paid by rumors (including malicious comments) and rumors (written false news) due to the flood of fake news, false reports, rumors and stabbings, among other social challenges. In addition, fake news can distort normal communication channels, undermine human mutual trust, and reduce social capital at the same time. The final purpose of the study is to upgrade the study to a topic that is difficult to distinguish between false and exaggerated, fake and hypocrisy, sincere and false, fraud and error, truth and false.

A Comparison of Cluster Analyses and Clustering of Sensory Data on Hanwoo Bulls (군집분석 비교 및 한우 관능평가데이터 군집화)

  • Kim, Jae-Hee;Ko, Yoon-Sil
    • The Korean Journal of Applied Statistics
    • /
    • v.22 no.4
    • /
    • pp.745-758
    • /
    • 2009
  • Cluster analysis is the automated search for groups of related observations in a data set. To group the observations into clusters many techniques has been proposed, and a variety measures aimed at validating the results of a cluster analysis have been suggested. In this paper, we compare complete linkage, Ward's method, K-means and model-based clustering and compute validity measures such as connectivity, Dunn Index and silhouette with simulated data from multivariate distributions. We also select a clustering algorithm and determine the number of clusters of Korean consumers based on Korean consumers' palatability scores for Hanwoo bull in BBQ cooking method.

Construction and Validation of Hospital-Based Cancer Registry Using Various Health Records to Detect Patients with Newly Diagnosed Cancer: Experience at Asan Medical Center (의무기록의 다각적 활용을 통한 충실도 높은 병원 암등록 체계의 구축: 서울아산병원의 경험)

  • Kim, Hwa-Jung;Cho, Jin-Hee;Lyu, Yong-Man;Lee, Sun-Hye;Hwang, Kyeong-Ha;Lee, Moo-Song
    • Journal of Preventive Medicine and Public Health
    • /
    • v.43 no.3
    • /
    • pp.257-264
    • /
    • 2010
  • Objectives: An accurate estimation of cancer patients is the basis of epidemiological studies and health services. However in Korea, cancer patients visiting out-patient clinics are usually ruled out of such studies and so these studies are suspected of underestimating the cancer patient population. The purpose of this study is to construct a more complete, hospital-based cancer patient registry using multiple sources of medical information. Methods: We constructed a cancer patient detection algorithm using records from various sources that were obtained from both the in-patients and out-patients seen at Asan Medical Center (AMC) for any reason. The medical data from the potentially incident cancer patients was reviewed four months after first being detected by the algorithm to determine whether these patients actually did or did not have cancer. Results: Besides the traditional practice of reviewing the charts of in-patients upon their discharge, five more sources of information were added for this algorithm, i.e., pathology reports, the national severe disease registry, the reason for treatment, prescriptions of chemotherapeutic agents and radiation therapy reports. The constructed algorithm was observed to have a PPV of 87.04%. Compared to the results of traditional practice, 36.8% of registry failures were avoided using the AMC algorithm. Conclusions: To minimize loss in the cancer registry, various data sources should be utilized, and the AMC algorithm can be a successful model for this. Further research will be required in order to apply novel and innovative technology to the electronic medical records system in order to generate new signals from data that has not been previously used.

Imputation Accuracy from Low to Moderate Density Single Nucleotide Polymorphism Chips in a Thai Multibreed Dairy Cattle Population

  • Jattawa, Danai;Elzo, Mauricio A.;Koonawootrittriron, Skorn;Suwanasopee, Thanathip
    • Asian-Australasian Journal of Animal Sciences
    • /
    • v.29 no.4
    • /
    • pp.464-470
    • /
    • 2016
  • The objective of this study was to investigate the accuracy of imputation from low density (LDC) to moderate density SNP chips (MDC) in a Thai Holstein-Other multibreed dairy cattle population. Dairy cattle with complete pedigree information (n = 1,244) from 145 dairy farms were genotyped with GeneSeek GGP20K (n = 570), GGP26K (n = 540) and GGP80K (n = 134) chips. After checking for single nucleotide polymorphism (SNP) quality, 17,779 SNP markers in common between the GGP20K, GGP26K, and GGP80K were used to represent MDC. Animals were divided into two groups, a reference group (n = 912) and a test group (n = 332). The SNP markers chosen for the test group were those located in positions corresponding to GeneSeek GGP9K (n = 7,652). The LDC to MDC genotype imputation was carried out using three different software packages, namely Beagle 3.3 (population-based algorithm), FImpute 2.2 (combined family- and population-based algorithms) and Findhap 4 (combined family- and population-based algorithms). Imputation accuracies within and across chromosomes were calculated as ratios of correctly imputed SNP markers to overall imputed SNP markers. Imputation accuracy for the three software packages ranged from 76.79% to 93.94%. FImpute had higher imputation accuracy (93.94%) than Findhap (84.64%) and Beagle (76.79%). Imputation accuracies were similar and consistent across chromosomes for FImpute, but not for Findhap and Beagle. Most chromosomes that showed either high (73%) or low (80%) imputation accuracies were the same chromosomes that had above and below average linkage disequilibrium (LD; defined here as the correlation between pairs of adjacent SNP within chromosomes less than or equal to 1 Mb apart). Results indicated that FImpute was more suitable than Findhap and Beagle for genotype imputation in this Thai multibreed population. Perhaps additional increments in imputation accuracy could be achieved by increasing the completeness of pedigree information.