• Title/Summary/Keyword: Count Data

Search Result 1,124, Processing Time 0.022 seconds

Analysis of Traffic Accident by Circular Intersection Type in Korea Using Count Data Model (가산자료 모형을 이용한 국내 원형교차로 유형별 교통사고 분석)

  • Kim, Tae Yang;Lee, Min Yeong;Park, Byung Ho
    • Journal of the Korean Society of Safety
    • /
    • v.32 no.5
    • /
    • pp.129-134
    • /
    • 2017
  • This study aims to develop the traffic accident models by circular intersection type using count data model. The number of accident, the number of fatal and injured persons(FSI), and EPDO are calculated from the traffic accident data of TAAS. The circular intersection accident models are developed through Poisson and negative binomial regression analysis. The main results of this study are as follows. First, the null hypotheses that there are differences in the number of traffic accidents, FSI and EPDO by type of circular intersections are rejected. Second, the scale of intersection(median, large), number of approach road, mean width and length of exit road, area of the circulating roadway and central island are selected as factors influencing the number of traffic accidents, FSI and EPDO in rotary. Third, the scale of intersection(median), guide signs(limited speed, direction, roundabout), number of approach road, entry angle, area of the intersection and central island are adopted as factors influencing the number of traffic accidents, FSI and EPDO in roundabout. Finally, transferring from rotary to roundabout could be expected to make the accident decrease.

Analysis of Failutr Count Data Based on NHPP Models (NHPP모형에 기초한 고장 수 자료의 분석)

  • Kim, Seong-Hui;Jeong, Hyang-Suk;Kim, Yeong-Sun;Park, Jung-Yang
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.2
    • /
    • pp.395-400
    • /
    • 1997
  • An important quality characteristic of a software reliability.Software reliablilty growh models prvied the tools to evluate and moniter the reliabolty growth behavior of the sofwate during the testing phase Therefore failure data collected during the testing phase should be continmuosly analyzed on the basis of some selected software reliability growth models.For the cases where nonhomogeneous Poisson proxess models are the candiate models,we suggest Poisson regression model, which expresses the relationship between the expeted and actual failures counts in disjonint time intervals,for analyzing the failure count data.The weighted lest squares method is then used to-estimate the paramethers in the parameters in the model:The resulting estimators are equivalent to the maximum likelihood estimators. The method is illustrated by analyzing the failutr count data gathered from a large- scale switchong system.

  • PDF

Comparison between Word Embedding Techniques in Traditional Korean Medicine for Data Analysis: Implementation of a Natural Language Processing Method (한의학 고문헌 데이터 분석을 위한 단어 임베딩 기법 비교: 자연어처리 방법을 적용하여)

  • Oh, Junho
    • Journal of Korean Medical classics
    • /
    • v.32 no.1
    • /
    • pp.61-74
    • /
    • 2019
  • Objectives : The purpose of this study is to help select an appropriate word embedding method when analyzing East Asian traditional medicine texts as data. Methods : Based on prescription data that imply traditional methods in traditional East Asian medicine, we have examined 4 count-based word embedding and 2 prediction-based word embedding methods. In order to intuitively compare these word embedding methods, we proposed a "prescription generating game" and compared its results with those from the application of the 6 methods. Results : When the adjacent vectors are extracted, the count-based word embedding method derives the main herbs that are frequently used in conjunction with each other. On the other hand, in the prediction-based word embedding method, the synonyms of the herbs were derived. Conclusions : Counting based word embedding methods seems to be more effective than prediction-based word embedding methods in analyzing the use of domesticated herbs. Among count-based word embedding methods, the TF-vector method tends to exaggerate the frequency effect, and hence the TF-IDF vector or co-word vector may be a more reasonable choice. Also, the t-score vector may be recommended in search for unusual information that could not be found in frequency. On the other hand, prediction-based embedding seems to be effective when deriving the bases of similar meanings in context.

The Gamma-Ray Detection Circuit design of RI Use Instrument for Hand Carry (휴대용 RI 이용 계기의 감마선 검출 회로설계)

  • Seong, Nak-Jin;Kim, Sang-Jin;Kim, Ki-Joon
    • Proceedings of the Korean Institute of Electrical and Electronic Material Engineers Conference
    • /
    • 2003.05b
    • /
    • pp.154-158
    • /
    • 2003
  • In this study, to measure the density of compaction, it is designed to use the 5 gamma-ray detectors. The developed instrument consists of measuring circuits for gamma-rays and thermal neutrons, a high voltage supply unit, stable circuit unit, count circuit unit and a microprocessor. To read count pulse from gamma-ray detectors are very accurate and it can be count to data calibration excluded count of ripple.

  • PDF

Highly Cited Articles in Periacetabular Osteotomy Research

  • Colin K. Cantrell;Cody J. Goedderz;Ryan S. Selley;Ernest L. Sink;Michael D. Stover
    • Hip & pelvis
    • /
    • v.34 no.2
    • /
    • pp.87-95
    • /
    • 2022
  • Purpose: The aim of this study is to identify highly cited articles and examine trends and characteristics in research on periacetabular osteotomy. Materials and Methods: The 50 most highly cited articles on periacetabular osteotomy research were identified using Scopus. Data regarding article demographics and publication were collected from each article and an analysis was performed. Results: The mean citation count was 125±37. The article with the highest total citation count (796), five-year citation count (327), and five-year citation density (65/year) was reported by Reinhold Ganz. The five-year citation density showed strong correlation with total citation density (r=0.930, P<0.001). Reinhold Ganz, the most productive author, was listed on 13 articles in the cohort with 455 weighted citation points. Conclusion: This study provides a collection of articles examining periacetabular osteotomies and demonstrates that citation count can be regarded as an acceptable measure of the contemporary academic influence of an article.

A study on MPEG-7 descriptor combining method using borda count method (Borda count 방법을 이용한 다중 MPEG-7 서술자 조합에 관한 연구)

  • Eom, Min-Young;Choe, Yoon-Sik
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.1 s.307
    • /
    • pp.39-44
    • /
    • 2006
  • In this paper, search result list synthesis method is proposed using borda count method for still image retrieval based on MPEG-7 descriptors. MPEG-7 standardizes descriptors that extract feature information from media data. In many cases, using a single descriptor lacks of correctness, it is suggested to use multiple descriptors to enhance retrieval efficiency. In this paper, retrieval efficiency enhancement is achieved by combining multiple search results which are from each descriptor. In combining search result, newly calculated borda count method is proposed. Comparing current frequency compensated calculation, rank considered frequency compensation is used to score animage in database. This combining method is considered in Content based image retrieval system with relevance feedback algorithm which uses high level information from system user. In each relevance iteration step, adoptive borda count method is used to calculate score of images.

A Performance Analysis Based on Hadoop Application's Characteristics in Cloud Computing (클라우드 컴퓨팅에서 Hadoop 애플리케이션 특성에 따른 성능 분석)

  • Keum, Tae-Hoon;Lee, Won-Joo;Jeon, Chang-Ho
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.5
    • /
    • pp.49-56
    • /
    • 2010
  • In this paper, we implement a Hadoop based cluster for cloud computing and evaluate the performance of this cluster based on application characteristics by executing RandomTextWriter, WordCount, and PI applications. A RandomTextWriter creates given amount of random words and stores them in the HDFS(Hadoop Distributed File System). A WordCount reads an input file and determines the frequency of a given word per block unit. PI application induces PI value using the Monte Carlo law. During simulation, we investigate the effect of data block size and the number of replications on the execution time of applications. Through simulation, we have confirmed that the execution time of RandomTextWriter was proportional to the number of replications. However, the execution time of WordCount and PI were not affected by the number of replications. Moreover, the execution time of WordCount was optimum when the block size was 64~256MB. Therefore, these results show that the performance of cloud computing system can be enhanced by using a scheduling scheme that considers application's characteristics.

Estimating the Economic Value of Recreation Sea Fishing in the Yellow Sea: An Application of Count Data Model (가산자료모형을 이용한 서해 태안군 유어객의 편익추정)

  • Choi, Jong Du
    • Environmental and Resource Economics Review
    • /
    • v.23 no.2
    • /
    • pp.331-347
    • /
    • 2014
  • The purpose of this study is to estimate the economic value of the recreational sea fishing in the Yellow Sea using count data model. For estimating consumer surplus, we used several count data model of travel cost recreation demand such as a poisson model(PM), a negative binomial model(NBM), a truncated poisson model(TPM), and a truncated negative binomial model(TNBM). Model results show that there is no exist the over-dispersion problem and a NBM was statistically more suitable than the other models. All parameters estimated are statistically significant and theoretically valid. The NBM was applied to estimate the travel demand and consumer surplus. The consumer surplus pre trip was estimated to be 254,453won, total consumer surplus per person and per year 1,536,896won.

Evaluation of Biological Activated Carbon Using Particle Counter (입자계수기를 이용한 생물활성탄 공정의 효율평가)

  • Kim, Hee gune;Ryon, Dong choon;Kim, Hyun sil;Ryu, Byung soon;Moon, Seong yong;Kim, Seung Hyun;Kim, Won Kyong
    • Journal of Korean Society of Water and Wastewater
    • /
    • v.20 no.6
    • /
    • pp.823-828
    • /
    • 2006
  • For this study, an online particle counter was installed before and after the activated carbon filtration process of D water treatment plant where has advanced water treatment processes, produces average 900,000ton/day of drinking water and supply the produced drinking water to Busan citizens. We collected and analyzed particle count data for about 1 year. We inspected particle breakthrough in three out of sixteen filter processes operated at same conditions, i.e. 5th filter, 6th filter and 7th filter. According to the monitoring results, 6th and 7th filters showed similar results while 5th filter showed different results. When compared seasonal effect, the particle count for dry season was below 10 particles/ml while the particle count for August when monthly average rainfall is over 200mm was much higher than for dry season. In January and August, there was a difference in breakthrough particle size. In January, small particles in 2~3um were mainly detected while in August 10um particles were mainly detected and the size distribution was 40% of total count.

Bayesian Analysis of a Zero-inflated Poisson Regression Model: An Application to Korean Oral Hygienic Data (영과잉 포아송 회귀모형에 대한 베이지안 추론: 구강위생 자료에의 적용)

  • Lim, Ah-Kyoung;Oh, Man-Suk
    • The Korean Journal of Applied Statistics
    • /
    • v.19 no.3
    • /
    • pp.505-519
    • /
    • 2006
  • We consider zero-inflated count data, which is discrete count data but has too many zeroes compared to the Poisson distribution. Zero-inflated data can be found in various areas. Despite its increasing importance in practice, appropriate statistical inference on zero-inflated data is limited. Classical inference based on a large number theory does not fit unless the sample size is very large. And regular Poisson model shows lack of St due to many zeroes. To handle the difficulties, a mixture of distributions are considered for the zero-inflated data. Specifically, a mixture of a point mass at zero and a Poisson distribution is employed for the data. In addition, when there exist meaningful covariates selected to the response variable, loglinear link is used between the mean of the response and the covariates in the Poisson distribution part. We propose a Bayesian inference for the zero-inflated Poisson regression model by using a Markov Chain Monte Carlo method. We applied the proposed method to a Korean oral hygienic data and compared the inference results with other models. We found that the proposed method is superior in that it gives small parameter estimation error and more accurate predictions.