• Title/Summary/Keyword: cumulative distribution

Search Result 685, Processing Time 0.03 seconds

The Development and Application of Multi-metric Water Quality Assessment Model for Reservoir Managements in Korea. (우리나라 인공호 관리를 위한 다변수 수질평가 모델의 개발 및 적용)

  • Lee, Hyun-Joon;An, Kwang-Guk
    • Korean Journal of Ecology and Environment
    • /
    • v.42 no.2
    • /
    • pp.242-252
    • /
    • 2009
  • The purpose of this study was to develop a Multi-metric Water Quality Assessment (MWQA) model and apply it to dataset sampled from Paldang and Daechung reservoir in 2008. The various water dataset used to this study included 5 year data sets (2003${\sim}$2007) in Korean reservoirs which were obtained from the Ministry of Environment, Korea. In this study, suggested MWQA model has 4 metrics that were composed of 4 parameters such as chemical, physical, biological, and hydrological variables. And, each of the variables attributed total phosphorus (TP) concentration in water, secchi depth (SD) measure in water, chlorophyll-${\alpha}$(Chl-${\alpha}$) concentration in water and the ratio of inflow of water into lakes and efflux of water from lakes, input/output (I/O). First, we established the criteria for trophic boundaries. The boundary between oligotrophic and mesotrophic categories was defined by the lower third of the cumulative distribution of the values. The mesotrophic-eutrophic boundary was defined by the upper third of the distribution. Second, each metric was given by a point-oligo=1, meso=3, eu=5. And then, obtained total score from each metric was divided 5 grade-Excellent, Good, Fair, Poor, and Very poor. As the results of applying the proposed MWQA model, the Paldang reservoir obtained "Fair" or "Poor" grade and Daechung reservoir obtained "Excellent" or "Good" grade. The suggested MWQA model through these procedures will enable to manage efficiently the reservoir. And, more studies such as metric numbers and attributes should be done for the accurate application of the new model.

Study on Probabilistic Analysis for Fire·Explosion Accidents of LPG Vaporizer with Jet Fire (Jet Fire를 수반한 국내외 LPG 기화기의 화재·폭발사고에 관한 확률론적 분석에 관한 연구)

  • Ko, Jae-Sun
    • Fire Science and Engineering
    • /
    • v.26 no.4
    • /
    • pp.31-41
    • /
    • 2012
  • This study collected 5,100 cases of gas accident occurred in Korea for 14 years from 1995 to 2008, established Database and based on it, analyzed them by detailed forms and reasons. As the result of analyzing the whole city gas accidents with Poisson analysis, the item of "Careless work-Explosion-Pipeline' showed the highest rate of accidents for the next 5 years. And, "Joint Losening and corrosion-Release-Pipeline" showed the lowest rate of accident. In addition, for the result of analyzing only accidents related to LPG vaporizer, "LPG-Vaporizer-Fire" showed the highest rate of accident and "LPG-Vaporizer-Products Faults" showed the lowest rate of accident. Also, as the result of comparing and analyzing foreign LPG accident accompanied by Jet fire, facility's defect which is liquid outflow cut-off device and heat exchanger's defect were analyzed as the main reason causing jet fire, like the case of Korea, but the number of accidents for the next 5 years was the highest in "LPG-Mechanical-Jet fire" and "LPG-Mechanical-Vapor Cloud" showed the highest rate of accidents. By grafting Poisson distribution theory onto gas accident expecting program of the future, it's expected to suggest consistent standard and be used as the scale which can be used in actual field.

An Assessment of Applicability of Heat Waves Using Extreme Forecast Index in KMA Climate Prediction System (GloSea5) (기상청 현업 기후예측시스템(GloSea5)에서의 극한예측지수를 이용한 여름철 폭염 예측 성능 평가)

  • Heo, Sol-Ip;Hyun, Yu-Kyung;Ryu, Young;Kang, Hyun-Suk;Lim, Yoon-Jin;Kim, Yoonjae
    • Atmosphere
    • /
    • v.29 no.3
    • /
    • pp.257-267
    • /
    • 2019
  • This study is to assess the applicability of the Extreme Forecast Index (EFI) algorithm of the ECMWF seasonal forecast system to the Global Seasonal Forecasting System version 5 (GloSea5), operational seasonal forecast system of the Korea Meteorological Administration (KMA). The EFI is based on the difference between Cumulative Distribution Function (CDF) curves of the model's climate data and the current ensemble forecast distribution, which is essential to diagnose the predictability in the extreme cases. To investigate its applicability, the experiment was conducted during the heat-wave cases (the year of 1994 and 2003) and compared GloSea5 hindcast data based EFI with anomaly data of ERA-Interim. The data also used to determine quantitative estimates of Probability Of Detection (POD), False Alarm Ratio (FAR), and spatial pattern correlation. The results showed that the area of ERA-Interim indicating above 4-degree temperature corresponded to the area of EFI 0.8 and above. POD showed high ratio (0.7 and 0.9, respectively), when ERA-Interim anomaly data were the highest (on Jul. 11, 1994 (> $5^{\circ}C$) and Aug. 8, 2003 (> $7^{\circ}C$), respectively). The spatial pattern showed a high correlation in the range of 0.5~0.9. However, the correlation decreased as the lead time increased. Furthermore, the case of Korea heat wave in 2018 was conducted using GloSea5 forecast data to validate EFI showed successful prediction for two to three weeks lead time. As a result, the EFI forecasts can be used to predict the probability that an extreme weather event of interest might occur. Overall, we expected these results to be available for extreme weather forecasting.

A study on Bayesian beta regressions for modelling rates and proportions (비율자료 모델링을 위한 베이지안 베타회귀모형의 비교 연구)

  • Jeongin Lee;Jaeoh Kim;Seongil Jo
    • The Korean Journal of Applied Statistics
    • /
    • v.37 no.3
    • /
    • pp.339-353
    • /
    • 2024
  • In cases where the response variable in proportional data is confined to a limited interval, a regression model based on the assumption of normality can yield inaccurate results due to issues such as asymmetry and heteroscedasticity. In such cases, the beta regression model can be considered as an alternative. This model reparametrizes the beta distribution in terms of mean and precision parameters, assuming that the response variable follows a beta distribution. This allows for easy consideration of heteroscedasticity in the data. In this paper, we therefore aim to analyze proportional data using the beta regression model in two empirical analyses. Specifically, we investigate the relationship between smoking rates and coffee consumption using data from the 6th National Health Survey, and examine the association between regional characteristics in the U.S. and cumulative mortality rates based on COVID-19 data. In each analysis, we apply the ordinary least squares regression model, the beta regression model, and the extended beta regression model to analyze the data and interpret the results with the selected optimal model. The results demonstrate the appropriateness of applying the beta regression model and its extended version in proportional data.

Distributional Characteristics of Fault Segments in Cretaceous and Tertiary Rocks from Southeastern Gyeongsang Basin (경상분지 남동부 일대의 백악기 및 제3기 암류에서 발달하는 단층분절의 분포특성)

  • Park, Deok-Won
    • The Journal of the Petrological Society of Korea
    • /
    • v.27 no.3
    • /
    • pp.109-120
    • /
    • 2018
  • The distributional characteristics of fault segments in Cretaceous and Tertiary rocks from southeastern Gyeongsang Basin were derived. The 267 sets of fault segments showing linear type were extracted from the curved fault lines delineated on the regional geological map. First, the directional angle(${\theta}$)-length(L) chart for the whole fault segments was made. From the related chart, the general d istribution pattern of fault segments was derived. The distribution curve in the chart was divided into four sections according to its overall shape. NNE, NNW and WNW directions, corresponding to the peaks of the above sections, indicate those of the Yangsan, Ulsan and Gaeum fault systems. The fault segment population show near symmetrical distribution with respect to $N19^{\circ}E$ direction corresponding to the maximum peak. Second, the directional angle-frequency(N), mean length(Lm), total length(Lt) and density(${\rho}$) chart was made. From the related chart, whole domain of the above chart was divided into 19 domains in terms of the phases of the distribution curve. The directions corresponding to the peaks of the above domains suggest the directions of representative stresses acted on rock body. Third, the length-cumulative frequency graphs for the 18 sub-populations were made. From the related chart, the value of exponent(${\lambda}$) increase in the clockwise direction($N10{\sim}20^{\circ}E{\rightarrow}N50{\sim}60^{\circ}E$) and counterclockwise direction ($N10{\sim}20^{\circ}W{\rightarrow}N50{\sim}60^{\circ}W$). On the other hand, the width of distribution of lengths and mean length decrease. The chart for the above sub-populations having mutually different evolution characteristics, reveals a cross section of evolutionary process. Fourth, the general distribution chart for the 18 graphs was made. From the related chart, the above graphs were classified into five groups(A~E) according to the distribution area. The lengths of fault segments increase in order of group E ($N80{\sim}90^{\circ}E{\cdot}N70{\sim}80^{\circ}E{\cdot}N80{\sim}90^{\circ}W{\cdot}N50{\sim}60^{\circ}W{\cdot}N30{\sim}40^{\circ}W{\cdot}N40{\sim}50^{\circ}W$) < D ($N70{\sim}80^{\circ}W{\cdot}N60{\sim}70^{\circ}W{\cdot}N60{\sim}70^{\circ}E{\cdot}N50{\sim}60^{\circ}E{\cdot}N40{\sim}50^{\circ}E{\cdot}N0{\sim}10^{\circ}W$) < C ($N20{\sim}30^{\circ}W{\cdot}N10{\sim}20^{\circ}W$) < B ($N0{\sim}10^{\circ}E{\cdot}N30{\sim}40^{\circ}E$) < A ($N20{\sim}30^{\circ}E{\cdot}N10{\sim}20^{\circ}E$). Especially the forms of graph gradually transition from a uniform distribution to an exponential one. Lastly, the values of the six parameters for fault-segment length were divided into five groups. Among the six parameters, mean length and length of the longest fault segment decrease in the order of group III ($N10^{\circ}W{\sim}N20^{\circ}E$) > IV ($N20{\sim}60^{\circ}E$) > II ($N10{\sim}60^{\circ}W$) > I ($N60{\sim}90^{\circ}W$) > V ($N60{\sim}90^{\circ}E$). Frequency, longest length, total length, mean length and density of fault segments, belonging to group V, show the lowest values. The above order of arrangement among five groups suggests the interrelationship with the relative formation ages of fault segments.

Keyword Network Analysis for Technology Forecasting (기술예측을 위한 특허 키워드 네트워크 분석)

  • Choi, Jin-Ho;Kim, Hee-Su;Im, Nam-Gyu
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.227-240
    • /
    • 2011
  • New concepts and ideas often result from extensive recombination of existing concepts or ideas. Both researchers and developers build on existing concepts and ideas in published papers or registered patents to develop new theories and technologies that in turn serve as a basis for further development. As the importance of patent increases, so does that of patent analysis. Patent analysis is largely divided into network-based and keyword-based analyses. The former lacks its ability to analyze information technology in details while the letter is unable to identify the relationship between such technologies. In order to overcome the limitations of network-based and keyword-based analyses, this study, which blends those two methods, suggests the keyword network based analysis methodology. In this study, we collected significant technology information in each patent that is related to Light Emitting Diode (LED) through text mining, built a keyword network, and then executed a community network analysis on the collected data. The results of analysis are as the following. First, the patent keyword network indicated very low density and exceptionally high clustering coefficient. Technically, density is obtained by dividing the number of ties in a network by the number of all possible ties. The value ranges between 0 and 1, with higher values indicating denser networks and lower values indicating sparser networks. In real-world networks, the density varies depending on the size of a network; increasing the size of a network generally leads to a decrease in the density. The clustering coefficient is a network-level measure that illustrates the tendency of nodes to cluster in densely interconnected modules. This measure is to show the small-world property in which a network can be highly clustered even though it has a small average distance between nodes in spite of the large number of nodes. Therefore, high density in patent keyword network means that nodes in the patent keyword network are connected sporadically, and high clustering coefficient shows that nodes in the network are closely connected one another. Second, the cumulative degree distribution of the patent keyword network, as any other knowledge network like citation network or collaboration network, followed a clear power-law distribution. A well-known mechanism of this pattern is the preferential attachment mechanism, whereby a node with more links is likely to attain further new links in the evolution of the corresponding network. Unlike general normal distributions, the power-law distribution does not have a representative scale. This means that one cannot pick a representative or an average because there is always a considerable probability of finding much larger values. Networks with power-law distributions are therefore often referred to as scale-free networks. The presence of heavy-tailed scale-free distribution represents the fundamental signature of an emergent collective behavior of the actors who contribute to forming the network. In our context, the more frequently a patent keyword is used, the more often it is selected by researchers and is associated with other keywords or concepts to constitute and convey new patents or technologies. The evidence of power-law distribution implies that the preferential attachment mechanism suggests the origin of heavy-tailed distributions in a wide range of growing patent keyword network. Third, we found that among keywords that flew into a particular field, the vast majority of keywords with new links join existing keywords in the associated community in forming the concept of a new patent. This finding resulted in the same outcomes for both the short-term period (4-year) and long-term period (10-year) analyses. Furthermore, using the keyword combination information that was derived from the methodology suggested by our study enables one to forecast which concepts combine to form a new patent dimension and refer to those concepts when developing a new patent.

A Stochastic Study for the Emergency Treatment of Carbon Monoxide Poisoning in Korea (일산화탄소중독(一酸化炭素中毒)의 진료대책(診療對策) 수립(樹立)을 위한 추계학적(推計學的) 연구(硏究))

  • Kim, Yong-Ik;Yun, Dork-Ro;Shin, Young-Soo
    • Journal of Preventive Medicine and Public Health
    • /
    • v.16 no.1
    • /
    • pp.135-152
    • /
    • 1983
  • Emergency medical service is an important part of the health care delivery system, and the optimal allocation of resources and their efficient utilization are essentially demanded. Since these conditions are the prerequisite to prompt treatment which, in turn, will be crucial for life saving and in reducing the undesirable sequelae of the event. This study, taking the hyperbaric chamber for carbon monoxide poisoning as an example, is to develop a stochastic approach for solving the problems of optimal allocation of such emergency medical facility in Korea. The hyperbaric chamber, in Korea, is used almost exclusively for the treatment of acute carbon monoxide poisoning, most of which occur at home, since the coal briquette is used as domestic fuel by 69.6 per cent of the Korean population. The annual incidence rate of the comatous and fatal carbon monoxide poisoning is estimated at 45.5 per 10,000 of coal briquette-using population. It offers a serious public health problem and occupies a large portion of the emergency outpatients, especially in the winter season. The requirement of hyperbaric chambers can be calculated by setting the level of the annual queueing rate, which is here defined as the proportion of the annual number of the queued patients among the annual number of the total patients. The rate is determined by the size of the coal briquette-using population which generate a certain number of carbon monoxide poisoning patients in terms of the annual incidence rate, and the number of hyperbaric chambers per hospital to which the patients are sent, assuming that there is no referral of the patients among hospitals. The queueing occurs due to the conflicting events of the 'arrival' of the patients and the 'service' of the hyperbaric chambers. Here, we can assume that the length of the service time of hyperbaric chambers is fixed at sixty minutes, and the service discipline is based on 'first come, first served'. The arrival pattern of the carbon monoxide poisoning is relatively unique, because it usually occurs while the people are in bed. Diurnal variation of the carbon monoxide poisoning can hardly be formulated mathematically, so empirical cumulative distribution of the probability of the hourly arrival of the patients was used for Monte Carlo simulation to calculate the probability of queueing by the number of the patients per day, for the cases of one, two or three hyperbaric chambers assumed to be available per hospital. Incidence of the carbon monoxide poisoning also has strong seasonal variation, because of the four distinctive seasons in Korea. So the number of the patients per day could not be assumed to be distributed according to the Poisson distribution. Testing the fitness of various distributions of rare event, it turned out to be that the daily distribution of the carbon monoxide poisoning fits well to the Polya-Eggenberger distribution. With this model, we could forecast the number of the poisonings per day by the size of the coal-briquette using population. By combining the probability of queueing by the number of patients per day, and the probability of the number of patients per day in a year, we can estimate the number of the queued patients and the number of the patients in a year by the number of hyperbaric chamber per hospital and by the size of coal briquette-using population. Setting 5 per cent as the annual queueing rate, the required number of hyperbaric chambers was calculated for each province and for the whole country, in the cases of 25, 50, 75 and 100 per cent of the treatment rate which stand for the rate of the patients treated by hyperbaric chamber among the patients who are to be treated. Findings of the study were as follows. 1. Probability of the number of patients per day follows Polya-Eggenberger distribution. $$P(X=\gamma)=\frac{\Pi\limits_{k=1}^\gamma[m+(K-1)\times10.86]}{\gamma!}\times11.86^{-{(\frac{m}{10.86}+\gamma)}}$$ when$${\gamma}=1,2,...,n$$$$P(X=0)=11.86^{-(m/10.86)}$$ when $${\gamma}=0$$ Hourly arrival pattern of the patients turned out to be bimodal, the large peak was observed in $7 : 00{\sim}8 : 00$ a.m., and the small peak in $11 : 00{\sim}12 : 00$ p.m. 2. In the cases of only one or two hyperbaric chambers installed per hospital, the annual queueing rate will be at the level of more than 5 per cent. Only in case of three chambers, however, the rate will reach 5 per cent when the average number of the patients per day is 0.481. 3. According to the results above, a hospital equipped with three hyperbaric chambers will be able to serve 166,485, 83,242, 55,495 and 41,620 of population, when the treatmet rate are 25, 50, 75 and 100 per cent. 4. The required number of hyperbaric chambers are estimated at 483, 963, 1,441 and 1,923 when the treatment rate are taken as 25, 50, 75 and 100 per cent. Therefore, the shortage are respectively turned out to be 312, 791. 1,270 and 1,752. The author believes that the methodology developed in this study will also be applicable to the problems of resource allocation for the other kinds of the emergency medical facilities.

  • PDF

Evaluation for Rock Cleavage Using Distribution of Microcrack Spacings (V) (미세균열의 간격 분포를 이용한 결의 평가(V))

  • Park, Deok-Won
    • The Journal of the Petrological Society of Korea
    • /
    • v.26 no.3
    • /
    • pp.297-309
    • /
    • 2017
  • Jurassic granite from Geochang was analysed with respect to the characteristics of the rock cleavage. The comprehensive evaluation for rock cleavages was performed through the combination of the 16 parameters derived from the enlarged photomicrographs of the thin section and the spacing-cumulative frequency diagrams. The results of analysis for the representative values of these spacing parameters with respect to the rock cleavage are summarized as follows. First, the above parameters can be classified into group I (spacing frequency (N), total spacing ($1m{\geq}$), constant (a), exponent (${\lambda}$), slope of exponential straight line (${\theta}$), length of line (oa') and trigonometric ratios ($sin{\theta}$, $tan{\theta}$) and group II (mean spacing (Sm), difference value between mean spacing and median spacing (Sm-Sme), density (${\rho}$), lengths of lines (oa and aa'), area of a right-angled triangle (${\Delta}oaa^{\prime}$) and trigonometric ratio($cos{\theta}$). The values of the 8 parameters belonging to group I show an order of H(hardway, H1+H2)

Developing a Traffic Accident Prediction Model for Freeways (고속도로 본선에서의 교통사고 예측모형 개발)

  • Mun, Sung-Ra;Lee, Young-Ihn;Lee, Soo-Beom
    • Journal of Korean Society of Transportation
    • /
    • v.30 no.2
    • /
    • pp.101-116
    • /
    • 2012
  • Accident prediction models have been utilized to predict accident possibilities in existing or projected freeways and to evaluate programs or policies for improving safety. In this study, a traffic accident prediction model for freeways was developed for the above purposes. When selecting variables for the model, the highest priority was on the ease of both collecting data and applying them into the model. The dependent variable was set as the number of total accidents and the number of accidents including casualties in the unit of IC(or JCT). As a result, two models were developed; the overall accident model and the casualty-related accident model. The error structure adjusted to each model was the negative binomial distribution and the Poisson distribution, respectively. Among the two models, a more appropriate model was selected by statistical estimation. Major nine national freeways were selected and five-year dada of 2003~2007 were utilized. Explanatory variables should take on either a predictable value such as traffic volumes or a fixed value with respect to geometric conditions. As a result of the Maximum Likelihood estimation, significant variables of the overall accident model were found to be the link length between ICs(or JCTs), the daily volumes(AADT), and the ratio of bus volume to the number of curved segments between ICs(or JCTs). For the casualty-related accident model, the link length between ICs(or JCTs), the daily volumes(AADT), and the ratio of bus volumes had a significant impact on the accident. The likelihood ratio test was conducted to verify the spatial and temporal transferability for estimated parameters of each model. It was found that the overall accident model could be transferred only to the road with four or more than six lanes. On the other hand, the casualty-related accident model was transferrable to every road and every time period. In conclusion, the model developed in this study was able to be extended to various applications to establish future plans and evaluate policies.

Comparative Study of Reliability Design Methods by Application to Donghae Harbor Breakwaters. 1. Stability of Amor Blocks (동해항 방파제를 대상으로 한 신뢰성 설계법의 비교 연구. 1 피복 블록의 안정성)

  • Kim Seung-Woo;Suh Kyung-Duck;Oh Young Min
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.17 no.3
    • /
    • pp.188-201
    • /
    • 2005
  • This is the first part of a two-part paper which describes comparison of reliability design methods by application to Donghae Harbor Breakwaters. This paper, Part 1, is restricted to stability of armor blocks, while Part 2 deals with sliding of caissons. Reliability design methods have been developed fur breakwater designs since the mid-1980s. The reliability design method is classified into three categories depending on the level of probabilistic concepts being employed. In the Level 1 method, partial safety factors are used, which are predetermined depending on the allowable probability of failure. In the Level 2 method, the probability of failure is evaluated with the reliability index, which is calculated using the means and standard deviations of the load and resistance. The load and resistance are assumed to distribute normally. In the Level 3 method, the cumulative quantity of failure (e.g. cumulative damage of armor blocks) during the lifetime of the breakwater is calculated without assumptions of normal distribution of load and resistance. Each method calculates different design parameters, but they can be expressed in terms of probability of failure so that tile difference can be compared among the different methods. In this study, we applied the reliability design methods to the stability of armor blocks of the breakwaters of Donghae Harbor, which was constructed by traditional deterministic design methods to be damaged in 1987. Analyses are made for the breakwaters before the damage and after reinforcement. The probability of failure before the damage is much higher than the target probability of failure while that for the reinforced breakwater is much lower than the target value, indicating that the breakwaters before damage and after reinforcement were under- and over-designed, respectively. On the other hand, the results of the different reliability design methods were in fairly good agreement, confirming that there is not much difference among different methods.