• Title/Summary/Keyword: statistical approach

Search Result 2,344, Processing Time 0.028 seconds

Support Vector Machine for Interval Regression

  • Hong Dug Hun;Hwang Changha
    • Proceedings of the Korean Statistical Society Conference
    • /
    • 2004.11a
    • /
    • pp.67-72
    • /
    • 2004
  • Support vector machine (SVM) has been very successful in pattern recognition and function estimation problems for crisp data. This paper proposes a new method to evaluate interval linear and nonlinear regression models combining the possibility and necessity estimation formulation with the principle of SVM. For data sets with crisp inputs and interval outputs, the possibility and necessity models have been recently utilized, which are based on quadratic programming approach giving more diverse spread coefficients than a linear programming one. SVM also uses quadratic programming approach whose another advantage in interval regression analysis is to be able to integrate both the property of central tendency in least squares and the possibilistic property In fuzzy regression. However this is not a computationally expensive way. SVM allows us to perform interval nonlinear regression analysis by constructing an interval linear regression function in a high dimensional feature space. In particular, SVM is a very attractive approach to model nonlinear interval data. The proposed algorithm here is model-free method in the sense that we do not have to assume the underlying model function for interval nonlinear regression model with crisp inputs and interval output. Experimental results are then presented which indicate the performance of this algorithm.

  • PDF

A Development of the Ship Weight Estimating Method by a Statistical Approach (통계적 접근법에 의한 선박 중량추정 방법 개발)

  • Cho, Yong-Jin
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.48 no.5
    • /
    • pp.426-434
    • /
    • 2011
  • Accurate weight prediction methods are an essential of the ship design in both ship cost managements and performance satisfactions. When no parent or similar ships are available, an adequate method of the ship weight estimating is required. In this study, there was carried out to develop the ship weight estimating method for the preliminary design phase. The weight estimating methods were first surveyed by the references and summarized their characteristics. The weight estimation method by statistical approach was developed for the container ship because the containerized transportation markets is gradually growing and ship's size and loading capacity are rapidly enlarged. The correlation analysis and the multiple regression analysis were used for developing the weight estimating method. As a results of evaluating the developed method, the error ratio of the variation between estimated weight and ship's data was about 5%. And it was only 1% difference with the calculating weight of conceptual design results by shipyard design team that the estimating weight of ultra-large container ship was predicted by the developed method.

A Statistical Model-Based Voice Activity Detection Employing the Conditional MAP Criterion with Spectral Deviation (조건 사후 최대 확률과 음성 스펙트럼 변이 조건을 이용한 통계적 모델 기반의 음성 검출기)

  • Kim, Sang-Kyun;Chang, Joon-Hyuk
    • The Journal of the Acoustical Society of Korea
    • /
    • v.30 no.6
    • /
    • pp.324-329
    • /
    • 2011
  • In this paper, we propose a novel approach to improve the performance of a statistical model-based voice activity detection (VAD) which is based on the conditional maximum a posteriori (CMAP) with deviation. In our approach, the VAD decision rule is expressed as the geometric mean of likelihood ratios (LRs) based on adapted threshold according to the speech presence probability conditioned on both the speech activity decisions and spectral deviation in the pervious frame. Experimental results show that the proposed approach yields better results compared to the CMAP-based VAD using the LR test.

Comparative Study of Probabilistic Ecological Risk Assessment (PERA) used in Developed Countries and Proposed PERA approach for Korean Water Environment (확률생태위해성평가(PERA) 선진국 사례분석 및 국내수계에 적합한 PERA 기법 제안)

  • An, Youn-Joo;Nam, Sun-Hwa;Lee, Woo-Mi
    • Journal of Korean Society on Water Environment
    • /
    • v.25 no.4
    • /
    • pp.494-501
    • /
    • 2009
  • Probabilistic Ecological risk assessment (PERA) is extensive approach to qualify and quantify risk on the multi species based on species sensitivity distribution (SSD). As a while, deterministic ecological risk assessment (DERA) considers the comparison of predicted no-effect concentration (PNEC) and predicted exposure concentration (PEC). DERA is used to determine if there is potential risk or no risk, and it doesn't consider the nature variability and the species sensitivity. But PERA can be more realistic and reasonable approach to estimate likelihood or risk. In this study, we compared PERA used in developed countries, and proposed PERA applicable for the Korean water environment. Taxonomic groups were classified as "class" level including Actinopterygill, Branchiopoda, Chlorophyceae, Maxillapoda, Insects, Bivalvia, Gastropoda, Secernentea, Polychaeta, Monocotyldoneae, and Chanophyceae in this study. Statistical extrapolation method (SEM), statistical extrapolation method $_{acutechronicratio}$ ($SEM_{ACR}$) and assessment factor method (AFM) were used to calculate the ecological protective concentration based on qualitative and quantitative levels of taxonomic toxicity data. This study would be useful to establish the PERA for the protection of aquatic ecosystem in Korea.

Dimension reduction for right-censored survival regression: transformation approach

  • Yoo, Jae Keun;Kim, Sung-Jin;Seo, Bi-Seul;Shin, Hyejung;Sim, Su-Ah
    • Communications for Statistical Applications and Methods
    • /
    • v.23 no.3
    • /
    • pp.259-268
    • /
    • 2016
  • High-dimensional survival data with large numbers of predictors has become more common. The analysis of such data can be facilitated if the dimensions of predictors are adequately reduced. Recent studies show that a method called sliced inverse regression (SIR) is an effective dimension reduction tool in high-dimensional survival regression. However, it faces incapability in implementation due to a double categorization procedure. This problem can be overcome in the right-censoring type by transforming the observed survival time and censoring status into a single variable. This provides more flexibility in the categorization, so the applicability of SIR can be enhanced. Numerical studies show that the proposed transforming approach is equally good to (or even better) than the usual SIR application in both balanced and highly-unbalanced censoring status. The real data example also confirms its practical usefulness, so the proposed approach should be an effective and valuable addition to usual statistical practitioners.

AN ASSESSMENT OF UNCERTAINTY ON A LOFT L2-5 LBLOCA PCT BASED ON THE ACE-RSM APPROACH: COMPLEMENTARY WORK FOR THE OECD BEMUSE PHASE-III PROGRAM

  • Ahn, Kwang-Il;Chung, Bub-Dong;Lee, John C.
    • Nuclear Engineering and Technology
    • /
    • v.42 no.2
    • /
    • pp.163-174
    • /
    • 2010
  • As pointed out in the OECD BEMUSE Program, when a high computation time is taken to obtain the relevant output values of a complex physical model (or code), the number of statistical samples that must be evaluated through it is a critical factor for the sampling-based uncertainty analysis. Two alternative methods have been utilized to avoid the problem associated with the size of these statistical samples: one is based on Wilks' formula, which is based on simple random sampling, and the other is based on the conventional nonlinear regression approach. While both approaches provide a useful means for drawing conclusions on the resultant uncertainty with a limited number of code runs, there are also some unique corresponding limitations. For example, a conclusion based on the Wilks' formula can be highly affected by the sampled values themselves, while the conventional regression approach requires an a priori estimate on the functional forms of a regression model. The main objective of this paper is to assess the feasibility of the ACE-RSM approach as a complementary method to the Wilks' formula and the conventional regression-based uncertainty analysis. This feasibility was assessed through a practical application of the ACE-RSM approach to the LOFT L2-5 LBLOCA PCT uncertainty analysis, which was implemented as a part of the OECD BEMUSE Phase III program.

Korean Part-of-Speech Tagging System Using Resolution Rules for Individual Ambiguous Word (어절별 중의성 해소 규칙을 이용한 혼합형 한국어 품사 태깅 시스템)

  • Park, Hee-Geun;Ahn, Young-Min;Seo, Young-Hoon
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.13 no.6
    • /
    • pp.427-431
    • /
    • 2007
  • In this paper we describe a Korean part-of-speech tagging approach using resolution rules for individual ambiguous word and statistical information. Our tagging approach resolves lexical ambiguities by common rules, rules for individual ambiguous word, and statistical approach. Common rules are ones for idioms and phrases of common use including phrases composed of main and auxiliary verbs. We built resolution rules for each word which has several distinct morphological analysis results to enhance tagging accuracy. Each rule may have morphemes, morphological tags, and/or word senses of not only an ambiguous word itself but also words around it. Statistical approach based on HMM is then applied for ambiguous words which are not resolved by rules. Experiment shows that the part-of-speech tagging approach has high accuracy and broad coverage.

A Statistical Approach for the Determination of Location and Range of Fishing Grounds in the Coastal Vessel Fishery (통계학적 접근에 의한 연안어업의 조업어장 위치 및 범위결정 에 관한 연구)

  • Kang, Y.J.;Kim, K.S.;Yi, S.B.
    • The Journal of Fisheries Business Administration
    • /
    • v.38 no.1 s.73
    • /
    • pp.97-113
    • /
    • 2007
  • This study attempts to suggest a new approach of the the determination of location and range of fishing ground in the coastal vessel fishery using the anaiysis of variance(ANOVA) and least significant difference test(LSD test) in statistics. The important result of the empirical study is that there is no significant difference in fisfishing ground based on between fishermen's argument and spot investigation. But there is stastitically significant difference in fisfishing ground based on between fishermen's argument and the official documents of fishing records for past three years. The results suggest us that there exists a possibility of misuse of tax free oil providied for subsidizing fisheries. But because of the range limitation of data, it is impossible for us to generalize the above results.

  • PDF

Variable Selection and Outlier Detection for Automated K-means Clustering

  • Kim, Sung-Soo
    • Communications for Statistical Applications and Methods
    • /
    • v.22 no.1
    • /
    • pp.55-67
    • /
    • 2015
  • An important problem in cluster analysis is the selection of variables that define cluster structure that also eliminate noisy variables that mask cluster structure; in addition, outlier detection is a fundamental task for cluster analysis. Here we provide an automated K-means clustering process combined with variable selection and outlier identification. The Automated K-means clustering procedure consists of three processes: (i) automatically calculating the cluster number and initial cluster center whenever a new variable is added, (ii) identifying outliers for each cluster depending on used variables, (iii) selecting variables defining cluster structure in a forward manner. To select variables, we applied VS-KM (variable-selection heuristic for K-means clustering) procedure (Brusco and Cradit, 2001). To identify outliers, we used a hybrid approach combining a clustering based approach and distance based approach. Simulation results indicate that the proposed automated K-means clustering procedure is effective to select variables and identify outliers. The implemented R program can be obtained at http://www.knou.ac.kr/~sskim/SVOKmeans.r.

Statistical approach to a SHM benchmark problem

  • Casciati, Sara
    • Smart Structures and Systems
    • /
    • v.6 no.1
    • /
    • pp.17-27
    • /
    • 2010
  • The approach to damage detection and localization adopted in this paper is based on a statistical comparison of models built from the response time histories collected at different stages during the structure lifetime. Some of these time histories are known to have been recorded when the structural system was undamaged. The consistency of the models associated to two different stages, both undamaged, is first recognized. By contrast, the method detects the discrepancies between the models from measurements collected for a damaged situation and for the undamaged reference situation. The damage detection and localization is pursued by a comparison of the SSE (sum of the squared errors) histograms. The validity of the proposed approach is tested by applying it to the analytical benchmark problem developed by the ASCE Task Group on Structural Health Monitoring (SHM). In the paper, the results of the benchmark studies are presented and the performance of the method is discussed.