• Title/Summary/Keyword: Variable Statistics

Search Result 1,353, Processing Time 0.024 seconds

REAL-VARIABLE CHARACTERIZATIONS OF VARIABLE HARDY SPACES ON LIPSCHITZ DOMAINS OF ℝn

  • Liu, Xiong
    • Bulletin of the Korean Mathematical Society
    • /
    • v.58 no.3
    • /
    • pp.745-765
    • /
    • 2021
  • Let Ω be a proper open subset of ℝn and p(·) : Ω → (0, ∞) be a variable exponent function satisfying the globally log-Hölder continuous condition. In this article, the author introduces the "geometrical" variable Hardy spaces Hp(·)r (Ω) and Hp(·)z (Ω) on Ω, and then obtains the grand maximal function characterizations of Hp(·)r (Ω) and Hp(·)z (Ω) when Ω is a strongly Lipschitz domain of ℝn. Moreover, the author further introduces the "geometrical" variable local Hardy spaces hp(·)r (Ω), and then establishes the atomic characterization of hp(·)r (Ω) when Ω is a bounded Lipschitz domain of ℝn.

Variable selection with quantile regression tree (분위수 회귀나무를 이용한 변수선택 방법 연구)

  • Chang, Youngjae
    • The Korean Journal of Applied Statistics
    • /
    • v.29 no.6
    • /
    • pp.1095-1106
    • /
    • 2016
  • The quantile regression method proposed by Koenker et al. (1978) focuses on conditional quantiles given by independent variables, and analyzes the relationship between response variable and independent variables at the given quantile. Considering the linear programming used for the estimation of quantile regression coefficients, the model fitting job might be difficult when large data are introduced for analysis. Therefore, dimension reduction (or variable selection) could be a good solution for the quantile regression of large data sets. Regression tree methods are applied to a variable selection for quantile regression in this paper. Real data of Korea Baseball Organization (KBO) players are analyzed following the variable selection approach based on the regression tree. Analysis result shows that a few important variables are selected, which are also meaningful for the given quantiles of salary data of the baseball players.

Comments on the regression coefficients (다중회귀에서 회귀계수 추정량의 특성)

  • Kahng, Myung-Wook
    • The Korean Journal of Applied Statistics
    • /
    • v.34 no.4
    • /
    • pp.589-597
    • /
    • 2021
  • In simple and multiple regression, there is a difference in the meaning of regression coefficients, and not only are the estimates of regression coefficients different, but they also have different signs. Understanding the relative contribution of explanatory variables in a regression model is an important part of regression analysis. In a standardized regression model, the regression coefficient can be interpreted as the change in the response variable with respect to the standard deviation when the explanatory variable increases by the standard deviation in a situation where the values of the explanatory variables other than the corresponding explanatory variable are fixed. However, the size of the standardized regression coefficient is not a proper measure of the relative importance of each explanatory variable. In this paper, the estimator of the regression coefficient in multiple regression is expressed as a function of the correlation coefficient and the coefficient of determination. Furthermore, it is considered in terms of the effect of an additional explanatory variable and additional increase in the coefficient of determination. We also explore the relationship between estimates of regression coefficients and correlation coefficients in various plots. These results are specifically applied when there are two explanatory variables.

Regression Trees with. Unbiased Variable Selection (변수선택 편향이 없는 회귀나무를 만들기 위한 알고리즘)

  • 김진흠;김민호
    • The Korean Journal of Applied Statistics
    • /
    • v.17 no.3
    • /
    • pp.459-473
    • /
    • 2004
  • It has well known that an exhaustive search algorithm suggested by Breiman et. a1.(1984) has a trend to select the variable having relatively many possible splits as an splitting rule. We propose an algorithm to overcome this variable selection bias problem and then construct unbiased regression trees based on the algorithm. The proposed algorithm runs two steps of selecting a split variable and determining a split rule for binary split based on the split variable. Simulation studies were performed to compare the proposed algorithm with Breiman et a1.(1984)'s CART(Classification and Regression Tree) in terms of degree of variable selection bias, variable selection power, and MSE(Mean Squared Error). Also, we illustrate the proposed algorithm with real data sets.

Variable Selection and Outlier Detection for Automated K-means Clustering

  • Kim, Sung-Soo
    • Communications for Statistical Applications and Methods
    • /
    • v.22 no.1
    • /
    • pp.55-67
    • /
    • 2015
  • An important problem in cluster analysis is the selection of variables that define cluster structure that also eliminate noisy variables that mask cluster structure; in addition, outlier detection is a fundamental task for cluster analysis. Here we provide an automated K-means clustering process combined with variable selection and outlier identification. The Automated K-means clustering procedure consists of three processes: (i) automatically calculating the cluster number and initial cluster center whenever a new variable is added, (ii) identifying outliers for each cluster depending on used variables, (iii) selecting variables defining cluster structure in a forward manner. To select variables, we applied VS-KM (variable-selection heuristic for K-means clustering) procedure (Brusco and Cradit, 2001). To identify outliers, we used a hybrid approach combining a clustering based approach and distance based approach. Simulation results indicate that the proposed automated K-means clustering procedure is effective to select variables and identify outliers. The implemented R program can be obtained at http://www.knou.ac.kr/~sskim/SVOKmeans.r.

Variable selection in Poisson HGLMs using h-likelihoood

  • Ha, Il Do;Cho, Geon-Ho
    • Journal of the Korean Data and Information Science Society
    • /
    • v.26 no.6
    • /
    • pp.1513-1521
    • /
    • 2015
  • Selecting relevant variables for a statistical model is very important in regression analysis. Recently, variable selection methods using a penalized likelihood have been widely studied in various regression models. The main advantage of these methods is that they select important variables and estimate the regression coefficients of the covariates, simultaneously. In this paper, we propose a simple procedure based on a penalized h-likelihood (HL) for variable selection in Poisson hierarchical generalized linear models (HGLMs) for correlated count data. For this we consider three penalty functions (LASSO, SCAD and HL), and derive the corresponding variable-selection procedures. The proposed method is illustrated using a practical example.

WEAK HERZ-TYPE HARDY SPACES WITH VARIABLE EXPONENTS AND APPLICATIONS

  • Souad Ben Seghier
    • Journal of the Korean Mathematical Society
    • /
    • v.60 no.1
    • /
    • pp.33-69
    • /
    • 2023
  • Let α ∈ (0, ∞), p ∈ (0, ∞) and q(·) : ℝn → [1, ∞) satisfy the globally log-Hölder continuity condition. We introduce the weak Herz-type Hardy spaces with variable exponents via the radial grand maximal operator and to give its maximal characterizations, we establish a version of the boundedness of the Hardy-Littlewood maximal operator M and the Fefferman-Stein vector-valued inequality on the weak Herz spaces with variable exponents. We also obtain the atomic and the molecular decompositions of the weak Herz-type Hardy spaces with variable exponents. As an application of the atomic decomposition we provide various equivalent characterizations of our spaces by means of the Lusin area function, the Littlewood-Paley g-function and the Littlewood-Paley $g^*_{\lambda}$-function.

Censored varying coefficient regression model using Buckley-James method

  • Shim, Jooyong;Seok, Kyungha
    • Journal of the Korean Data and Information Science Society
    • /
    • v.28 no.5
    • /
    • pp.1167-1177
    • /
    • 2017
  • The censored regression using the pseudo-response variable proposed by Buckley and James has been one of the most well-known models. Recently, the varying coefficient regression model has received a great deal of attention as an important tool for modeling. In this paper we propose a censored varying coefficient regression model using Buckley-James method to consider situations where the regression coefficients of the model are not constant but change as the smoothing variables change. By using the formulation of least squares support vector machine (LS-SVM), the coefficient estimators of the proposed model can be easily obtained from simple linear equations. Furthermore, a generalized cross validation function can be easily derived. In this paper, we evaluated the proposed method and demonstrated the adequacy through simulate data sets and real data sets.

Evaluating Properties of Variable Sampling Interval EWMA Control Charts for Mean Vector

  • Kwon, Yong-Man;Chang, Duk-Joon
    • Journal of the Korean Data and Information Science Society
    • /
    • v.16 no.3
    • /
    • pp.639-650
    • /
    • 2005
  • Theoretical and numerical comparison have shown that variable sampling interval (VSI) charts are substantially more efficient than fixed sampling interval(FSI) charts in term of ATS(average time to signal). But the frequency of switching between different sampling intervals is a complicating factor in VSI procedures. VSI EWMA charts for monitoring mean vector of related qualify characteristics are investigated. To compare the efficiencies of the proposed charts, the performances are evaluated for matched FSI and VSI charts in terms of average time to signal(ATS) and average number of samples to signal(ANSS). For the switching behavior of the proposed VSI charts, average number of switches(ANSW) are also investigated.

  • PDF

Application of Logit Model in Qualitative Dependent Variables (로짓모형을 이용한 질적 종속변수의 분석)

  • Lee, Kil-Soon;Yu, Wann
    • Journal of Families and Better Life
    • /
    • v.10 no.1 s.19
    • /
    • pp.131-138
    • /
    • 1992
  • Regression analysis has become a standard statistical tool in the behavioral science. Because of its widespread popularity. regression has been often misused. Such is the case when the dependent variable is a qualitative measure rather than a continuous, interval measure. Regression estimates with a qualitative dependent variable does not meet the assumptions underlying regression. It can lead to serious errors in the standard statistical inference. Logit model is recommended as alternatives to the regression model for qualitative dependent variables. Researchers can employ this model to measure the relationship between independent variables and qualitative dependent variables without assuming that logit model was derived from probabilistic choice theory. Coefficients in logit model are typically estimated by the method of Maximum Likelihood Estimation in contrast to ordinary regression model which estimated by the method of Least Squares Estimation. Goodness of fit in logit model is based on the likelihood ratio statistics and the t-statistics is used for testing the null hypothesis.

  • PDF