Journal of Applied Statistics (J APPL STAT )

Publisher: Taylor & Francis


Journal of Applied Statistics provides a forum for communication between both applied statisticians and users of applied statistical techniques across a wide range of disciplines. These areas include business, computing, economics, ecology, education, management, medicine, operational research and sociology, but papers from other areas are also considered. The editorial policy is to publish rigorous but clear and accessible papers on applied techniques. Purely theoretical papers are avoided but those on theoretical developments which clearly demonstrate significant applied potential are welcomed. Each paper is submitted to at least two independent referees. Each issue aims for a balance of methodological innovation, thorough evaluation of existing techniques, case studies, speculative articles, book reviews and letters. Gopal Kanji, the Editor, has been running the Journal of Applied Statistics for 25 years in 1998. Journal of Applied Statistics includes a supplement on Advances in Applied Statistics. Each annual edition of the supplement aims to provide a comprehensive and modern account of a subject at the cutting edge of applied statistics. Individual articles and entire thematic issues are invited and commissioned from authors in the forefront of their speciality, linking established themes to current and future developments.

  • Impact factor
    Show impact factor history
    Impact factor
  • 5-year impact
  • Cited half-life
  • Immediacy index
  • Eigenfactor
  • Article influence
  • Website
    Journal of Applied Statistics website
  • Other titles
    Journal of applied statistics (Online)
  • ISSN
  • OCLC
  • Material type
    Document, Periodical, Internet resource
  • Document type
    Internet Resource, Computer File, Journal / Magazine / Newspaper

Publisher details

Taylor & Francis

  • Pre-print
    • Author can archive a pre-print version
  • Post-print
    • Author cannot archive a post-print version
  • Restrictions
    • 12 month embargo for STM, Behavioural Science and Public Health Journals
    • 18 month embargo for SSH journals
  • Conditions
    • Some individual journals may have policies prohibiting pre-print archiving
    • Pre-print on authors own website, Institutional or Subject Repository
    • Post-print on authors own website, Institutional or Subject Repository
    • Publisher's version/PDF cannot be used
    • On a non-profit server
    • Published source must be acknowledged
    • Must link to publisher version
    • Set statements to accompany deposits (see policy)
    • Publisher will deposit to PMC on behalf of NIH authors.
    • STM: Science, Technology and Medicine
    • SSH: Social Science and Humanities
    • 'Taylor & Francis (Psychology Press)' is an imprint of 'Taylor & Francis'
  • Classification
    ​ yellow

Publications in this journal

  • Journal of Applied Statistics 04/2014;
  • [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, the scheme of the inspection plan, namely the tightened normal tightened (nT, nN; k) is considered and procedures and necessary tables are developed for the selection of the variables sampling scheme, indexed through crossover point (COP). The importance of COP, the properties and advantages of the operating characteristic curve with respect to COP are studied.
    Journal of Applied Statistics 01/2014; 41(7):1504-1515.
  • [Show abstract] [Hide abstract]
    ABSTRACT: Within the context of California's public report of coronary artery bypass graft (CABG) surgery outcomes, we first thoroughly review popular statistical methods for profiling healthcare providers. Extensive simulation studies are then conducted to compare profiling schemes based on hierarchical logistic regression (LR) modeling under various conditions. Both Bayesian and frequentist's methods are evaluated in classifying hospitals into ‘better’, ‘normal’ or ‘worse’ service providers. The simulation results suggest that no single method would dominate others on all accounts. Traditional schemes based on LR tend to identify too many false outliers, while those based on hierarchical modeling are relatively conservative. The issue of over shrinkage in hierarchical modeling is also investigated using the 2005–2006 California CABG data set. The article provides theoretical and empirical evidence in choosing the right methodology for provider profiling.
    Journal of Applied Statistics 01/2014; 41(1).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Robust parameter design has been widely used to improve the quality of products and processes. Although a product array, in which an orthogonal array for control factors is crossed with an orthogonal array for noise factors, is commonly used for parameter design experiments, this may lead to an unacceptably large number of experimental runs. The compound noise strategy proposed by Taguchi [30] can be used to reduce the number of experimental runs. In this strategy, a compound noise factor is formed based on the directionality of the effects of noise factors. However, the directionality is usually unknown in practice. Recently, Singh et al. [28] proposed a random compound noise strategy, in which a compound noise factor is formed by randomly selecting a setting of the levels of noise factors. The present paper evaluates the random compound noise strategy in terms of the precision of the estimators of the response mean and the response variance. In addition, the variances of the estimators in the random compound noise strategy are compared with those in the n-replication design. The random compound noise strategy is shown to have smaller variances of the estimators than the 2-replication design, especially when the control-by-noise-interactions are strong.
    Journal of Applied Statistics 01/2014; 41(9).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Verification bias may occur when the test results of not all subjects are verified by using a gold standard. The correction for this bias can be made using different approaches depending on whether missing gold standard test results are random or not. Some of these approaches with binary test and gold standard results include the correction method by Begg and Greenes, lower and upper limits for diagnostic measurements by Zhou, logistic regression method, multiple imputation method, and neural networks. In this study, all these approaches are compared by employing a real and simulated data under different conditions.
    Journal of Applied Statistics 01/2014; 41(5).
  • [Show abstract] [Hide abstract]
    ABSTRACT: This article deals with the statistical inference and prediction on Burr Type XII parameters based on Type II censored sample. It is observed that the maximum likelihood estimators (MLEs) cannot be obtained in closed form. We use the expectation-maximization algorithm to compute the MLEs. We also obtain the Bayes estimators under symmetric and asymmetric loss functions such as squared error and Linex By applying Lindley's approximation and Markov chain Monte Carlo (MCMC) technique. Further, MCMC samples are used to calculate the highest posterior density credible intervals. Monte Carlo simulation study and two real-life data-sets are presented to illustrate all of the methods developed here. Furthermore, we obtain a prediction of future order statistics based on the observed ordered because of its important application in different fields such as medical and engineering sciences. A numerical example carried out to illustrate the procedures obtained for prediction of future order statistics.
    Journal of Applied Statistics 01/2014; 41(1).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Probability plots allow us to determine whether a set of sample observations is distributed according to a theoretical distribution. Plotting positions are fundamental elements in statistics and, in particular, for the construction of probability plots. In this paper, a new plotting position to construct different probability plots, such as Q–Q Plot, P–P Plot and S–P Plot, is proposed. The proposed definition is based on the median of the ith order statistic of the theoretical distribution considered. The main feature of this plotting position formula is that it is independent of the theoretical distribution selected. Moreover, the procedure developed is ‘almost’ exact, reaching, without a high cost of time, an accuracy as great as we want, which avoids using approximations (proposed by other authors).
    Journal of Applied Statistics 01/2014; 41(1).
  • [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, we consider joint modelling of repeated measurements and competing risks failure time data. For competing risks time data, a semiparametric mixture model in which proportional hazards model are specified for failure time models conditional on cause and a multinomial model for the marginal distribution of cause conditional on covariates. We also derive a score test based on joint modelling of repeated measurements and competing risks failure time data to identify longitudinal biomarkers or surrogates for a time to event outcome in competing risks data.
    Journal of Applied Statistics 01/2014; 41(10).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Empirical research frequently involves regression analysis with binary categorical variables, which are traditionally handled through dummy explanatory variables. This paper argues that separate category-wise models may provide a more logical and comprehensive tool for analysing data with binary categories. Exploring different aspects of both methods, we contrast the two with a Monte Carlo simulation and an empirical example to provide a practical insight.
    Journal of Applied Statistics 01/2014; 41(2).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Accurate diagnosis of a molecularly defined subtype of cancer is often an important step toward its effective control and treatment. For the diagnosis of some subtypes of a cancer, a gold standard with perfect sensitivity and specificity may be unavailable. In those scenarios, tumor subtype status is commonly measured by multiple imperfect diagnostic markers. Additionally, in many such studies, some subjects are only measured by a subset of diagnostic tests and the missing probabilities may depend on the unknown disease status. In this paper, we present statistical methods based on the EM algorithm to evaluate incomplete multiple imperfect diagnostic tests under a missing at random assumption and one missing not at random scenario. We apply the proposed methods to a real data set from the National Cancer Institute (NCI) colon cancer family registry on diagnosing microsatellite instability for hereditary non-polyposis colorectal cancer to estimate diagnostic accuracy parameters (i.e. sensitivities and specificities), prevalence, and potential differential missing probabilities for 11 biomarker tests. Simulations are also conducted to evaluate the small-sample performance of our methods.
    Journal of Applied Statistics 01/2014; 41(3).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Image warping is the process of deforming an image through a transformation of its domain, which is typically a subset of R 2. Given the destination of a collection of points, the problem becomes one of finding a suitable smooth interpolation for the destinations of the remaining points of the domain. A common solution is to use the thin plate spline (TPS). We find that the TPS often introduces unintended distortions of image structures. In this paper, we will analyze interpolation by TPS, experiment with other radial basis functions, and suggest two alternative functions that provide better results.
    Journal of Applied Statistics 01/2014; 41(2).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Variance estimation is an important topic in nonparametric regression. In this paper, we propose a pairwise regression method for estimating the residual variance. Specifically, we regress the squared difference between observations on the squared distance between design points, and then estimate the residual variance as the intercept. Unlike most existing difference-based estimators that require a smooth regression function, our method applies to regression models with jump discontinuities. Our method also applies to the situations where the design points are unequally spaced. Finally, we conduct extensive simulation studies to evaluate the finite-sample performance of the proposed method and compare it with some existing competitors.
    Journal of Applied Statistics 01/2014; 41(3).
  • [Show abstract] [Hide abstract]
    ABSTRACT: This paper considers the problem of analysis of covariance (ANCOVA) under the assumption of inverse Gaussian distribution for response variable. We develop the essential methodology for estimating the model parameters via maximum likelihood method. The general form of the maximum likelihood estimator is obtained in color closed form. Adjusted treatment effects and adjusted covariate effects are given, too. We also provide the asymptotic distribution of the proposed estimators. A simulation study and a real world application are also performed to illustrate and evaluate the proposed methodology.
    Journal of Applied Statistics 01/2014; 41(6).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Motivated from extreme value (EV) analysis for large non-metallic inclusions in engineering steels and a real data set, the benefit of choosing a multivariate EV approach is discussed. An extensive simulation study shows that the common univariate setup may lead to a high proportion of mis-specifications of the true EV distribution, as well as that the statistical analysis is considerably improved when being based on the respective data of r largest observations, with r appropriately chosen. Results for several underlying distributions and various values of r are presented along with effects on estimators for the parameters of the generalized EV family of distributions.
    Journal of Applied Statistics 01/2014; 41(3).
  • [Show abstract] [Hide abstract]
    ABSTRACT: We evaluate the effects of college choice on earnings using Swedish register databases. This case study is used to motivate the introduction of a novel procedure to analyse the sensitivity of such an observational study to the assumption made that there are no unobserved confounders – variables affecting both college choice and earnings. This assumption is not testable without further information, and should be considered an approximation of reality. To perform a sensitivity analysis, we measure the departure from the unconfoundedness assumption with the correlation between college choice and earnings when conditioning on observed covariates. The use of a correlation as a measure of dependence allows us to propose a standardised procedure by advocating the use of a fixed value for the correlation, typically 1% or 5%, when checking the sensitivity of an evaluation study. A correlation coefficient is, moreover, intuitive to most empirical scientists, which makes the results of our sensitivity analysis easier to communicate than those of previously proposed methods. In our evaluation of the effects of college choice on earnings, the significantly positive effect obtained could not be questioned by a sensitivity analysis allowing for unobserved confounders inducing at most 5% correlation between college choice and earnings.
    Journal of Applied Statistics 01/2014; 41(8).
  • [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, maximum likelihood and Bayes estimators of the parameters, reliability and hazard functions have been obtained for two-parameter bathtub-shaped lifetime distribution when sample is available from progressive Type-II censoring scheme. The Markov chain Monte Carlo (MCMC) method is used to compute the Bayes estimates of the model parameters. It has been assumed that the parameters have gamma priors and they are independently distributed. Gibbs within the Metropolis–Hasting algorithm has been applied to generate MCMC samples from the posterior density function. Based on the generated samples, the Bayes estimates and highest posterior density credible intervals of the unknown parameters as well as reliability and hazard functions have been computed. The results of Bayes estimators are obtained under both the balanced-squared error loss and balanced linear-exponential (BLINEX) loss. Moreover, based on the asymptotic normality of the maximum likelihood estimators the approximate confidence intervals (CIs) are obtained. In order to construct the asymptotic CI of the reliability and hazard functions, we need to find the variance of them, which are approximated by delta and Bootstrap methods. Two real data sets have been analyzed to demonstrate how the proposed methods can be used in practice.
    Journal of Applied Statistics 01/2014; 41(4).
  • [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, we formulate a very flexible family of models which unifies most recent lifetime distributions. The main idea is to obtain a cumulative distribution function to transform the baseline distribution with an activation mechanism characterized by a latent threshold variable. The new family has a strong biological interpretation from the competitive risks point of view and the Box–Cox transformation provides an elegant manner to interpret the effect on the baseline distribution to obtain this alternative model. Several structural properties of the new model are investigated. A Bayesian analysis using Markov Chain Monte Carlo procedure is developed to illustrate with a real data the usefulness of the proposed family.
    Journal of Applied Statistics 01/2014; 41(9).
  • [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, we propose a Bayesian partition modeling for lifetime data in the presence of a cure fraction by considering a local structure generated by a tessellation which depends on covariates. In this modeling we include information of nominal qualitative variables with more than two categories or ordinal qualitative variables. The proposed modeling is based on a promotion time cure model structure but assuming that the number of competing causes follows a geometric distribution. It is an alternative modeling strategy to the conventional survival regression modeling generally used for modeling lifetime data in the presence of a cure fraction, which models the cure fraction through a (generalized) linear model of the covariates. An advantage of our approach is its ability to capture the effects of covariates in a local structure. The flexibility of having a local structure is crucial to capture local effects and features of the data. The modeling is illustrated on two real melanoma data sets.
    Journal of Applied Statistics 01/2014; 41(3).
  • Journal of Applied Statistics 01/2014; 41(6).
  • [Show abstract] [Hide abstract]
    ABSTRACT: Compositional tables represent a continuous counterpart to well-known contingency tables. Their cells contain quantitatively expressed relative contributions of a whole, carrying exclusively relative information and are popularly represented in proportions or percentages. The resulting factors, corresponding to rows and columns of the table, can be inspected similarly as with contingency tables, e.g. for their mutual independent behaviour. The nature of compositional tables requires a specific geometrical treatment, represented by the Aitchison geometry on the simplex. The properties of the Aitchison geometry allow a decomposition of the original table into its independent and interactive parts. Moreover, the specific case of 2×2 compositional tables allows the construction of easily interpretable orthonormal coordinates (resulting from the isometric logratio transformation) for the original table and its decompositions. Consequently, for a sample of compositional tables both explorative statistical analysis like graphical inspection of the independent and interactive parts or any statistical inference (odds-ratio-like testing of independence) can be performed. Theoretical advancements of the presented approach are demonstrated using two economic applications.
    Journal of Applied Statistics 01/2014; 41(5).

Related Journals