Ching-Kang Ing

Ching-Kang Ing
National Tsing Hua University | NTHU · Institute of Statistics

About

55
Publications
5,049
Reads
How we measure 'reads'
A 'read' is counted each time someone views a publication summary (such as the title, abstract, and list of authors), clicks on a figure, or views or downloads the full-text. Learn more
872
Citations

Publications

Publications (55)
Article
Although existing literature on high-dimensional regression models is rich, the vast majority of studies have focused on independent and homogeneous error terms. In this article, we consider the problem of selecting high-dimensional regression models with heteroscedastic and time series errors, which have broad applications in economics, quantitati...
Preprint
Principal component analysis (PCA) is the most commonly used statistical procedure for dimension reduction. An important issue for applying PCA is to determine the rank, which is the number of dominant eigenvalues of the covariance matrix. The Akaike information criterion (AIC) and Bayesian information criterion (BIC) are among the most widely used...
Preprint
Full-text available
We investigate the prediction capability of the orthogonal greedy algorithm (OGA) in high-dimensional regression models with dependent observations. The rates of convergence of the prediction error of OGA are obtained under a variety of sparsity conditions. To prevent OGA from overfitting, we introduce a high-dimensional Akaike's information criter...
Article
Consider finite parametric time series models. “I have n observations and k models, which model should I choose on the basis of the data alone” is a frequently asked question in many practical situations. This poses the key problem of selecting a model from a collection of candidate models, none of which is necessarily the true data generating proc...
Article
Full-text available
We introduce a fast stepwise regression method, called the orthogonal greedy algorithm (OGA), that selects input variables to enter a p-dimensional linear regression model (with p n, the sample size) sequentially so that the selected variable at each step minimizes the residual sum squares. We derive the convergence rate of OGA and develop a consis...
Preprint
Full-text available
We establish a negative moment bound for the sample autocovariance matrix of a stationary process driven by conditional heteroscedastic errors. This moment bound enables us to asymptotically express the mean squared prediction error (MSPE) of the least squares predictor as the sum of three terms related to model complexity, model misspecification,...
Article
We consider model selection for linear mixed‐effects models with clustered structure, where conditional Kullback‐Leibler (CKL) loss is applied to measure the efficiency of the selection. We estimate the CKL loss by substituting the empirical best linear unbiased predictors (EBLUPs) into random effects with model parameters estimated by maximum like...
Article
This paper studies an important sequential decision making problem known as the multi-armed stochastic bandit problem with covariates. Under a linear bandit framework with high-dimensional covariates, we propose a general multi-stage arm allocation algorithm that integrates both arm elimination and randomized assignment strategies. By employing a c...
Article
Full-text available
Principal component analysis (PCA) is a commonly used statistical tool for dimension reduction. An important issue in PCA is to determine the rank, which is the number of dominant eigenvalues of the covariance matrix. Among information-based criteria, the Akaike information criterion (AIC) and the Bayesian information criterion (BIC) are the two mo...
Article
Estimating the orders of the autoregressive fractionally integrated moving average (ARFIMA) model has been a long-standing problem in time series analysis. This paper tackles this challenge by establishing the consistency of the Bayesian information criterion (BIC) for ARFIMA models with independent errors. Since the memory parameter of the model c...
Article
We consider a linear mixed-effects model with a clustered structure, where the parameters are estimated using maximum likelihood (ML) based on possibly unbalanced data. Inference with this model is typically done based on asymptotic theory, assuming that the number of clusters tends to infinity with the sample size. However, when the number of clus...
Preprint
Full-text available
The framework of model-X knockoffs provides a flexible tool for exact finite-sample false discovery rate (FDR) control in variable selection. It also completely bypasses the use of conventional p-values, making it especially appealing in high-dimensional nonlinear models. Existing works have focused on the setting of independent and identically dis...
Article
The concepts of Industry 4.1 for achieving Zero-Defect (ZD) manufacturing were disclosed in IEEE Robotics and Automation Letters in January 2016. ZD of all the deliverables can be achieved by discarding the defective products via a real-time and online total inspection technology, such as Automatic Virtual Metrology (AVM). Further, the Key-variab...
Preprint
We consider a linear mixed-effects model with a clustered structure, where the parameters are estimated using maximum likelihood (ML) based on possibly unbalanced data. Inference with this model is typically done based on asymptotic theory, assuming that the number of clusters tends to infinity with the sample size. However, when the number of clus...
Article
We are interested in constructing confidence intervals for the autoregressive (AR) coefficient of a first‐order AR model with i.i.d. positive errors via an extreme value estimate (EVE). We assume that the error distribution has a density function fε(x) behaving like as x→0, where b1,0 and α0 are unknown positive constants. These specifications impl...
Article
Prediction has long been a vibrant topic in modern probability and statistics. In addition to finding optimal forecasts and for model selection, it is argued in this paper that the prediction principle can also be used to analyze critical phenomena, in particular, in stationary and unstable time series. Although the notion of nearly unstable models...
Article
Full-text available
We consider a stochastic process model with time trend and measurement error. We establish consistency and derive the limiting distributions of the maximum likelihood (ML) estimators of the covariance function parameters under a general asymptotic framework, including both the fixed domain and the increasing domain frameworks, even when the time tr...
Article
Motivated by applications to root-cause identification of faults in multistage manufacturing processes which involve a large number of tools or equipments at each stage, we consider multiple testing in regression models whose outputs represent the quality characteristics of a multistage manufacturing process. Because of the large number of input va...
Article
Full-text available
This work aims at estimating inverse autocovariance matrices of long memory processes admitting a linear representation. A modified Cholesky decomposition is used in conjunction with an increasing order autoregressive model to achieve this goal. The spectral norm consistency of the proposed estimate is established. We then extend this result to lin...
Article
Threshold autoregressive (TAR) model is an important class of nonlinear time series models that possess many desirable features such as asymmetric limit cycles and amplitude dependent frequencies. Statistical inference for TAR model encounters a major difficulty in the estimation of thresholds, however. This paper develops an efficient procedure to...
Article
There is an extensive literature on fixed-size confidence regions for the regression parameters in a linear model with p regressors, attaining a prescribed coverage probability when p is fixed and the size d approaches 0. Motivated by recent developments in regression modeling in response to applications for which p is considerably larger than the...
Article
Consider a regression model with in?nitely many parameters and time series errors. We are interested in choosing weights for averaging across generalized least squares (GLS) estimators obtained from a set of approximating models. However, GLS estimators, depending on the unknown inverse covariance matrix of the errors, are usually infeasible. We th...
Article
Full-text available
Information criteria, such as Akaike's information criterion and Bayesian information criterion are often applied in model selection. However, their asymptotic behaviors for selecting geostatistical regression models have not been well studied, particularly under the fixed domain asymptotic framework with more and more data observed in a bounded fi...
Article
In this paper, we assume that observations are generated by a linear regression model with short- or long-memory dependent errors. We establish inverse moment bounds for -dimensional sample autocovariance matrices based on the least squares residuals (also known as the detrended time series), where , and n is the sample size. These results are then...
Article
Let observations y 1, …, yn be generated from a first-order autoregressive (AR) model with positive errors. In both the stationary and unit root cases, we derive moment bounds and limiting distributions of an extreme value estimator, [Inline formula], of the AR coefficient. These results enable us to provide asymptotic expressions for the mean squa...
Article
Full-text available
A moment bound for the normalized conditional-sum-of-squares (CSS) estimate of a general autoregressive fractionally integrated moving average (ARFIMA) model with an arbitrary unknown memory parameter is derived in this paper. To achieve this goal, a uniform moment bound for the inverse of the normalized objective function is established. An import...
Article
Consider l(q)-hulls, 0 < q <= 1, from a dictionary of M functions in L-P space for 1 < p < cc. Their precise metric entropy orders are derived. Sparse linear approximation bounds are obtained to characterize the number of terms needed to achieve accurate approximation of the best function in a l(q)-hull that is closest to a target function. Further...
Article
Full-text available
non-stationary autoregressions
Article
We show that Akaike’s Information Criterion (AIC) and its variants are asymptotically efficient in integrated autoregressive processes of infinite order (AR(∞)). This result, together with its stationary counterpart established previously in the literature, ensures that AIC can ultimately achieve prediction efficiency in an AR(∞) process, without k...
Article
The article [Sequential Anal. 30, No. 4, 356–399 (2011; Zbl 1228.62096)] by M. Aoshima and K. Yata considers two-stage procedures to construct fixed-size confidence regions and to select variables in high-dimensional models with a bounded number of nonzero parameters. In this discussion, we describe some recent work on alternative variable selectio...
Article
Full-text available
In this paper, a uniform (over some parameter space) moment bound for the inverse of Fisher’s information matrix is established. This result is then applied to develop moment bounds for the normalized least squares estimate in (nonlinear) stochastic regression models. The usefulness of these results is illustrated using time series models. In parti...
Article
Full-text available
Assume that observations are generated from nonstationary autoregressive (AR) processes of infinite order. We adopt a finite-order approximation model to predict future observations and obtain an asymptotic expression for the mean-squared prediction error (MSPE) of the least squares predictor. This expression provides the first exact assessment of...
Article
Full-text available
This paper investigates multistep prediction errors for non-stationary autoregressive processes with both model order and true parameters unknown. We give asymptotic expressions for the multistep mean squared prediction errors and accumulated prediction errors of two important methods, plug-in and direct prediction. These expressions not only chara...
Article
Full-text available
Findings from literature showed inconsistent results for applying service quality scale in hospitals. Moreover, hospitalization services are provided by diversified departments and a scale designed to measure the overall hospitalization quality is difficult and capturing special characteristics of different departments is also not an easy task. Thi...
Article
Full-text available
A major research area of Ching-Zong Wei (1949--2004) was time series models and their applications in econometrics and engineering, to which he made many important contributions. A conference on time series and related topics in memory of him was held on December 12--14, 2005, at Academia Sinica in Taipei, where he was Director of the Institute of...
Article
Full-text available
In this article asymptotic expressions for the final prediction error (FPE) and the accumulated prediction error (APE) of the least squares predictor are obtained in regression models with nonstationary regressors. It is shown that the term of order $1/n$ in FPE and the term of order $\log n$ in APE share the same constant, where $n$ is the sample...
Article
Full-text available
Assume that observations are generated from an infinite-order autoregressive [AR($\infty$)] process. Shibata [Ann. Statist. 8 (1980) 147--164] considered the problem of choosing a finite-order AR model, allowing the order to become infinite as the number of observations does in order to obtain a better approximation. He showed that, for the purpose...
Article
Full-text available
We establish a maximal moment inequality for the weighted sum of a sequence of random variables with finite second moments. An extension of the J. Hájek and A. Rény [Acta Math. Acad. Sci. Hung. 6, 281–283 (1956; Zbl 0067.10701)] and Y. S. Chow [Proc. Am. Math. Soc. 11, 107–111 (1960; Zbl 0102.13501)] inequalities is then obtained. When certain seco...
Article
Full-text available
The predictive capability of a modification of Rissanen's accumulated prediction error (APE) criterion, APE$_{\delta_{n}}$,is investigated in infinite-order autoregressive (AR($\infty$)) models. Instead of accumulating squares of sequential prediction errors from the beginning, APE$_{\delta_{n}}$ is obtained by summing these squared errors from sta...
Article
Full-text available
We consider the problem of choosing the optimal (in the sense of mean-squared prediction error) multistep predictor for an autoregressive (AR) process of finite but unknown order. If a working AR model (which is possibly misspecified) is adopted for multistep predictions, then two competing types of multistep predictors (i.e., plug-in and direct pr...
Article
We investigate the predictive ability of the accumulated prediction error (APE) of Rissanen in an infinite-order autoregressive (AR(#)) model. Since there are infinitely many parameters in the model, all finite-order AR models are misspecified. We first show that APE is asymptotically equivalent to Bayesian information criterion (BIC) and is not as...
Article
Zhang and Shaman considered the problem of estimating the conditional mean-squared prediciton error (CMSPE) for a Gaussian autoregressive (AR) process. They used the final prediction error (FPE) of Akaike to estimate CMSPE and proposed that FPE's effectiveness be judged by its asymptotic correlation with CMSPE. However, as pointed out by Kabaila an...
Article
Let observations come from an infinite-order autoregressive (AR) process. For predicting the future of the observed time series (referred to as the same-realization prediction), we use the least-squares predictor obtained by fitting a finite-order AR model. We also allow the order to become infinite as the number of observations does in order to ob...
Article
Full-text available
In this paper, two competing types of multistep predictors, i.e., plug-in and direct predictors, are considered in autoregressive (AR) processes. When a working model AR(k) is used for the h-step prediction with h 1, the plug-in predictor is obtained from repeatedly using the fitted (by least squares) AR(k) model with an unknown future value replac...
Article
An asymptotic expression for the mean-squared prediction error (MSPE) of the least squares predictor is obtained in the random walk model. It is shown that the term of order 1/n in this error, where n is the sample size, is twice as large as the one obtained from the first-order autoregressive (AR(1)) model satisfying the stationary assumption. Mor...
Article
Full-text available
We investigate the predictive ability of the accumulated prediction error (APE) of Rissanen in an infinite-order autoregressive (AR(∞)) model. Since there are infinitely many parameters in the model, all finite-order AR models are misspecified. We first show that APE is asymptotically equivalent to Bayesian information criterion (BIC) and is not as...
Article
Full-text available
This paper considers the order (predictor) selection problem in autoregressive (AR) processes. When the underlying AR model is known to be stationary and of infinite order, Shibata (1980) and Ing and Wei (2005) showed that AIC is asymptotically efficient for independent-and same-realization predictions, respectively. Ing (2007) recently removed the...
Article
Full-text available
This paper investigates multistep prediction errors for nonstationary au- toregressive processes with both model order and true parameters unknown. We give asymptotic expressions for the multistep mean squared prediction errors and accumulated prediction errors of two important methods, plug-in and direct prediction. These expres- sions not only ch...

Network

Cited By