David Beaudoin’s research while affiliated with University of Quebec and other places
What is this page?
This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.
In this paper, we introduce a new ranking system where the data are preferences resulting from paired comparisons. When direct preferences are missing or unclear, then preferences are determined through indirect comparisons. Given that a ranking of n subjects implies (2n) paired preferences, the resultant computational problem is the determination of an optimal ranking where the agreement between the implied preferences via the ranking and the data preferences is maximized. Comparisons are carried out via simulation studies where the proposed rankings outperform Bradley–Terry in a particular predictive comparison.
The main objective of this paper is to investigate the extent to which the margin of victory can be predicted solely by the rankings of the opposing teams in NCAA Division I men's basketball games. Several past studies have modeled this relationship for the games played during the March Madness tournament, and this work aims at verifying if the models advocated in these papers still perform well for regular season games. Indeed, most previous articles have shown that a simple quadratic regression model provides fairly accurate predictions of the margin of victory when team rankings only range from 1 to 16. Does that still hold true when team rankings can go as high as 351? Do the model assumptions hold? Can we find semi- or non-parametric methods that yield even better results (i.e. predicted margins of victory that more closely resemble actual results)? The analyses presented in this paper suggest that the answer is "yes" on all three counts!
The main objective of this paper is to investigate the extent to which the margin of victory can be predicted solely by the rankings of the opposing teams in NCAA Division I men's basketball games. Several past studies have modeled this relationship for the games played during the March Madness tournament, and this work aims at verifying if the models advocated in these papers still perform well for regular season games. Indeed, most previous articles have shown that a simple quadratic regression model provides fairly accurate predictions of the margin of victory when team rankings only range from 1 to 16. Does that still hold true when team rankings can go as high as 351? Do the model assumptions hold? Can we find semi- or non-parametric methods that yield even better results (i.e. predicted margins of victory that more closely resemble actual results)? The analyses presented in this paper suggest that the answer is "yes" on all three counts!
This paper develops a simulator for matches in the National Hockey League with the intent of assessing strategies for pulling the goaltender. Aspects of the approach that are novel include breaking the game down into ner and more realistic situations, introducing the eect of penalties and including the home-ice advantage. Parameter estimates used in the simulator are obtained through the analysis of an extensive data set using constrained Bayesian estimation via Markov chain methods. Some surprising strategies are obtained which do not appear to be used in practice.
Many proposals have been made recently for goodness-of-fit testing of copula models. After reviewing them briefly, the authors concentrate on "blanket tests", i.e., those whose implementation requires neither an arbitrary categorization of the data nor any strategic choice of smoothing parameter, weight function, kernel, window, etc. The authors present a critical review of these procedures and suggest new ones. They describe and interpret the results of a large Monte Carlo experiment designed to assess the effect of the sample size and the strength of dependence on the level and power of the blanket tests for various combinations of copula models under the null hypothesis and the alternative. To circumvent problems in the determination of the limiting distribution of the test statistics under composite null hypotheses, they recommend the use of a double parametric bootstrap procedure, whose implementation is detailed. They conclude with a number of practical recommendations.
We investigate the nonparametric estimation of Kendall's coefficient of concordance, τ, for measuring the association between two variables under bivariate censoring. The proposed estimator is a modification of the estimator introduced by Oakes (1982), using a Horvitz-Thompson-type correction for the pairs that are not orderable. With censored data, a pair is orderable if one can establish whether the uncensored pair is discordant or concordant using the data available for that pair. Our estimator is shown to be consistent and asymptotically normally distributed. A simulation study shows that the proposed estimator performs well when compared with competing alternatives. The various methods are illustrated with a real data set.
One-sided truncated survival data arise when a pair of time-to-event variables (X, Y) is observed only when X<Y. Existing methods of analysis rely on the assumption of quasi-independence between X and Y. Recently, Lakhal-Chaieb et al. (Biometrika 2006; 93:655-669) modeled potential dependency between these random variables via a semi-survival Archimedean copula. In this paper, we present a model selection procedure to rank a set of semi-survival Archimedean copula families according to their ability to fit a given data set subject to dependent truncation. The proposed procedure is based on a truncated version of Kendall's tau (J. Multivariate Anal. 1996; 56:60-74). The performance of the proposal is illustrated through simulations and three real data sets.
This paper considers the estimation of Kendall's tau for bivariate data (X,Y) when only Y is subject to right-censoring. Although τ is estimable under weak regularity conditions, the estimators proposed by Brown et al. [1974. Nonparametric tests of independence for censored data, with applications to heart transplant studies. Reliability and Biometry, 327–354], Weier and Basu [1980. An investigation of Kendall's τ modified for censored data with applications. J. Statist. Plann. Inference 4, 381–390] and Oakes [1982. A concordance test for independence in the presence of censoring. Biometrics 38, 451–455], which are standard in this context, fail to be consistent when τ≠0 because they only use information from the marginal distributions. An exception is the renormalized estimator of Oakes [2006. On consistency of Kendall's tau under censoring. Technical Report, Department of Biostatistics and Computational Biology, University of Rochester, Rochester, NY], whose consistency has been established for all possible values of τ, but only in the context of the gamma frailty model. Wang and Wells [2000. Estimation of Kendall's tau under censoring. Statist. Sinica 10, 1199–1215] were the first to propose an estimator which accounts for joint information. Four more are developed here: the first three extend the methods of Brown et al. [1974. Nonparametric tests of independence for censored data, with applications to heart transplant studies. Reliability and Biometry, 327–354], Weier and Basu [1980, An investigation of Kendall's τ modified for censored data with applications. J. Statist. Plann. Inference 4, 381–390] and Oakes [1982, A concordance test for independence in the presence of censoring. Biometrics 38, 451–455] to account for information provided by X, while the fourth estimator inverts an estimation of Pr(Yi⩽y|Xi=xi,Yi>ci) to get an imputation of the value of Yi censored at Ci=ci. Following Lim [2006. Permutation procedures with censored data. Comput. Statist. Data Anal. 50, 332–345], a nonparametric estimator is also considered which averages the obtained from a large number of possible configurations of the observed data (X1,Z1),…,(Xn,Zn), where Zi=min(Yi,Ci). Simulations are presented which compare these various estimators of Kendall's tau. An illustration involving the well-known Stanford heart transplant data is also presented.
Abstract This paper concerns the search for optimal or nearly optimal batting orders in one-day cricket. A search is conducted over the space of permutations of batting orders where simulated annealing is used to explore the space. A non-standard aspect of the optimization is that the objective function (which is the mean,number,of runs per innings) is unavailable and is approximated,via simulation. The simulation component,generates runs ball by ball during an innings taking into account the state of the match and estimated characteristics of individual batsmen. The methods developed in the paper are applied to the national team of India based on their performance,in one-day international cricket matches. 2004 Elsevier Ltd. All rights reserved. Keywords: Log-linear models; Markov chain methods; Monte Carlo simulation; Simulated annealing; WinBUGS software
... The research circles involve in methodological advancements ritualistically facilitate the attainment of above documented delicacies through the application of paired comparison (PC) models or choice models. For example, Cattelan et al. (2013), demonstrated the applicability of PC mechanism to assess the outcomes of sports events while allowing the time-varying ability (Beaudoin & Swartz, 2018). Similarly, Johnson et al. (2019) elucidated the applicability of PC models in public health administration while facilitating the arduous task of project prioritization. ...
... Silva and Swartz (2016) investigated the problem of optimal substitution times in soccer, and as a by-product of their analysis, found that teams that were leading in a soccer match were more likely to have the next goal scored against them than if the match had been tied. In the National Hockey League (NHL), Figure 2 from Beaudoin, Schulte and Swartz (2016) indicates that the probability of shots on goal by the home team increases as the goal differential in favour of the road team increases. This finding was corroborated by Thomas (2017) who showed that there is an increased probability for tied matches than would be expected by independent Poisson scoring models. ...
... The hope is to score a quick goal to get back in the game, but the risk is falling further behind. Beaudoin and Swartz (2010) show that NHL coaches do not always employ the optimal strategies, usually by waiting too long to pull their goalies. Skinner (2011) develops a general framework for these desperation strategies, which include the onside kick in American football, pulling the infield and/or outfield in baseball, and of course, the fabled Hack-a-Shaq strategy in basketball. ...
... Many estimators of Kendall's τ , with right-censored data, have been proposed and studied by some statisticians. See Lim and Meier (2006); Beaudoin et al. (2007); Wang and Wells (2000); Weier and Basu (1980); Oakes (1982Oakes ( , 2008; Lakhal et al. (2008). Hesieh and Li (2017) studied the estimation of bivariate, left-truncated variables. ...
... The purpose of each individual batsman is to accumulate runs while simultaneously defending their wicket [36]. Players score runs by striking the ball and running to the opposite end of the pitch before any fielder dislodges the bails of the wickets [37]. Likewise, players obtain runs by striking the ball beyond the boundary of the field [37]. ...
... The parameters and 95% uncertainty range of each copula model are estimated via Markov Chain Monte Carlo (MCMC) simulation within a Bayesian framework. The best bivariate copula model is selected considering Root Mean Square Error (RMSE), Nash-Sutcliff efficiency (NSE), and underlying uncertainties (Genest et al., 2009). ...
... Kendall's tau estimator represented by Equations (9), (11) and (13) is valid if all realizations of the two-dimensional variable are uncensored data. Otherwise, the estimator can be used [49], [53], [54]: ...
... Much has been written about the limitations of different classes of copulas. We may mention, for instance: 1) when working with an Archimedean copula with generator function φ, when conditioning on realizations, the conditional distribution only depend on the sum of inverse values of the generator function of the realizations, thus, for data showing dependence behaviour close to that of the Archimedean class may present problems with the dependence of the parameters of conditional copulas on this joint probability (Stöber, Joe, and Czado 2013); 2) when using an Archimedean copula to high dimension applications it is inevitably to make some simplifying assumptions (Oh and Patton 2018); 3) when response variables are binary, modelling with an Archimedean copula, the associations become more challenging due to the stringent constraints imposed on the dependence parameters (Deng and Chaganty 2017); 4) Archimedean copula families do not have ability to fit a given data set subject to dependent truncation (Beaudoin and Lakhal-Chaieb 2008); 5) elliptical copulas do not have closed form expressions and are restricted to have radial symmetry (Embrechts, Lindskog, and McNeil 2003); 6) the copula family generated by the sub-Gaussian αstable distribution is unable to cover the size of tail dependence observed in different kinds of data (Frahm, Junker, and Szimayer 2003). All of the previous limitations do not apply in the present research. ...