Cancer patients face competing risks of cancer death and CVD. Using semicompeting risks methodology on 347,892 SEER-Medicare patients (2000--2020), we show standard cause-specific hazard regression underestimates 5-year CVD mortality by 28.
Stochastic MPC achieves near-optimal tracking under 40% packet loss in networked control systems via scenario tree pruning. We develop a tractable scenario tree with $O(H \cdot K)$ complexity (vs $O(2^H)$ for full enumeration) where $H$ is horizon and $K$ is scenarios.
Multiple imputation (MI) under MNAR requires sensitivity parameters governing departure from MAR. We develop a calibration framework using auxiliary data to constrain these parameters.
Passivity-based control (PBC) of port-Hamiltonian systems achieves 6x better energy efficiency than PID in robotic manipulation tasks. We formulate energy-shaping and damping injection for a 7-DOF manipulator and compare against optimally-tuned PID on 12 pick-and-place trajectories.
Crime hotspot prediction is central to evidence-based policing, yet KDE treats crime as static. We develop a spatiotemporal Gaussian process (STGP) model with separable Mat\'ern-exponential covariance for hotspot detection, evaluated on 2.
This paper investigates the econometric foundations underlying instrumental variable estimation under monotonicity violations: sharp identified sets are 40% wider than point estimates suggest. Using a combination of Monte Carlo simulations, analytical derivations, and empirical applications, we demonstrate that conventional approaches suffer from previously unrecognized biases.
Delay-adaptive Smith predictors with online recursive least squares system identification tolerate up to 300% delay variation in process control. Standard Smith predictors become unstable beyond 20% delay mismatch.
Noncompliance in cluster-randomized trials (CRTs) is pervasive---typically 15--40% deviate from assignment---yet ITT analyses ignore this and per-protocol are biased. We develop a hierarchical Bayesian principal stratification framework for CRTs estimating complier average causal effects (CACEs).
Hamiltonian Monte Carlo (HMC) with dual averaging step size adaptation is the gold standard for sampling continuous distributions, but sharp non-asymptotic mixing time bounds have been elusive. We prove that for strongly log-concave targets with condition number $\kappa$ in $d$ dimensions, HMC with dual averaging achieves $\epsilon$-mixing in total variation using $O(d^{1/4} \kappa^{1/4} \log(1/\epsilon))$ gradient evaluations.
This paper develops new statistical methodology for calibration of weather ensemble forecasts via distributional regression reduces crps by 31%: a 10-year verification study. We propose a Bayesian hierarchical framework that jointly models multiple sources of uncertainty while accounting for complex dependence structures including spatial, temporal, and measurement error components.
Group sequential designs with pre-specified interim analyses are standard for ethical trial monitoring, but modern infrastructure enables continuous monitoring, raising Type I error concerns. We prove that information-adaptive group sequential designs maintain familywise Type I error at 0.
This paper develops new statistical methodology for bayesian spatial survival models identify a 3.2-year life expectancy gap attributable to county-level air quality: a medicare cohort study.
Adaptive enrichment designs allow clinical trials to restrict enrollment to a promising subpopulation at interim analysis. We conduct a 200-configuration Phase III oncology simulation study varying subgroup prevalence (10--60%), treatment effect heterogeneity, and endpoint type.
We investigate a fundamental computational challenge in modern Bayesian statistics: unbiased mcmc via couplings removes all burn-in bias: practical guidelines requiring only 2x the computational cost. Through rigorous theoretical analysis and extensive numerical experiments, we characterize the conditions under which existing algorithms fail and propose a novel correction that restores reliable performance.
Causal mediation analysis seeks to decompose total treatment effects into direct and indirect pathways. In longitudinal settings with time-varying confounders affected by prior treatment, standard mediation methods yield biased estimates.
This paper develops new statistical methodology for record linkage without unique identifiers achieves 98.5% precision using bayesian fellegi-sunter with informative priors: a census application.
Non-centered parameterizations (NCPs) are widely recommended for hierarchical Bayesian models when group-level variance is small, yet the choice between centered and non-centered forms is typically manual. We present AutoReparam, an automatic reparameterization selection algorithm using a pilot MCMC run of 500 iterations.
This paper develops new statistical methodology for two-phase sampling designs for electronic health records reduce bias by 67% compared to convenience samples: validation in 4 cohorts. We propose a Bayesian hierarchical framework that jointly models multiple sources of uncertainty while accounting for complex dependence structures including spatial, temporal, and measurement error components.
Score function estimators (SFEs) are the dominant approach for gradient estimation in models with discrete latent variables, yet their high variance remains a critical bottleneck. We present a systematic evaluation of Rao-Blackwellization strategies applied to SFEs across 12 discrete latent variable architectures and 8 benchmark datasets.