Hasil untuk "stat.ME"

Menampilkan 15 dari ~221788 hasil · dari Semantic Scholar, CrossRef, arXiv

JSON API
S2 Open Access 2026
Implementation and Workflows for INLA-Based Approximate Bayesian Structural Equation Modelling

Haziq Jamil, H. Rue

Bayesian structural equation modelling (BSEM) offers many advantages such as principled uncertainty quantification, small-sample regularisation, and flexible model specification. However, the Markov chain Monte Carlo (MCMC) methods on which it relies are computationally prohibitive for the iterative cycle of specification, criticism, and refinement that careful psychometric practice demands. We present INLAvaan, an R package for fast, approximate Bayesian SEM built around the Integrated Nested Laplace Approximation (INLA) framework for structural equation models developed by Jamil&Rue (2026, arXiv:2603.25690 [stat.ME]). This paper serves as a companion manuscript that describes the architectural decisions and computational strategies underlying the package. Two substantive applications -- a 256-parameter bifactor circumplex model and a multilevel mediation model with full-information missing-data handling -- demonstrate the approach on specifications where MCMC would require hours of run time and careful convergence work. In constrast, INLAvaan delivers calibrated posterior summaries in seconds.

en Mathematics
arXiv Open Access 2025
Comment on García-Donato et al. (2025) "Model uncertainty and missing data: An objective Bayesian perspective"

Joris Mulder

Garcia-Donato et al. (2025) present a methodology for handling missing data in a model selection problem using an objective Bayesian approach. The current comment discusses an alternative, existing objective Bayesian method for this problem. First, rather than using the g prior, O'Hagan's fractional Bayes factor (O'Hagan, 1995) is utilized based on a minimal fraction. Second, and more importantly due to the focus on missing data, Rubin's rules for multiple imputation can directly be used as the fractional Bayes factor can be written as a Savage-Dickey density ratio for a variable selection problem. The current comment derives the methodology for a variable selection problem. Moreover, its implied behavior is illustrated in a numerical experiment, showing competitive results as the method of Garcia-Donato et al. (2025).

arXiv Open Access 2025
An exploration of sequential Bayesian variable selection -- A comment on García-Donato et al. (2025). "Model uncertainty and missing data: An objective Bayesian perspective"

Sebastian Arnold, Alexander Ly

Our comment on García-Donato et al. (2025). "Model uncertainty and missing data: An objective Bayesian perspective" explores a further extension of the proposed methodology. Specifically, we consider the sequential setting where (potentially missing) data accumulate over time, with the goal of continuously monitoring statistical evidence, as opposed to assessing it only once data collection terminates. We explore a new variable selection method based on sequential model confidence sets, as proposed by Arnold et al. (2024), and show that it can help stabilise the inference of García-Donato et al. (2025). To be published as "Invited discussion" in Bayesian Analysis.

en stat.ME, math.ST
S2 Open Access 2024
Adaptive Unit Root Inference in Autoregressions using the Lasso Solution Path

Martin C. Arnold, Thilo Reinschlussel

We show that the activation knot of a potentially non-stationary regressor on the adaptive Lasso solution path in autoregressions can be leveraged for selection-free inference about a unit root. The resulting test has asymptotic power against local alternatives in $1/T$ neighbourhoods, unlike post-selection inference methods based on consistent model selection. Exploiting the information enrichment principle devised by Reinschl\"ussel and Arnold arXiv:2402.16580 [stat.ME] to improve the Lasso-based selection of ADF models, we propose a composite statistic and analyse its asymptotic distribution and local power function. Monte Carlo evidence shows that the combined test dominates the comparable post-selection inference methods of Tibshirani et al. [JASA, 2016, 514, 600-620] and may surpass the power of established unit root tests against local alternatives. We apply the new tests to groundwater level time series for Germany and find evidence rejecting stochastic trends to explain observed long-term declines in mean water levels.

1 sitasi en Mathematics
CrossRef Open Access 2024
Documenting and communicating the contributions of embedded statisticians: Show me the value!

Terrie Vasilopoulos, Amy Crisp, Gerard Garvan et al.

Academic research productivity relies upon the contribution of statisticians, who are typically clustered in statistics and biostatistics departments, isolated from clinical researchers. Most academic health centres have created consultation hubs or research incubators to make statisticians available for individual collaboration to support the clinical research enterprise. Additionally, some clinical departments within academic health centres have recognized the value in colocating statisticians within their clinical departments to improve availability for collaboration with physicians/researchers. Embedded statisticians encounter the same challenges of isolated statisticians regarding professional support and networking, mentorship and clear role expectations. While for all collaborative statisticians, it is important to effectively communicate value to both collaborators and supervisors, this may be especially problematic for embedded statisticians in clinical departments where their supervisors may not have backgrounds in research or statistics. Previous papers have reported valuable metrics for statisticians, particularly those associated with Biostatistics, Epidemiology and Research Design Cores. There is a knowledge gap regarding metrics tailored to meet the needs of the embedded statistician and clinical supervisors. This paper is a first step towards addressing this important need.In this paper, we explore (1) the critical role of collaborative statisticians and the benefits and challenges of the embedded statistician model, (2) the need for additional metrics specific to embedded statisticians which measure value and (3) how to design a value report. We offer a framework for evaluation of the contributions of the embedded statistician with the following domains: (1) collaboration, (2) research output/productivity, (3) mentoring and (4) education. Metrics that are particularly specific to embedded statisticians and that are not routinely captured include time from project initiation to completion/outcome, time from initial statistical consultation to statistical outcome completion and summary of level of contribution for manuscripts and presentations in addition to author order. We conclude with thoughts on future directions for development of metrics and reporting systems for statisticians embedded in clinical departments.

S2 Open Access 2023
On Convergence Rate of the Generalized Diversity Subsampling Method

Bo Shang

arXiv:2206.10812v1 [stat.ME] proposes a useful algorithm, named generalized Diversity Subsampling (g-DS) algorithm, to select a subsample following some target probability distribution from a finite data set and demonstrates its effectiveness numerically. While the asymptotic performances of g-DS when the true data distribution is known was discussed in arXiv:2206.10812v1 [stat.ME], it remains an interesting question how the estimation errors in the density estimation step, which is an unavoidable step to use g-DS in real-world data sets, influences its asymptotic performance. In this paper, we study the pointwise convergence rate of probability density function (p.d.f) the g-DS subsample to the target p.d.f value, as the data set size approaches infinity, under consideration of the pointwise bias and variance of the estimated data p.d.f.

1 sitasi en Mathematics
arXiv Open Access 2023
A Fast Bootstrap Algorithm for Causal Inference with Large Data

Matthew Kosko, Lin Wang, Michele Santacatterina

Estimating causal effects from large experimental and observational data has become increasingly prevalent in both industry and research. The bootstrap is an intuitive and powerful technique used to construct standard errors and confidence intervals of estimators. Its application however can be prohibitively demanding in settings involving large data. In addition, modern causal inference estimators based on machine learning and optimization techniques exacerbate the computational burden of the bootstrap. The bag of little bootstraps has been proposed in non-causal settings for large data but has not yet been applied to evaluate the properties of estimators of causal effects. In this paper, we introduce a new bootstrap algorithm called causal bag of little bootstraps for causal inference with large data. The new algorithm significantly improves the computational efficiency of the traditional bootstrap while providing consistent estimates and desirable confidence interval coverage. We describe its properties, provide practical considerations, and evaluate the performance of the proposed algorithm in terms of bias, coverage of the true 95% confidence intervals, and computational time in a simulation study. We apply it in the evaluation of the effect of hormone therapy on the average time to coronary heart disease using a large observational data set from the Women's Health Initiative.

en stat.ME, stat.AP
arXiv Open Access 2022
Modelling and forecasting patient recruitment in clinical trials with patients' dropout

Vladimir Anisimov, Guillaume Mijoule, Armando Turchetta et al.

This paper focuses on statistical modelling and prediction of patient recruitment in clinical trials accounting for patients dropout. The recruitment model is based on a Poisson-gamma model introduced by Anisimov and Fedorov (2007), where the patients arrive at different centres according to Poisson processes with rates viewed as gamma-distributed random variables. Each patient can drop the study during some screening period. Managing the dropout process is of a major importance but data related to dropout are rarely correctly collected. In this paper, a few models of dropout are proposed. The technique for estimating parameters and predicting the number of recruited patients over time and the recruitment time is developed. Simulation results confirm the applicability of the technique and thus, the necessity to account for patients dropout at the stage of forecasting recruitment in clinical trials.

en stat.ME, stat.OT
S2 Open Access 2020
Contextual Stochastic Block Model: Sharp Thresholds and Contiguity

Chen Lu, S. Sen

We study community detection in the contextual stochastic block model arXiv:1807.09596 [cs.SI], arXiv:1607.02675 [stat.ME]. In arXiv:1807.09596 [cs.SI], the second author studied this problem in the setting of sparse graphs with high-dimensional node-covariates. Using the non-rigorous cavity method from statistical physics, they conjectured the sharp limits for community detection in this setting. Further, the information theoretic threshold was verified, assuming that the average degree of the observed graph is large. It is expected that the conjecture holds as soon as the average degree exceeds one, so that the graph has a giant component. We establish this conjecture, and characterize the sharp threshold for detection and weak recovery.

23 sitasi en Mathematics, Computer Science
CrossRef Open Access 2020
The cure model in perinatal epidemiology

Emil A Stoltenberg, Hedvig ME Nordeng, Eivind Ystrom et al.

In the statistical literature, the class of survival analysis models known as cure models has received much attention in recent years. Cure models seem not, however, to be part of the statistical toolbox of perinatal epidemiologists. In this paper, we demonstrate that in perinatal epidemiological studies where one investigates the relation between a gestational exposure and a condition that can only be ascertained after several years, cure models may provide the correct statistical framework. The reason for this is that the hypotheses being tested often concern an unobservable outcome that, in view of the hypothesis, should be thought of as occurring at birth, even though it is only detectable much later in life. The outcome of interest can therefore be viewed as a censored binary variable. We illustrate our argument with a simple cure model analysis of the possible relation between gestational exposure to paracetamol and attention-deficit hyperactivity disorder, using data from the Norwegian Mother, Father and Child Cohort Study conducted by the Norwegian Institute of Public Health, and information about the attention-deficit hyperactivity disorder diagnoses obtained from the Norwegian Patient Registry.

4 sitasi en
S2 Open Access 2020
agtboost: Adaptive and Automatic Gradient Tree Boosting Computations

Berent AAnund Stromnes Lunde, T. S. Kleppe

agtboost is an R package implementing fast gradient tree boosting computations in a manner similar to other established frameworks such as xgboost and LightGBM, but with significant decreases in computation time and required mathematical and technical knowledge. The package automatically takes care of split/no-split decisions and selects the number of trees in the gradient tree boosting ensemble, i.e., agtboost adapts the complexity of the ensemble automatically to the information in the data. All of this is done during a single training run, which is made possible by utilizing developments in information theory for tree algorithms {\tt arXiv:2008.05926v1 [stat.ME]}. agtboost also comes with a feature importance function that eliminates the common practice of inserting noise features. Further, a useful model validation function performs the Kolmogorov-Smirnov test on the learned distribution.

en Computer Science, Mathematics
arXiv Open Access 2020
Notes on Exact Power Calculations for t Tests and Analysis of Covariance

Yongqiang Tang

Tang derived the exact power formulae for t tests and analysis of covariance (ANCOVA) in superiority, noninferiority and equivalence trials. The power calculation in equivalence trials can be simplified by using Owen's Q function, which is available in standard statistical software. We extend the exact power determination method for ANCOVA to unstratified and stratified multi-arm randomized trials. The method is applied to the design of multi-arm trials and gold standard noninferiority trials.

S2 Open Access 2019
A DECOMPOSITION ANALYSIS OF DIFFUSION OVER A LARGE NETWORK

Kyungchul Song

Diffusion over a network refers to the phenomenon of a change of state of a cross-sectional unit in one period leading to a change of state of its neighbors in the network in the next period. One may estimate or test for diffusion by estimating a cross-sectionally aggregated correlation between neighbors over time from data. However, the estimated diffusion can be misleading if the diffusion is confounded by omitted covariates. This paper focuses on the measure of diffusion proposed by He and Song (2022, Preprint, arXiv:1812.04195v4 [stat.ME]), provides a method of decomposition analysis to measure the role of the covariates on the estimated diffusion, and develops an asymptotic inference procedure for the decomposition analysis in such a situation. This paper also presents results from a Monte Carlo study on the small sample performance of the inference procedure.

1 sitasi en Mathematics

Halaman 1 dari 11090