1 code implementation • 20 Jun 2023 • David Fleischer, David A. Stephens, Archer Yang
Generalized random forests arXiv:1610. 01271 build upon the well-established success of conventional forests (Breiman, 2001) to offer a flexible and powerful non-parametric method for estimating local solutions of heterogeneous estimating equations.
no code implementations • 23 Mar 2021 • Ayoub El Hanchi, David A. Stephens
Despite the strong theoretical guarantees that variance-reduced finite-sum optimization algorithms enjoy, their applicability remains limited to cases where the memory overhead they introduce (SAG/SAGA), or the periodic full gradient computation they require (SVRG/SARAH) are manageable.
no code implementations • NeurIPS 2020 • Ayoub El Hanchi, David A. Stephens
Reducing the variance of the gradient estimator is known to improve the convergence rate of stochastic gradient-based optimization and sampling algorithms.
1 code implementation • 26 Jul 2017 • Mohamad Elmasri, Maxwell Farrell, David A. Stephens
As many interaction networks are constructed from presence-only data, we extend the model by integrating a correction mechanism for missing interactions, which proves valuable in reducing uncertainty in unobserved interactions.
Applications Populations and Evolution
1 code implementation • 15 Jan 2017 • Olli Saarela, Léo R. Belzile, David A. Stephens
In causal inference confounding may be controlled either through regression adjustment in an outcome model, or through propensity score adjustment or inverse probability of treatment weighting, or both.
Methodology 62F15