no code implementations • 1 Apr 2024 • Tsuyoshi Idé, Dzung T. Phan, Rudy Raymond
This paper presents two methodological advancements in decentralized multi-task learning under privacy constraints, aiming to pave the way for future developments in next-generation Blockchain platforms.
1 code implementation • NeurIPS 2021 • Tsuyoshi Idé, Georgios Kollias, Dzung T. Phan, Naoki Abe
In this paper, we propose a mathematically well-defined sparse causal learning framework based on a cardinality-regularized Hawkes process, which remedies the pathological issues of existing approaches.
1 code implementation • NeurIPS 2021 • Hoang Thanh Lam, Gabriele Picco, Yufang Hou, Young-suk Lee, Lam M. Nguyen, Dzung T. Phan, Vanessa López, Ramon Fernandez Astudillo
In many machine learning tasks, models are trained to predict structure data such as graphs.
Ranked #2 on
AMR Parsing
on LDC2020T02
(using extra training data)
1 code implementation • 5 Mar 2021 • Quoc Tran-Dinh, Nhan H. Pham, Dzung T. Phan, Lam M. Nguyen
These new algorithms can handle statistical and system heterogeneity, which are the two main challenges in federated learning, while achieving the best known communication complexity.
no code implementations • NeurIPS 2020 • Haoran Zhu, Pavankumar Murali, Dzung T. Phan, Lam M. Nguyen, Jayant R. Kalagnanam
Several recent publications report advances in training optimal decision trees (ODT) using mixed-integer programs (MIP), due to algorithmic advances in integer programming and a growing interest in addressing the inherent suboptimality of heuristic approaches such as CART.
1 code implementation • 1 Mar 2020 • Nhan H. Pham, Lam M. Nguyen, Dzung T. Phan, Phuong Ha Nguyen, Marten van Dijk, Quoc Tran-Dinh
We propose a novel hybrid stochastic policy gradient estimator by combining an unbiased policy gradient estimator, the REINFORCE estimator, with another biased one, an adapted SARAH estimator for policy optimization.
no code implementations • 19 Feb 2020 • Lam M. Nguyen, Quoc Tran-Dinh, Dzung T. Phan, Phuong Ha Nguyen, Marten van Dijk
We also study uniformly randomized shuffling variants with different learning rates and model assumptions.
no code implementations • 8 Jul 2019 • Quoc Tran-Dinh, Nhan H. Pham, Dzung T. Phan, Lam M. Nguyen
We introduce a new approach to develop stochastic optimization algorithms for a class of stochastic composite and possibly nonconvex optimization problems.
no code implementations • 15 May 2019 • Quoc Tran-Dinh, Nhan H. Pham, Dzung T. Phan, Lam M. Nguyen
We introduce a hybrid stochastic estimator to design stochastic gradient algorithms for solving stochastic optimization problems.
1 code implementation • 15 Feb 2019 • Nhan H. Pham, Lam M. Nguyen, Dzung T. Phan, Quoc Tran-Dinh
We also specify the algorithm to the non-composite case that covers existing state-of-the-arts in terms of complexity bounds.
no code implementations • 22 Jan 2019 • Lam M. Nguyen, Phuong Ha Nguyen, Dzung T. Phan, Jayant R. Kalagnanam, Marten van Dijk
This paper has some inconsistent results, i. e., we made some failed claims because we did some mistakes for using the test criterion for a series.
no code implementations • 22 Jan 2019 • Lam M. Nguyen, Marten van Dijk, Dzung T. Phan, Phuong Ha Nguyen, Tsui-Wei Weng, Jayant R. Kalagnanam
The total complexity (measured as the total number of gradient computations) of a stochastic first-order optimization algorithm that finds a first-order stationary point of a finite-sum smooth nonconvex objective function $F(w)=\frac{1}{n} \sum_{i=1}^n f_i(w)$ has been proven to be at least $\Omega(\sqrt{n}/\epsilon)$ for $n \leq \mathcal{O}(\epsilon^{-2})$ where $\epsilon$ denotes the attained accuracy $\mathbb{E}[ \|\nabla F(\tilde{w})\|^2] \leq \epsilon$ for the outputted approximation $\tilde{w}$ (Fang et al., 2018).
no code implementations • 9 Oct 2018 • Marten van Dijk, Lam M. Nguyen, Phuong Ha Nguyen, Dzung T. Phan
We study Stochastic Gradient Descent (SGD) with diminishing step sizes for convex objective functions.
no code implementations • 18 Jan 2018 • Lam M. Nguyen, Nam H. Nguyen, Dzung T. Phan, Jayant R. Kalagnanam, Katya Scheinberg
In this paper, we consider a general stochastic optimization problem which is often at the core of supervised learning, such as deep learning and linear classification.