no code implementations • 19 Oct 2023 • Ashok Vardhan Makkuva, Marco Bondaschi, Thijs Vogels, Martin Jaggi, Hyeji Kim, Michael C. Gastpar
On the latter, we obtain $50$-$64 \%$ improvement in perplexity over our baselines for noisy channels.
1 code implementation • 25 Sep 2023 • Vinitra Swamy, Malika Satayeva, Jibril Frej, Thierry Bossy, Thijs Vogels, Martin Jaggi, Tanja Käser, Mary-Anne Hartley
Predicting multiple real-world tasks in a single model often requires a particularly diverse feature space.
1 code implementation • 5 Jan 2023 • Thijs Vogels, Hadrien Hendrikx, Martin Jaggi
This paper aims to paint an accurate picture of sparsely-connected distributed optimization.
1 code implementation • 12 Nov 2022 • Cécile Trottet, Thijs Vogels, Martin Jaggi, Mary-Anne Hartley
Data-driven Clinical Decision Support Systems (CDSS) have the potential to improve and standardise care with personalised probabilistic guidance.
1 code implementation • 7 Jun 2022 • Thijs Vogels, Hadrien Hendrikx, Martin Jaggi
In data-parallel optimization of machine learning models, workers collaborate to improve their estimates of the model: more accurate gradients allow them to use larger learning rates and optimize faster.
1 code implementation • NeurIPS 2021 • Thijs Vogels, Lie He, Anastasia Koloskova, Tao Lin, Sai Praneeth Karimireddy, Sebastian U. Stich, Martin Jaggi
A key challenge, primarily in decentralized deep learning, remains the handling of differences between the workers' local data distributions.
1 code implementation • NeurIPS 2020 • Thijs Vogels, Sai Praneeth Karimireddy, Martin Jaggi
Lossy gradient compression has become a practical tool to overcome the communication bottleneck in centrally coordinated distributed training of machine learning models.
2 code implementations • 4 Aug 2020 • Thijs Vogels, Sai Praneeth Karimireddy, Martin Jaggi
Lossy gradient compression has become a practical tool to overcome the communication bottleneck in centrally coordinated distributed training of machine learning models.
no code implementations • ICML 2020 • Prabhu Teja Sivaprasad, Florian Mai, Thijs Vogels, Martin Jaggi, François Fleuret
The performance of optimizers, particularly in deep learning, depends considerably on their chosen hyperparameter configuration.
no code implementations • 25 Sep 2019 • Prabhu Teja S*, Florian Mai*, Thijs Vogels, Martin Jaggi, Francois Fleuret
There is no consensus yet on the question whether adaptive gradient methods like Adam are easier to use than non-adaptive optimization methods like SGD.
1 code implementation • NeurIPS 2019 • Thijs Vogels, Sai Praneeth Karimireddy, Martin Jaggi
We study gradient compression methods to alleviate the communication bottleneck in data-parallel distributed optimization.
no code implementations • ACM Transactions on Graphics 2017 • Steve Bako, Thijs Vogels, Brian McWilliams, Mark Meyer, Jan Novák, Alex Harvill, Pradeep Sen, Tony Derose, Fabrice Rousselle
In a second approach, we introduce a novel, kernel-prediction network which uses the CNN to estimate the local weighting kernels used to compute each denoised pixel from its neighbors.