no code implementations • 27 Oct 2022 • Juan Cervino, Juan Andres Bazerque, Miguel Calvo-Fullana, Alejandro Ribeiro
In this paper we draw intuition from the two extreme learning scenarios -- a single function for all tasks, and a task-specific function that ignores the other tasks dependencies -- to propose a bias-variance trade-off.
no code implementations • 24 Oct 2020 • Juan Cervino, Juan Andres Bazerque, Miguel Calvo-Fullana, Alejandro Ribeiro
In this paper we consider a problem known as multi-task learning, consisting of fitting a set of classifier or regression functions intended for solving different tasks.
no code implementations • 16 Oct 2020 • Santiago Paternain, Juan Andres Bazerque, Alejandro Ribeiro
To that end we compute unbiased stochastic gradients of the value function which we use as ascent directions to update the policy.