1 code implementation • 30 Apr 2019 • Charles Weill, Javier Gonzalvo, Vitaly Kuznetsov, Scott Yang, Scott Yak, Hanna Mazzawi, Eugen Hotaj, Ghassen Jerfel, Vladimir Macko, Ben Adlam, Mehryar Mohri, Corinna Cortes
AdaNet is a lightweight TensorFlow-based (Abadi et al., 2015) framework for automatically learning high-quality ensembles with minimal expert intervention.
no code implementations • NeurIPS 2018 • Corinna Cortes, Vitaly Kuznetsov, Mehryar Mohri, Dmitry Storcheus, Scott Yang
In this paper, we design efficient gradient computation algorithms for two broad families of structured prediction loss functions: rational and tropical losses.
no code implementations • 9 May 2018 • Vitaly Kuznetsov, Zelda Mariet
The availability of large amounts of time series data, paired with the performance of deep-learning algorithms on a broad class of problems, has recently led to significant interest in the use of sequence-to-sequence models for time series forecasting.
no code implementations • 18 Apr 2018 • Corinna Cortes, Vitaly Kuznetsov, Mehryar Mohri, Holakou Rahmanian, Manfred K. Warmuth
We study the problem of online path learning with non-additive gains, which is a central problem appearing in several applications, including ensemble structured prediction.
no code implementations • 15 Mar 2018 • Vitaly Kuznetsov, Mehryar Mohri
We present data-dependent learning bounds for the general scenario of non-stationary non-mixing stochastic processes.
no code implementations • NeurIPS 2017 • Vitaly Kuznetsov, Mehryar Mohri
In this paper, we introduce and analyze Discriminative State-Space Models for forecasting non-stationary time series.
no code implementations • 29 Oct 2017 • Corinna Cortes, Giulia Desalvo, Vitaly Kuznetsov, Mehryar Mohri, Scott Yang
We show that the notion of discrepancy can be used to design very general algorithms and a unified framework for the analysis of multi-armed rested bandit problems with non-stationary rewards.
2 code implementations • ICML 2017 • Corinna Cortes, Xavi Gonzalvo, Vitaly Kuznetsov, Mehryar Mohri, Scott Yang
We present new algorithms for adaptively learning artificial neural networks.
no code implementations • NeurIPS 2016 • Corinna Cortes, Mehryar Mohri, Vitaly Kuznetsov, Scott Yang
We give new data-dependent margin guarantees for structured prediction for a very wide family of loss functions and a general family of hypotheses, with an arbitrary factor graph decomposition.
no code implementations • NeurIPS 2015 • Vitaly Kuznetsov, Mehryar Mohri
We present data-dependent learning bounds for the general scenario of non-stationary non-mixing stochastic processes.
no code implementations • 14 Sep 2015 • Corinna Cortes, Prasoon Goyal, Vitaly Kuznetsov, Mehryar Mohri
This paper presents an algorithm, Voted Kernel Regularization , that provides the flexibility of using potentially very complex kernel functions such as predictors based on much higher-degree polynomial kernels, while benefitting from strong learning guarantees.
no code implementations • NeurIPS 2014 • Vitaly Kuznetsov, Mehryar Mohri, Umar Syed
We give new data-dependent learning bounds for convex ensembles in the multi-class classification setting expressed in terms of the Rademacher complexities of the sub-families composing the base classifier set, and the mixture weight assigned to each sub-family.