Search Results for author: Gábor Lugosi

Found 17 papers, 1 papers with code

Estimating the history of a random recursive tree

no code implementations14 Mar 2024 Simon Briend, Christophe Giraud, Gábor Lugosi, Déborah Sulem

This paper studies the problem of estimating the order of arrival of the vertices in a random recursive tree.

On the quality of randomized approximations of Tukey's depth

no code implementations11 Sep 2023 Simon Briend, Gábor Lugosi, Roberto Imbuzeiro Oliveira

Tukey's depth (or halfspace depth) is a widely used measure of centrality for multivariate data.

Online-to-PAC Conversions: Generalization Bounds via Regret Analysis

no code implementations31 May 2023 Gábor Lugosi, Gergely Neu

We establish a connection between the online and statistical learning setting by showing that the existence of an online learning algorithm with bounded regret in this game implies a bound on the generalization error of the statistical learning algorithm, up to a martingale concentration term that is independent of the complexity of the statistical learning method.

Generalization Bounds

Archaeology of random recursive dags and Cooper-Frieze random networks

no code implementations29 Jul 2022 Simon Briend, Francisco Calvillo, Gábor Lugosi

We study the problem of finding the root vertex in large growing networks.

Generalization Bounds via Convex Analysis

no code implementations10 Feb 2022 Gábor Lugosi, Gergely Neu

Since the celebrated works of Russo and Zou (2016, 2019) and Xu and Raginsky (2017), it has been well known that the generalization error of supervised learning algorithms can be bounded in terms of the mutual information between their input and the output, given that the loss of any fixed hypothesis has a subgaussian tail.

Generalization Bounds

Bandit problems with fidelity rewards

no code implementations25 Nov 2021 Gábor Lugosi, Ciara Pike-Burke, Pierre-André Savalle

The fidelity bandits problem is a variant of the $K$-armed bandit problem in which the reward of each arm is augmented by a fidelity reward that provides the player with an additional payoff depending on how 'loyal' the player has been to that arm in the past.

Learning to maximize global influence from local observations

no code implementations24 Sep 2021 Gábor Lugosi, Gergely Neu, Julia Olkhovskaya

The goal of the decision maker is to select the sequence of agents in a way that the total number of influenced nodes in the network.

Benign Overfitting in Linear Regression

no code implementations26 Jun 2019 Peter L. Bartlett, Philip M. Long, Gábor Lugosi, Alexander Tsigler

Motivated by this phenomenon, we consider when a perfect fit to training data in linear regression is compatible with accurate prediction.

regression

Learning partial correlation graphs and graphical models by covariance queries

no code implementations22 Jun 2019 Gábor Lugosi, Jakub Truszkowski, Vasiliki Velona, Piotr Zwiernik

We study the problem of recovering the structure underlying large Gaussian graphical models or, more generally, partial correlation graphs.

Online Influence Maximization with Local Observations

no code implementations28 May 2018 Julia Olkhovskaya, Gergely Neu, Gábor Lugosi

We consider an online influence maximization problem in which a decision maker selects a node among a large number of possibilities and places a piece of information at the node.

On the Hardness of Inventory Management with Censored Demand Data

no code implementations16 Oct 2017 Gábor Lugosi, Mihalis G. Markakis, Gergely Neu

Furthermore, we modify the proposed policy in order to perform well in terms of the tracking regret, that is, using as benchmark the best sequence of inventory decisions that switches a limited number of times.

Management

Boltzmann Exploration Done Right

no code implementations NeurIPS 2017 Nicolò Cesa-Bianchi, Claudio Gentile, Gábor Lugosi, Gergely Neu

Boltzmann exploration is a classic strategy for sequential decision-making under uncertainty, and is one of the most standard tools in Reinforcement Learning (RL).

Decision Making Decision Making Under Uncertainty +2

Algorithmic stability and hypothesis complexity

no code implementations ICML 2017 Tongliang Liu, Gábor Lugosi, Gergely Neu, DaCheng Tao

The bounds are based on martingale inequalities in the Banach space to which the hypotheses belong.

Sub-Gaussian estimators of the mean of a random vector

no code implementations1 Feb 2017 Gábor Lugosi, Shahar Mendelson

We study the problem of estimating the mean of a random vector $X$ given a sample of $N$ independent, identically distributed points.

Regularization, sparse recovery, and median-of-means tournaments

no code implementations15 Jan 2017 Gábor Lugosi, Shahar Mendelson

A regularized risk minimization procedure for regression function estimation is introduced that achieves near optimal accuracy and confidence under general conditions, including heavy-tailed predictor and response variables.

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.