Search Results for author: Hong Jun Jeon

Found 7 papers, 0 papers with code

An Information-Theoretic Analysis of In-Context Learning

no code implementations28 Jan 2024 Hong Jun Jeon, Jason D. Lee, Qi Lei, Benjamin Van Roy

Previous theoretical results pertaining to meta-learning on sequences build on contrived assumptions and are somewhat convoluted.

In-Context Learning Meta-Learning

Adaptive Crowdsourcing Via Self-Supervised Learning

no code implementations24 Jan 2024 Anmol Kagrecha, Henrik Marklund, Benjamin Van Roy, Hong Jun Jeon, Richard Zeckhauser

Common crowdsourcing systems average estimates of a latent quantity of interest provided by many crowdworkers to produce a group estimate.

Self-Supervised Learning

Continual Learning as Computationally Constrained Reinforcement Learning

no code implementations10 Jul 2023 Saurabh Kumar, Henrik Marklund, Ashish Rao, Yifan Zhu, Hong Jun Jeon, Yueyang Liu, Benjamin Van Roy

The design of such agents, which remains a long-standing challenge of artificial intelligence, is addressed by the subject of continual learning.

Continual Learning reinforcement-learning

An Information-Theoretic Analysis of Compute-Optimal Neural Scaling Laws

no code implementations2 Dec 2022 Hong Jun Jeon, Benjamin Van Roy

For a particular learning model inspired by barron 1993, we establish an upper bound on the minimal information-theoretically achievable expected error as a function of model and data set sizes.

Language Modelling

Is Stochastic Gradient Descent Near Optimal?

no code implementations18 Sep 2022 Yifan Zhu, Hong Jun Jeon, Benjamin Van Roy

However, existing computational theory suggests that, even for single-hidden-layer teacher networks, to attain small error for all such teacher networks, the computation required to achieve this sample complexity is intractable.

An Information-Theoretic Framework for Supervised Learning

no code implementations1 Mar 2022 Hong Jun Jeon, Yifan Zhu, Benjamin Van Roy

For a particular prior distribution on weights, we establish sample complexity bounds that are simultaneously width independent and linear in depth.

Reward-rational (implicit) choice: A unifying formalism for reward learning

no code implementations NeurIPS 2020 Hong Jun Jeon, Smitha Milli, Anca D. Dragan

It is often difficult to hand-specify what the correct reward function is for a task, so researchers have instead aimed to learn reward functions from human behavior or feedback.

Cannot find the paper you are looking for? You can Submit a new open access paper.