Search Results for author: Jane Yu

Found 4 papers, 1 papers with code

NaturalReasoning: Reasoning in the Wild with 2.8M Challenging Questions

no code implementations18 Feb 2025 Weizhe Yuan, Jane Yu, Song Jiang, Karthik Padthe, Yang Li, Dong Wang, Ilia Kulikov, Kyunghyun Cho, Yuandong Tian, Jason E Weston, Xian Li

Scaling reasoning capabilities beyond traditional domains such as math and coding is hindered by the lack of diverse and high-quality questions.

Knowledge Distillation Math

LLM Pretraining with Continuous Concepts

no code implementations12 Feb 2025 Jihoon Tack, Jack Lanchantin, Jane Yu, Andrew Cohen, Ilia Kulikov, Janice Lan, Shibo Hao, Yuandong Tian, Jason Weston, Xian Li

We propose Continuous Concept Mixing (CoCoMix), a novel pretraining framework that combines discrete next token prediction with continuous concepts.

Knowledge Distillation Language Modeling +3

Self-Consistency Preference Optimization

no code implementations6 Nov 2024 Archiki Prasad, Weizhe Yuan, Richard Yuanzhe Pang, Jing Xu, Maryam Fazel-Zarandi, Mohit Bansal, Sainbayar Sukhbaatar, Jason Weston, Jane Yu

Self-alignment, whereby models learn to improve themselves without human annotation, is a rapidly growing research area.

GSM8K Math

Quantifying Adaptability in Pre-trained Language Models with 500 Tasks

2 code implementations NAACL 2022 Belinda Z. Li, Jane Yu, Madian Khabsa, Luke Zettlemoyer, Alon Halevy, Jacob Andreas

When a neural language model (LM) is adapted to perform a new task, what aspects of the task predict the eventual performance of the model?

Language Modeling Language Modelling +3

Cannot find the paper you are looking for? You can Submit a new open access paper.