Search Results for author: Tatsunori B. Hashimoto

Found 25 papers, 18 papers with code

Length-Controlled AlpacaEval: A Simple Way to Debias Automatic Evaluators

1 code implementation6 Apr 2024 Yann Dubois, Balázs Galambosi, Percy Liang, Tatsunori B. Hashimoto

Even simple, known confounders such as preference for longer outputs remain in existing automated evaluation metrics.

Chatbot counterfactual

Benchmarking Multi-Domain Active Learning on Image Classification

no code implementations1 Dec 2023 Jiayi Li, Rohan Taori, Tatsunori B. Hashimoto

Active learning aims to enhance model performance by strategically labeling informative data points.

Active Learning Benchmarking +2

Proving Test Set Contamination in Black Box Language Models

1 code implementation26 Oct 2023 Yonatan Oren, Nicole Meister, Niladri Chatterji, Faisal Ladhak, Tatsunori B. Hashimoto

In contrast, the tendency for language models to memorize example order means that a contaminated language model will find certain canonical orderings to be much more likely than others.

Language Modelling

One Step of Gradient Descent is Provably the Optimal In-Context Learner with One Layer of Linear Self-Attention

no code implementations7 Jul 2023 Arvind Mahankali, Tatsunori B. Hashimoto, Tengyu Ma

Then, we find that changing the distribution of the covariates and weight vector to a non-isotropic Gaussian distribution has a strong impact on the learned algorithm: the global minimizer of the pre-training loss now implements a single step of $\textit{pre-conditioned}$ GD.

In-Context Learning regression

AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback

2 code implementations NeurIPS 2023 Yann Dubois, Xuechen Li, Rohan Taori, Tianyi Zhang, Ishaan Gulrajani, Jimmy Ba, Carlos Guestrin, Percy Liang, Tatsunori B. Hashimoto

As a demonstration of the research possible in AlpacaFarm, we find that methods that use a reward model can substantially improve over supervised fine-tuning and that our reference PPO implementation leads to a +10% improvement in win-rate against Davinci003.

Instruction Following

Benchmarking Large Language Models for News Summarization

1 code implementation31 Jan 2023 Tianyi Zhang, Faisal Ladhak, Esin Durmus, Percy Liang, Kathleen McKeown, Tatsunori B. Hashimoto

Large language models (LLMs) have shown promise for automatic summarization but the reasons behind their successes are poorly understood.

Benchmarking News Summarization

Coder Reviewer Reranking for Code Generation

1 code implementation29 Nov 2022 Tianyi Zhang, Tao Yu, Tatsunori B. Hashimoto, Mike Lewis, Wen-tau Yih, Daniel Fried, Sida I. Wang

Sampling diverse programs from a code language model and reranking with model likelihood is a popular method for code generation but it is prone to preferring degenerate solutions.

Code Generation Language Modelling

Data Feedback Loops: Model-driven Amplification of Dataset Biases

1 code implementation8 Sep 2022 Rohan Taori, Tatsunori B. Hashimoto

Datasets scraped from the internet have been critical to the successes of large-scale machine learning.

Image Classification Text Generation

Diffusion-LM Improves Controllable Text Generation

1 code implementation27 May 2022 Xiang Lisa Li, John Thickstun, Ishaan Gulrajani, Percy Liang, Tatsunori B. Hashimoto

Controlling the behavior of language models (LMs) without re-training is a major open problem in natural language generation.

Language Modelling Sentence +1

TempLM: Distilling Language Models into Template-Based Generators

1 code implementation23 May 2022 Tianyi Zhang, Mina Lee, Lisa Li, Ende Shen, Tatsunori B. Hashimoto

While pretrained language models (PLMs) have greatly improved text generation, they have also been known to produce unfaithful or inappropriate content.

Text Generation

DReCa: A General Task Augmentation Strategy for Few-Shot Natural Language Inference

no code implementations NAACL 2021 Shikhar Murty, Tatsunori B. Hashimoto, Christopher Manning

Meta-learning promises few-shot learners that can adapt to new distributions by repurposing knowledge acquired from previous training.

Clustering Few-Shot NLI +2

Distributionally Robust Neural Networks

1 code implementation ICLR 2020 Shiori Sagawa*, Pang Wei Koh*, Tatsunori B. Hashimoto, Percy Liang

Distributionally robust optimization (DRO) allows us to learn models that instead minimize the worst-case training loss over a set of pre-defined groups.

L2 Regularization Natural Language Inference +1

Learning Autocomplete Systems as a Communication Game

1 code implementation16 Nov 2019 Mina Lee, Tatsunori B. Hashimoto, Percy Liang

We study textual autocomplete---the task of predicting a full sentence from a partial sentence---as a human-machine communication game.

Sentence

Distributionally Robust Language Modeling

1 code implementation IJCNLP 2019 Yonatan Oren, Shiori Sagawa, Tatsunori B. Hashimoto, Percy Liang

Language models are generally trained on data spanning a wide range of topics (e. g., news, reviews, fiction), but they might be applied to an a priori unknown target distribution (e. g., restaurant reviews).

Language Modelling

A Retrieve-and-Edit Framework for Predicting Structured Outputs

1 code implementation NeurIPS 2018 Tatsunori B. Hashimoto, Kelvin Guu, Yonatan Oren, Percy Liang

For the task of generating complex outputs such as source code, editing existing outputs can be easier than generating complex outputs from scratch.

Retrieval

Fairness Without Demographics in Repeated Loss Minimization

1 code implementation ICML 2018 Tatsunori B. Hashimoto, Megha Srivastava, Hongseok Namkoong, Percy Liang

Machine learning models (e. g., speech recognizers) are usually trained to minimize average loss, which results in representation disparity---minority groups (e. g., non-native speakers) contribute less to the training objective and thus tend to suffer higher loss.

Fairness

Derivative free optimization via repeated classification

1 code implementation11 Apr 2018 Tatsunori B. Hashimoto, Steve Yadlowsky, John C. Duchi

We develop an algorithm for minimizing a function using $n$ batched function value measurements at each of $T$ rounds by using classifiers to identify a function's sublevel set.

Active Learning Classification +1

Unsupervised Transformation Learning via Convex Relaxations

1 code implementation NeurIPS 2017 Tatsunori B. Hashimoto, John C. Duchi, Percy Liang

Our goal is to extract meaningful transformations from raw images, such as varying the thickness of lines in handwriting or the lighting in a portrait.

Generating Sentences by Editing Prototypes

3 code implementations TACL 2018 Kelvin Guu, Tatsunori B. Hashimoto, Yonatan Oren, Percy Liang

We propose a new generative model of sentences that first samples a prototype sentence from the training corpus and then edits it into a new sentence.

Language Modelling Sentence +1

From random walks to distances on unweighted graphs

no code implementations NeurIPS 2015 Tatsunori B. Hashimoto, Yi Sun, Tommi S. Jaakkola

Using these techniques we generalize results on the degeneracy of hitting times and analyze a metric based on the Laplace transformed hitting time (LTHT).

Clustering

Word, graph and manifold embedding from Markov processes

no code implementations18 Sep 2015 Tatsunori B. Hashimoto, David Alvarez-Melis, Tommi S. Jaakkola

Continuous vector representations of words and objects appear to carry surprisingly rich semantic content.

Dimensionality Reduction Word Embeddings

Metric recovery from directed unweighted graphs

no code implementations20 Nov 2014 Tatsunori B. Hashimoto, Yi Sun, Tommi S. Jaakkola

We demonstrate empirically that the estimator performs well on simulated examples as well as on real-world co-purchasing graphs even with a small number of points and degree scaling as low as $\log(n)$.

Cannot find the paper you are looking for? You can Submit a new open access paper.