Search Results for author: Lala Li

Found 12 papers, 8 papers with code

Differentiable Product Quantization for Learning Compact Embedding Layers

no code implementations ICML 2020 Ting Chen, Lala Li, Yizhou Sun

Embedding layers are commonly used to map discrete symbols into continuous embedding vectors that reflect their semantic meanings.

Quantization

Guiding Image Captioning Models Toward More Specific Captions

no code implementations ICCV 2023 Simon Kornblith, Lala Li, ZiRui Wang, Thao Nguyen

We further explore the use of language models to guide the decoding process, obtaining small improvements over the Pareto frontier of reference-free vs. reference-based captioning metrics that arises from classifier-free guidance, and substantially improving the quality of captions generated from a model trained only on minimally curated web data.

Image Captioning Image Retrieval

FIT: Far-reaching Interleaved Transformers

1 code implementation22 May 2023 Ting Chen, Lala Li

We employ two types of transformer layers: local layers operate on data tokens within each group, while global layers operate on a smaller set of introduced latent tokens.

A Unified Sequence Interface for Vision Tasks

1 code implementation15 Jun 2022 Ting Chen, Saurabh Saxena, Lala Li, Tsung-Yi Lin, David J. Fleet, Geoffrey Hinton

Despite that, by formulating the output of each task as a sequence of discrete tokens with a unified interface, we show that one can train a neural network with a single model architecture and loss function on all these tasks, with no task-specific customization.

Image Captioning Instance Segmentation +2

Intriguing Properties of Contrastive Losses

3 code implementations NeurIPS 2021 Ting Chen, Calvin Luo, Lala Li

We construct datasets with explicit and controllable competing features, and show that, for contrastive learning, a few bits of easy-to-learn shared features can suppress, and even fully prevent, the learning of other sets of competing features.

Contrastive Learning Data Augmentation

Big Bidirectional Insertion Representations for Documents

no code implementations WS 2019 Lala Li, William Chan

The Insertion Transformer is well suited for long form text generation due to its parallel generation capabilities, requiring $O(\log_2 n)$ generation steps to generate $n$ tokens.

Sentence Text Generation +1

Learning Compact Embedding Layers via Differentiable Product Quantization

no code implementations25 Sep 2019 Ting Chen, Lala Li, Yizhou Sun

Embedding layers are commonly used to map discrete symbols into continuous embedding vectors that reflect their semantic meanings.

Quantization

Differentiable Product Quantization for End-to-End Embedding Compression

2 code implementations26 Aug 2019 Ting Chen, Lala Li, Yizhou Sun

Embedding layers are commonly used to map discrete symbols into continuous embedding vectors that reflect their semantic meanings.

Quantization

Which Algorithmic Choices Matter at Which Batch Sizes? Insights From a Noisy Quadratic Model

1 code implementation NeurIPS 2019 Guodong Zhang, Lala Li, Zachary Nado, James Martens, Sushant Sachdeva, George E. Dahl, Christopher J. Shallue, Roger Grosse

Increasing the batch size is a popular way to speed up neural network training, but beyond some critical batch size, larger batch sizes yield diminishing returns.

Cannot find the paper you are looking for? You can Submit a new open access paper.