Search Results for author: Zhengyang Geng

Found 9 papers, 8 papers with code

Medusa: Simple LLM Inference Acceleration Framework with Multiple Decoding Heads

1 code implementation19 Jan 2024 Tianle Cai, Yuhong Li, Zhengyang Geng, Hongwu Peng, Jason D. Lee, Deming Chen, Tri Dao

We present two levels of fine-tuning procedures for Medusa to meet the needs of different use cases: Medusa-1: Medusa is directly fine-tuned on top of a frozen backbone LLM, enabling lossless inference acceleration.

One-Step Diffusion Distillation via Deep Equilibrium Models

1 code implementation NeurIPS 2023 Zhengyang Geng, Ashwini Pokle, J. Zico Kolter

We demonstrate that the DEQ architecture is crucial to this capability, as GET matches a $5\times$ larger ViT in terms of FID scores while striking a critical balance of computational cost and image quality.

TorchDEQ: A Library for Deep Equilibrium Models

1 code implementation28 Oct 2023 Zhengyang Geng, J. Zico Kolter

Deep Equilibrium (DEQ) Models, an emerging class of implicit models that maps inputs to fixed points of neural networks, are of growing interest in the deep learning community.

Deep Equilibrium Approaches to Diffusion Models

1 code implementation23 Oct 2022 Ashwini Pokle, Zhengyang Geng, Zico Kolter

In this paper, we look at diffusion models through a different perspective, that of a (deep) equilibrium (DEQ) fixed point model.

Denoising

Eliminating Gradient Conflict in Reference-based Line-Art Colorization

1 code implementation13 Jul 2022 Zekun Li, Zhengyang Geng, Zhao Kang, Wenyu Chen, Yibo Yang

To understand the instability in training, we detect the gradient flow of attention and observe gradient conflict among attention branches.

Line Art Colorization SSIM

Deep Equilibrium Optical Flow Estimation

1 code implementation CVPR 2022 Shaojie Bai, Zhengyang Geng, Yash Savani, J. Zico Kolter

Many recent state-of-the-art (SOTA) optical flow models use finite-step recurrent update operations to emulate traditional algorithms by encouraging iterative refinements toward a stable flow estimation.

Optical Flow Estimation

Residual Relaxation for Multi-view Representation Learning

no code implementations NeurIPS 2021 Yifei Wang, Zhengyang Geng, Feng Jiang, Chuming Li, Yisen Wang, Jiansheng Yang, Zhouchen Lin

Multi-view methods learn representations by aligning multiple views of the same image and their performance largely depends on the choice of data augmentation.

Data Augmentation Representation Learning

Is Attention Better Than Matrix Decomposition?

2 code implementations ICLR 2021 Zhengyang Geng, Meng-Hao Guo, Hongxu Chen, Xia Li, Ke Wei, Zhouchen Lin

As an essential ingredient of modern deep learning, attention mechanism, especially self-attention, plays a vital role in the global correlation discovery.

Conditional Image Generation Semantic Segmentation

Cannot find the paper you are looking for? You can Submit a new open access paper.