Search Results for author: Jiayi Yuan

Found 20 papers, 6 papers with code

Interpreting and Steering LLMs with Mutual Information-based Explanations on Sparse Autoencoders

no code implementations21 Feb 2025 Xuansheng Wu, Jiayi Yuan, Wenlin Yao, Xiaoming Zhai, Ninghao Liu

Large language models (LLMs) excel at handling human queries, but they can occasionally generate flawed or unexpected responses.

Robot Learning with Super-Linear Scaling

no code implementations2 Dec 2024 Marcel Torne, Arhan Jain, Jiayi Yuan, Vidaaranya Macha, Lars Ankile, Anthony Simeonov, Pulkit Agrawal, Abhishek Gupta

In this work, we propose Crowdsourcing and Amortizing Human Effort for Real-to-Sim-to-Real(CASHER), a pipeline for scaling up data collection and learning in simulation where the performance scales superlinearly with human effort.

3D Reconstruction

InvestESG: A multi-agent reinforcement learning benchmark for studying climate investment as a social dilemma

1 code implementation15 Nov 2024 Xiaoxuan Hou, Jiayi Yuan, Joel Z. Leibo, Natasha Jaques

InvestESG is a novel multi-agent reinforcement learning (MARL) benchmark designed to study the impact of Environmental, Social, and Governance (ESG) disclosure mandates on corporate climate investments.

Multi-agent Reinforcement Learning

LoRATK: LoRA Once, Backdoor Everywhere in the Share-and-Play Ecosystem

no code implementations29 Feb 2024 Hongyi Liu, Shaochen Zhong, Xintong Sun, Minghao Tian, Mohsen Hariri, Zirui Liu, Ruixiang Tang, Zhimeng Jiang, Jiayi Yuan, Yu-Neng Chuang, Li Li, Soo-Hyun Choi, Rui Chen, Vipin Chaudhary, Xia Hu

We find that with a simple, efficient, yet specific recipe, a backdoor LoRA can be trained once and then seamlessly merged (in a training-free fashion) with multiple task-enhancing LoRAs, retaining both its malicious backdoor and benign downstream capabilities.

KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache

1 code implementation5 Feb 2024 Zirui Liu, Jiayi Yuan, Hongye Jin, Shaochen Zhong, Zhaozhuo Xu, Vladimir Braverman, Beidi Chen, Xia Hu

However, there is a lack of in-depth studies that explore the element distribution of KV cache to understand the hardness and limitation of KV cache quantization.

Quantization

NetBooster: Empowering Tiny Deep Learning By Standing on the Shoulders of Deep Giants

no code implementations23 Jun 2023 Zhongzhi Yu, Yonggan Fu, Jiayi Yuan, Haoran You, Yingyan Lin

Tiny deep learning has attracted increasing attention driven by the substantial demand for deploying deep learning on numerous intelligent Internet-of-Things devices.

Deep Learning

Gen-NeRF: Efficient and Generalizable Neural Radiance Fields via Algorithm-Hardware Co-Design

no code implementations24 Apr 2023 Yonggan Fu, Zhifan Ye, Jiayi Yuan, Shunyao Zhang, Sixu Li, Haoran You, Yingyan Celine Lin

Novel view synthesis is an essential functionality for enabling immersive experiences in various Augmented- and Virtual-Reality (AR/VR) applications, for which generalizable Neural Radiance Fields (NeRFs) have gained increasing popularity thanks to their cross-scene generalization capability.

Generalizable Novel View Synthesis NeRF +1

Robust Tickets Can Transfer Better: Drawing More Transferable Subnetworks in Transfer Learning

no code implementations24 Apr 2023 Yonggan Fu, Ye Yuan, Shang Wu, Jiayi Yuan, Yingyan Celine Lin

Transfer learning leverages feature representations of deep neural networks (DNNs) pretrained on source tasks with rich data to empower effective finetuning on downstream tasks.

Adversarial Robustness Transfer Learning

Large Language Models for Healthcare Data Augmentation: An Example on Patient-Trial Matching

no code implementations24 Mar 2023 Jiayi Yuan, Ruixiang Tang, Xiaoqian Jiang, Xia Hu

The process of matching patients with suitable clinical trials is essential for advancing medical research and providing optimal care.

Data Augmentation Text Generation

Recurrent Structure Attention Guidance for Depth Super-Resolution

no code implementations31 Jan 2023 Jiayi Yuan, Haobo Jiang, Xiang Li, Jianjun Qian, Jun Li, Jian Yang

Second, instead of the coarse concatenation guidance, we propose a recurrent structure attention block, which iteratively utilizes the latest depth estimation and the image features to jointly select clear patterns and boundaries, aiming at providing refined guidance for accurate depth recovery.

Depth Estimation Super-Resolution

Structure Flow-Guided Network for Real Depth Super-Resolution

no code implementations31 Jan 2023 Jiayi Yuan, Haobo Jiang, Xiang Li, Jianjun Qian, Jun Li, Jian Yang

Specifically, our framework consists of a cross-modality flow-guided upsampling network (CFUNet) and a flow-enhanced pyramid edge attention network (PEANet).

Depth Estimation Depth Prediction +1

DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks

2 code implementations2 Jun 2022 Yonggan Fu, Haichuan Yang, Jiayi Yuan, Meng Li, Cheng Wan, Raghuraman Krishnamoorthi, Vikas Chandra, Yingyan Celine Lin

Efficient deep neural network (DNN) models equipped with compact operators (e. g., depthwise convolutions) have shown great potential in reducing DNNs' theoretical complexity (e. g., the total number of weights/operations) while maintaining a decent model accuracy.

Cannot find the paper you are looking for? You can Submit a new open access paper.