no code implementations • 5 Apr 2025 • Ruiqi Zhang, Jingfeng Wu, Licong Lin, Peter L. Bartlett
We show that after at most $1/\gamma^2$ burn-in steps, GD achieves a risk upper bounded by $\exp(-\Theta(\eta))$, where $\gamma$ is the margin of the dataset.
no code implementations • 24 Mar 2025 • Jiacheng Wu, Ruiqi Zhang, Jie Chen, HUI ZHANG
Efficiently modeling relightable human avatars from sparse-view videos is crucial for AR/VR applications.
no code implementations • 19 Feb 2025 • Tianyu Guo, Hanlin Zhu, Ruiqi Zhang, Jiantao Jiao, Song Mei, Michael I. Jordan, Stuart Russell
We further propose a three-parameter model that supports the causal claims for the mechanisms to the training dynamics of the transformer.
no code implementations • 6 Dec 2024 • Ruiqi Zhang, Brandon Motes, Shaun Tan, Yongli Lu, Meng-Chen Shih, Yilun Hao, Karen Yang, Shreyas Srinivasan, Moungi G. Bawendi, Vladimir Bulovic
We demonstrate a machine learning (ML) approach that accurately predicts the current-voltage behavior of 3D/2D-structured (FAMA)Pb(IBr)3/OABr hybrid organic-inorganic halide perovskite (HOIP) solar cells under AM1. 5 illumination.
1 code implementation • 26 Oct 2024 • Hanshi Sun, Momin Haider, Ruiqi Zhang, Huitao Yang, Jiahao Qiu, Ming Yin, Mengdi Wang, Peter Bartlett, Andrea Zanette
The safe and effective deployment of Large Language Models (LLMs) involves a critical step called alignment, which ensures that the model's responses are in accordance with human preferences.
2 code implementations • 9 Oct 2024 • Chongyu Fan, Jiancheng Liu, Licong Lin, Jinghan Jia, Ruiqi Zhang, Song Mei, Sijia Liu
In this work, we address the problem of large language model (LLM) unlearning, aiming to remove unwanted data influences and associated model capabilities (e. g., copyrighted data or harmful content generation) while preserving essential model utilities, without the need for retraining from scratch.
2 code implementations • 19 Aug 2024 • Ruiqi Zhang, Jing Hou, Florian Walter, Shangding Gu, Jiayi Guan, Florian Röhrbein, Yali Du, Panpan Cai, Guang Chen, Alois Knoll
Reinforcement Learning (RL) is a potent tool for sequential decision-making and has achieved performance surpassing human capabilities across many challenging real-world tasks.
1 code implementation • 8 Apr 2024 • Ruiqi Zhang, Licong Lin, Yu Bai, Song Mei
LLM unlearning aims to eliminate the influence of undesirable data from the pre-trained model while preserving the model's utilities on other tasks.
no code implementations • 24 Feb 2024 • Ruiqi Zhang, Yuexiang Zhai, Andrea Zanette
Surprisingly, in this work, we demonstrate that even in such a data-starved setting it may still be possible to find a policy competitive with the optimal one.
no code implementations • 22 Feb 2024 • Ruiqi Zhang, Jingfeng Wu, Peter L. Bartlett
We study the \emph{in-context learning} (ICL) ability of a \emph{Linear Transformer Block} (LTB) that combines a linear attention component and a linear multi-layer perceptron (MLP) component.
no code implementations • 18 Feb 2024 • Zhaorun Chen, Zhuokai Zhao, Zhihong Zhu, Ruiqi Zhang, Xiang Li, Bhiksha Raj, Huaxiu Yao
Recent advancements in large language models (LLMs) have shown promise in multi-step reasoning tasks, yet their reliance on extensive manual labeling to provide procedural feedback remains a significant impediment.
no code implementations • 11 Dec 2023 • Jing Hou, Guang Chen, Ruiqi Zhang, Zhijun Li, Shangding Gu, Changjun Jiang
While existing parallel RL frameworks encompass a variety of RL algorithms and parallelization techniques, the excessively burdensome communication frameworks hinder the attainment of the hardware's limit for final throughput and training effects on a single desktop.
no code implementations • 7 Aug 2023 • Ruiqi Zhang, Jie Chen, Qiang Wang
This paper proposes a technique for efficiently modeling dynamic humans by explicifying the implicit neural fields via a Neural Explicit Surface (NES).
no code implementations • 16 Jun 2023 • Ruiqi Zhang, Spencer Frei, Peter L. Bartlett
We show that although gradient flow succeeds at finding a global minimum in this setting, the trained transformer is still brittle under mild covariate shifts.
1 code implementation • 19 Jul 2022 • Ruiqi Zhang, Jie Chen
However, the learned canonical representation is static and the current design of the deformation fields is not able to represent large movements or detailed geometry changes.
no code implementations • 10 Feb 2022 • Ruiqi Zhang, Xuezhou Zhang, Chengzhuo Ni, Mengdi Wang
We approach this problem using the Z-estimation theory and establish the following results: The FQE estimation error is asymptotically normal with explicit variance determined jointly by the tangent space of the function class at the ground truth, the reward structure, and the distribution shift due to off-policy learning; The finite-sample FQE error bound is dominated by the same variance term, and it can also be bounded by function class-dependent divergence, which measures how the off-policy distribution shift intertwines with the function approximator.
no code implementations • 31 Jan 2022 • Chengzhuo Ni, Ruiqi Zhang, Xiang Ji, Xuezhou Zhang, Mengdi Wang
Policy gradient (PG) estimation becomes a challenge when we are not allowed to sample with the target policy but only have access to a dataset generated by some unknown behavior policy.
no code implementations • 4 Mar 2021 • James W Furness, Ruiqi Zhang, Jianwei Sun
In chemistry and condensed matter physics the solution of simple paradigm systems, such as the hydrogen atom and the uniform electron gas, plays a critical role in understanding electron behaviors and developing electronic structure methods.
Chemical Physics