2 code implementations • 27 Mar 2024 • Jerry Wei, Chengrun Yang, Xinying Song, Yifeng Lu, Nathan Hu, Jie Huang, Dustin Tran, Daiyi Peng, Ruibo Liu, Da Huang, Cosmo Du, Quoc V. Le
Empirically, we demonstrate that LLM agents can outperform crowdsourced human annotators - on a set of ~16k individual facts, SAFE agrees with crowdsourced human annotators 72% of the time, and on a random subset of 100 disagreement cases, SAFE wins 76% of the time.
3 code implementations • 7 Sep 2023 • Chengrun Yang, Xuezhi Wang, Yifeng Lu, Hanxiao Liu, Quoc V. Le, Denny Zhou, Xinyun Chen
In this work, we propose Optimization by PROmpting (OPRO), a simple and effective approach to leverage large language models (LLMs) as optimizers, where the optimization task is described in natural language.
1 code implementation • 15 Apr 2022 • Chengrun Yang, Gabriel Bender, Hanxiao Liu, Pieter-Jan Kindermans, Madeleine Udell, Yifeng Lu, Quoc Le, Da Huang
The best neural architecture for a given machine learning problem depends on many factors: not only the complexity and structure of the dataset, but also on resource constraints including latency, compute, energy consumption, etc.
1 code implementation • ICLR 2022 • Chengrun Yang, Ziyang Wu, Jerry Chee, Christopher De Sa, Madeleine Udell
Low-precision arithmetic trains deep learning models using less energy, less memory and less time.
1 code implementation • 1 Jan 2021 • Chengrun Yang, Lijun Ding, Ziyang Wu, Madeleine Udell
Tensors are widely used to represent multiway arrays of data.
no code implementations • 7 Dec 2020 • Jicong Fan, Lijun Ding, Chengrun Yang, Zhao Zhang, Madeleine Udell
The theorems show that a relatively sharper regularizer leads to a tighter error bound, which is consistent with our numerical results.
1 code implementation • 7 Jun 2020 • Chengrun Yang, Jicong Fan, Ziyang Wu, Madeleine Udell
Data scientists seeking a good supervised learning model on a new dataset have many choices to make: they must preprocess the data, select features, possibly reduce the dimension, select an estimation algorithm, and choose hyperparameters for each of these pipeline components.
no code implementations • 4 May 2020 • Jicong Fan, Chengrun Yang, Madeleine Udell
RNLMF constructs a dictionary for the data space by factoring a kernelized feature space; a noisy matrix can then be decomposed as the sum of a sparse noise matrix and a clean data matrix that lies in a low dimensional nonlinear manifold.
1 code implementation • 9 Aug 2018 • Chengrun Yang, Yuji Akimoto, Dae Won Kim, Madeleine Udell
Algorithm selection and hyperparameter tuning remain two of the most challenging tasks in machine learning.