Search Results for author: Chengrun Yang

Found 9 papers, 7 papers with code

Long-form factuality in large language models

2 code implementations27 Mar 2024 Jerry Wei, Chengrun Yang, Xinying Song, Yifeng Lu, Nathan Hu, Jie Huang, Dustin Tran, Daiyi Peng, Ruibo Liu, Da Huang, Cosmo Du, Quoc V. Le

Empirically, we demonstrate that LLM agents can outperform crowdsourced human annotators - on a set of ~16k individual facts, SAFE agrees with crowdsourced human annotators 72% of the time, and on a random subset of 100 disagreement cases, SAFE wins 76% of the time.

16k

Large Language Models as Optimizers

3 code implementations7 Sep 2023 Chengrun Yang, Xuezhi Wang, Yifeng Lu, Hanxiao Liu, Quoc V. Le, Denny Zhou, Xinyun Chen

In this work, we propose Optimization by PROmpting (OPRO), a simple and effective approach to leverage large language models (LLMs) as optimizers, where the optimization task is described in natural language.

GSM8K

TabNAS: Rejection Sampling for Neural Architecture Search on Tabular Datasets

1 code implementation15 Apr 2022 Chengrun Yang, Gabriel Bender, Hanxiao Liu, Pieter-Jan Kindermans, Madeleine Udell, Yifeng Lu, Quoc Le, Da Huang

The best neural architecture for a given machine learning problem depends on many factors: not only the complexity and structure of the dataset, but also on resource constraints including latency, compute, energy consumption, etc.

Image Retrieval Neural Architecture Search +1

Euclidean-Norm-Induced Schatten-p Quasi-Norm Regularization for Low-Rank Tensor Completion and Tensor Robust Principal Component Analysis

no code implementations7 Dec 2020 Jicong Fan, Lijun Ding, Chengrun Yang, Zhao Zhang, Madeleine Udell

The theorems show that a relatively sharper regularizer leads to a tighter error bound, which is consistent with our numerical results.

Efficient AutoML Pipeline Search with Matrix and Tensor Factorization

1 code implementation7 Jun 2020 Chengrun Yang, Jicong Fan, Ziyang Wu, Madeleine Udell

Data scientists seeking a good supervised learning model on a new dataset have many choices to make: they must preprocess the data, select features, possibly reduce the dimension, select an estimation algorithm, and choose hyperparameters for each of these pipeline components.

AutoML

Robust Non-Linear Matrix Factorization for Dictionary Learning, Denoising, and Clustering

no code implementations4 May 2020 Jicong Fan, Chengrun Yang, Madeleine Udell

RNLMF constructs a dictionary for the data space by factoring a kernelized feature space; a noisy matrix can then be decomposed as the sum of a sparse noise matrix and a clean data matrix that lies in a low dimensional nonlinear manifold.

Clustering Denoising +2

OBOE: Collaborative Filtering for AutoML Model Selection

1 code implementation9 Aug 2018 Chengrun Yang, Yuji Akimoto, Dae Won Kim, Madeleine Udell

Algorithm selection and hyperparameter tuning remain two of the most challenging tasks in machine learning.

Active Learning AutoML +4

Cannot find the paper you are looking for? You can Submit a new open access paper.