Search Results for author: Weixin Liang

Found 23 papers, 15 papers with code

Mapping the Increasing Use of LLMs in Scientific Papers

no code implementations1 Apr 2024 Weixin Liang, Yaohui Zhang, Zhengxuan Wu, Haley Lepp, Wenlong Ji, Xuandong Zhao, Hancheng Cao, Sheng Liu, Siyu He, Zhi Huang, Diyi Yang, Christopher Potts, Christopher D Manning, James Y. Zou

To address this gap, we conduct the first systematic, large-scale analysis across 950, 965 papers published between January 2020 and February 2024 on the arXiv, bioRxiv, and Nature portfolio journals, using a population-level statistical framework to measure the prevalence of LLM-modified content over time.

What's documented in AI? Systematic Analysis of 32K AI Model Cards

1 code implementation7 Feb 2024 Weixin Liang, Nazneen Rajani, Xinyu Yang, Ezinwanne Ozoani, Eric Wu, Yiqun Chen, Daniel Scott Smith, James Zou

To evaluate the impact of model cards, we conducted an intervention study by adding detailed model cards to 42 popular models which had no or sparse model cards previously.

Informativeness

Navigating Dataset Documentations in AI: A Large-Scale Analysis of Dataset Cards on Hugging Face

1 code implementation24 Jan 2024 Xinyu Yang, Weixin Liang, James Zou

By analyzing all 7, 433 dataset documentation on Hugging Face, our investigation provides an overview of the Hugging Face dataset ecosystem and insights into dataset documentation practices, yielding 5 main findings: (1) The dataset card completion rate shows marked heterogeneity correlated with dataset popularity.

Can large language models provide useful feedback on research papers? A large-scale empirical analysis

1 code implementation3 Oct 2023 Weixin Liang, Yuhui Zhang, Hancheng Cao, Binglu Wang, Daisy Ding, Xinyu Yang, Kailas Vodrahalli, Siyu He, Daniel Smith, Yian Yin, Daniel McFarland, James Zou

We first quantitatively compared GPT-4's generated feedback with human peer reviewer feedback in 15 Nature family journals (3, 096 papers in total) and the ICLR machine learning conference (1, 709 papers).

Accuracy on the Curve: On the Nonlinear Correlation of ML Performance Between Data Subpopulations

1 code implementation4 May 2023 Weixin Liang, Yining Mao, Yongchan Kwon, Xinyu Yang, James Zou

Our work highlights the importance of understanding the nonlinear effects of model improvement on performance in different subpopulations, and has the potential to inform the development of more equitable and responsible machine learning models.

Fairness

GPT detectors are biased against non-native English writers

2 code implementations6 Apr 2023 Weixin Liang, Mert Yuksekgonul, Yining Mao, Eric Wu, James Zou

In this study, we evaluate the performance of several widely-used GPT detectors using writing samples from native and non-native English writers.

Fairness

SEAL : Interactive Tool for Systematic Error Analysis and Labeling

no code implementations11 Oct 2022 Nazneen Rajani, Weixin Liang, Lingjiao Chen, Meg Mitchell, James Zou

With the advent of Transformers, large language models (LLMs) have saturated well-known NLP benchmarks and leaderboards with high aggregate performance.

Data Budgeting for Machine Learning

no code implementations3 Oct 2022 Xinyi Zhao, Weixin Liang, James Zou

Data is the fuel powering AI and creates tremendous value for many domains.

Disparities in Dermatology AI Performance on a Diverse, Curated Clinical Image Set

no code implementations15 Mar 2022 Roxana Daneshjou, Kailas Vodrahalli, Roberto A Novoa, Melissa Jenkins, Weixin Liang, Veronica Rotemberg, Justin Ko, Susan M Swetter, Elizabeth E Bailey, Olivier Gevaert, Pritam Mukherjee, Michelle Phung, Kiana Yekrang, Bradley Fong, Rachna Sahasrabudhe, Johan A. C. Allerup, Utako Okata-Karigane, James Zou, Albert Chiou

To ascertain potential biases in algorithm performance in this context, we curated the Diverse Dermatology Images (DDI) dataset-the first publicly available, expertly curated, and pathologically confirmed image dataset with diverse skin tones.

Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation Learning

2 code implementations3 Mar 2022 Weixin Liang, Yuhui Zhang, Yongchan Kwon, Serena Yeung, James Zou

Our systematic analysis demonstrates that this gap is caused by a combination of model initialization and contrastive learning optimization.

Contrastive Learning Fairness +2

MetaShift: A Dataset of Datasets for Evaluating Contextual Distribution Shifts and Training Conflicts

1 code implementation ICLR 2022 Weixin Liang, James Zou

We present MetaShift--a collection of 12, 868 sets of natural images across 410 classes--to address this challenge.

Benchmarking

Improving Out-of-Distribution Robustness via Selective Augmentation

2 code implementations2 Jan 2022 Huaxiu Yao, Yu Wang, Sai Li, Linjun Zhang, Weixin Liang, James Zou, Chelsea Finn

Machine learning algorithms typically assume that training and test examples are drawn from the same distribution.

Neural Group Testing to Accelerate Deep Learning

1 code implementation21 Nov 2020 Weixin Liang, James Zou

A key challenge of neural group testing is to modify a deep neural network so that it could test multiple samples in one forward pass.

LRTA: A Transparent Neural-Symbolic Reasoning Framework with Modular Supervision for Visual Question Answering

2 code implementations21 Nov 2020 Weixin Liang, Feiyang Niu, Aishwarya Reganti, Govind Thattai, Gokhan Tur

We show that LRTA makes a step towards truly understanding the question while the state-of-the-art model tends to learn superficial correlations from the training data.

Answer Generation Question Answering +1

ALICE: Active Learning with Contrastive Natural Language Explanations

no code implementations EMNLP 2020 Weixin Liang, James Zou, Zhou Yu

We propose Active Learning with Contrastive Explanations (ALICE), an expert-in-the-loop training framework that utilizes contrastive natural language explanations to improve data efficiency in learning.

Active Learning Classification +1

DAWSON: A Domain Adaptive Few Shot Generation Framework

no code implementations2 Jan 2020 Weixin Liang, Zixuan Liu, Can Liu

Based on DAWSON, We also propose MUSIC MATINEE, which is the first few-shot music generation model.

Meta-Learning Music Generation

MOSS: End-to-End Dialog System Framework with Modular Supervision

1 code implementation12 Sep 2019 Weixin Liang, Youzhi Tian, Chengcai Chen, Zhou Yu

To utilize limited training data more efficiently, we propose Modular Supervision Network (MOSS), an encoder-decoder training framework that could incorporate supervision from various intermediate dialog system modules including natural language understanding, dialog state tracking, dialog policy learning, and natural language generation.

dialog state tracking Natural Language Understanding +1

Cannot find the paper you are looking for? You can Submit a new open access paper.