no code implementations • 19 Dec 2024 • Weijia Shi, Xiaochuang Han, Chunting Zhou, Weixin Liang, Xi Victoria Lin, Luke Zettlemoyer, Lili Yu
We also demonstrate that this framework can adapt existing vision-language models with multimodal generation ability.
no code implementations • 7 Nov 2024 • Weixin Liang, Lili Yu, Liang Luo, Srinivasan Iyer, Ning Dong, Chunting Zhou, Gargi Ghosh, Mike Lewis, Wen-tau Yih, Luke Zettlemoyer, Xi Victoria Lin
In the Transfusion setting, where text and image are trained with different objectives, a 7B MoT model matches the image modality performance of the dense baseline with one third of the FLOPs, and a 760M MoT model outperforms a 1. 4B dense baseline across key image generation metrics.
1 code implementation • 1 Apr 2024 • Weixin Liang, Yaohui Zhang, Zhengxuan Wu, Haley Lepp, Wenlong Ji, Xuandong Zhao, Hancheng Cao, Sheng Liu, Siyu He, Zhi Huang, Diyi Yang, Christopher Potts, Christopher D Manning, James Y. Zou
To address this gap, we conduct the first systematic, large-scale analysis across 950, 965 papers published between January 2020 and February 2024 on the arXiv, bioRxiv, and Nature portfolio journals, using a population-level statistical framework to measure the prevalence of LLM-modified content over time.
1 code implementation • 11 Mar 2024 • Weixin Liang, Zachary Izzo, Yaohui Zhang, Haley Lepp, Hancheng Cao, Xuandong Zhao, Lingjiao Chen, Haotian Ye, Sheng Liu, Zhi Huang, Daniel A. McFarland, James Y. Zou
We present an approach for estimating the fraction of text in a large corpus which is likely to be substantially modified or produced by a large language model (LLM).
1 code implementation • 7 Feb 2024 • Weixin Liang, Nazneen Rajani, Xinyu Yang, Ezinwanne Ozoani, Eric Wu, Yiqun Chen, Daniel Scott Smith, James Zou
To evaluate the impact of model cards, we conducted an intervention study by adding detailed model cards to 42 popular models which had no or sparse model cards previously.
1 code implementation • 24 Jan 2024 • Xinyu Yang, Weixin Liang, James Zou
By analyzing all 7, 433 dataset documentation on Hugging Face, our investigation provides an overview of the Hugging Face dataset ecosystem and insights into dataset documentation practices, yielding 5 main findings: (1) The dataset card completion rate shows marked heterogeneity correlated with dataset popularity.
1 code implementation • 3 Oct 2023 • Weixin Liang, Yuhui Zhang, Hancheng Cao, Binglu Wang, Daisy Ding, Xinyu Yang, Kailas Vodrahalli, Siyu He, Daniel Smith, Yian Yin, Daniel McFarland, James Zou
We first quantitatively compared GPT-4's generated feedback with human peer reviewer feedback in 15 Nature family journals (3, 096 papers in total) and the ICLR machine learning conference (1, 709 papers).
1 code implementation • NeurIPS 2023 • Kevin Fu Jiang, Weixin Liang, James Zou, Yongchan Kwon
Assessing the quality and impact of individual data points is critical for improving model performance and mitigating undesirable biases within the training dataset.
1 code implementation • 4 May 2023 • Weixin Liang, Yining Mao, Yongchan Kwon, Xinyu Yang, James Zou
Our work highlights the importance of understanding the nonlinear effects of model improvement on performance in different subpopulations, and has the potential to inform the development of more equitable and responsible machine learning models.
2 code implementations • 6 Apr 2023 • Weixin Liang, Mert Yuksekgonul, Yining Mao, Eric Wu, James Zou
In this study, we evaluate the performance of several widely-used GPT detectors using writing samples from native and non-native English writers.
no code implementations • 11 Oct 2022 • Nazneen Rajani, Weixin Liang, Lingjiao Chen, Meg Mitchell, James Zou
With the advent of Transformers, large language models (LLMs) have saturated well-known NLP benchmarks and leaderboards with high aggregate performance.
no code implementations • 3 Oct 2022 • Xinyi Zhao, Weixin Liang, James Zou
Data is the fuel powering AI and creates tremendous value for many domains.
1 code implementation • 30 Jun 2022 • Zhiying Zhu, Weixin Liang, James Zou
Motivated by this, we propose a novel task, dataset explanation.
no code implementations • 15 Mar 2022 • Roxana Daneshjou, Kailas Vodrahalli, Roberto A Novoa, Melissa Jenkins, Weixin Liang, Veronica Rotemberg, Justin Ko, Susan M Swetter, Elizabeth E Bailey, Olivier Gevaert, Pritam Mukherjee, Michelle Phung, Kiana Yekrang, Bradley Fong, Rachna Sahasrabudhe, Johan A. C. Allerup, Utako Okata-Karigane, James Zou, Albert Chiou
To ascertain potential biases in algorithm performance in this context, we curated the Diverse Dermatology Images (DDI) dataset-the first publicly available, expertly curated, and pathologically confirmed image dataset with diverse skin tones.
2 code implementations • 3 Mar 2022 • Weixin Liang, Yuhui Zhang, Yongchan Kwon, Serena Yeung, James Zou
Our systematic analysis demonstrates that this gap is caused by a combination of model initialization and contrastive learning optimization.
1 code implementation • ICLR 2022 • Weixin Liang, James Zou
We present MetaShift--a collection of 12, 868 sets of natural images across 410 classes--to address this challenge.
3 code implementations • 2 Jan 2022 • Huaxiu Yao, Yu Wang, Sai Li, Linjun Zhang, Weixin Liang, James Zou, Chelsea Finn
Machine learning algorithms typically assume that training and test examples are drawn from the same distribution.
no code implementations • 15 Nov 2021 • Roxana Daneshjou, Kailas Vodrahalli, Weixin Liang, Roberto A Novoa, Melissa Jenkins, Veronica Rotemberg, Justin Ko, Susan M Swetter, Elizabeth E Bailey, Olivier Gevaert, Pritam Mukherjee, Michelle Phung, Kiana Yekrang, Bradley Fong, Rachna Sahasrabudhe, James Zou, Albert Chiou
AI diagnostic tools may aid in early skin cancer detection; however most models have not been assessed on images of diverse skin tones or uncommon diseases.
1 code implementation • ACL 2021 • Weixin Liang, Kai-Hui Liang, Zhou Yu
Open-domain dialog systems have a user-centric goal: to provide humans with an engaging conversation experience.
1 code implementation • NAACL (maiworkshop) 2021 • Weixin Liang, Yanhao Jiang, Zixuan Liu
Images are more than a collection of objects or attributes -- they represent a web of relationships among interconnected objects.
Ranked #1 on Graph Question Answering on GQA
3 code implementations • 21 Nov 2020 • Weixin Liang, Feiyang Niu, Aishwarya Reganti, Govind Thattai, Gokhan Tur
We show that LRTA makes a step towards truly understanding the question while the state-of-the-art model tends to learn superficial correlations from the training data.
1 code implementation • 21 Nov 2020 • Weixin Liang, James Zou
A key challenge of neural group testing is to modify a deep neural network so that it could test multiple samples in one forward pass.
no code implementations • EMNLP 2020 • Weixin Liang, James Zou, Zhou Yu
We propose Active Learning with Contrastive Explanations (ALICE), an expert-in-the-loop training framework that utilizes contrastive natural language explanations to improve data efficiency in learning.
1 code implementation • ACL 2020 • Weixin Liang, James Zou, Zhou Yu
Our experiments show that CMADE achieves 89. 2% accuracy in the dialog comparison task.
no code implementations • 2 Jan 2020 • Weixin Liang, Zixuan Liu, Can Liu
Based on DAWSON, We also propose MUSIC MATINEE, which is the first few-shot music generation model.
2 code implementations • 12 Sep 2019 • Weixin Liang, Youzhi Tian, Chengcai Chen, Zhou Yu
To utilize limited training data more efficiently, we propose Modular Supervision Network (MOSS), an encoder-decoder training framework that could incorporate supervision from various intermediate dialog system modules including natural language understanding, dialog state tracking, dialog policy learning, and natural language generation.