no code implementations • 11 Nov 2024 • Young-Min Cho, Raphael Shu, Nilaksh Das, Tamer Alkhouli, Yi-An Lai, Jason Cai, Monica Sunkara, Yi Zhang
This study investigates the efficacy of Multi-Agent Systems in eliciting cross-agent communication and enhancing collective intelligence through group decision-making in a decentralized setting.
no code implementations • 5 Feb 2024 • James Y. Huang, Sailik Sengupta, Daniele Bonadiman, Yi-An Lai, Arshit Gupta, Nikolaos Pappas, Saab Mansour, Katrin Kirchhoff, Dan Roth
Current work focuses on alignment at model training time, through techniques such as Reinforcement Learning with Human Feedback (RLHF).
no code implementations • 4 Feb 2023 • Yi-An Lai, Elman Mansimov, Yuqing Xie, Yi Zhang
When upgrading neural models to a newer version, new errors that were not encountered in the legacy version can be introduced, known as regression errors.
no code implementations • 25 Jan 2023 • Raphael Schumann, Elman Mansimov, Yi-An Lai, Nikolaos Pappas, Xibin Gao, Yi Zhang
This method interpolates between the weights of the old and new model and we show in extensive experiments that it reduces negative flips without sacrificing the improved accuracy of the new model.
no code implementations • 7 Feb 2022 • Deng Cai, Elman Mansimov, Yi-An Lai, Yixuan Su, Lei Shu, Yi Zhang
First, we measure and analyze model update regression in different model update settings.
2 code implementations • ACL 2022 • Yixuan Su, Lei Shu, Elman Mansimov, Arshit Gupta, Deng Cai, Yi-An Lai, Yi Zhang
Pre-trained language models have been recently shown to benefit task-oriented dialogue (TOD) systems.
no code implementations • EMNLP (sustainlp) 2021 • Vin Sachidananda, Jason S. Kessler, Yi-An Lai
While adaptive tokenization incurs a 6% increase in model parameters in our experimentation, due to the introduction of 10k new domain-specific tokens, our approach, using 64 vCPUs, is 72x faster than further pretraining the language model on domain-specific corpora on 8 TPUs.
no code implementations • ACL 2021 • Yuqing Xie, Yi-An Lai, Yuanjun Xiong, Yi Zhang, Stefano Soatto
Behavior of deep neural networks can be inconsistent between different versions.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Yi-An Lai, Garima Lalwani, Yi Zhang
Pre-trained language models that learn contextualized word representations from a large un-annotated corpus have become a standard component for many state-of-the-art NLP systems.
no code implementations • LREC 2020 • Yi-An Lai, Xuan Zhu, Yi Zhang, Mona Diab
Summarizing data samples by quantitative measures has a long history, with descriptive statistics being a case in point.
no code implementations • CONLL 2019 • Yi-An Lai, Arshit Gupta, Yi Zhang
Hierarchical neural networks are often used to model inherent structures within dialogues.
no code implementations • 20 Oct 2018 • Wen-Hao Chen, Chin-Chi Hsu, Yi-An Lai, Vincent Liu, Mi-Yen Yeh, Shou-De Lin
Attribute-aware CF models aims at rating prediction given not only the historical rating from users to items, but also the information associated with users (e. g. age), items (e. g. price), or even ratings (e. g. rating time).
1 code implementation • NeurIPS 2017 • Yi-An Lai, Chin-Chi Hsu, Wen Hao Chen, Mi-Yen Yeh, Shou-De Lin
We investigate an unsupervised generative approach for network embedding.