1 code implementation • 3 Oct 2024 • Rui Meng, Ye Liu, Lifu Tu, Daqing He, Yingbo Zhou, Semih Yavuz
Phrases are fundamental linguistic units through which humans convey semantics.
no code implementations • 22 Aug 2024 • Can Qin, Congying Xia, Krithika Ramakrishnan, Michael Ryoo, Lifu Tu, Yihao Feng, Manli Shu, Honglu Zhou, Anas Awadalla, Jun Wang, Senthil Purushwalkam, Le Xue, Yingbo Zhou, Huan Wang, Silvio Savarese, Juan Carlos Niebles, Zeyuan Chen, ran Xu, Caiming Xiong
We present xGen-VideoSyn-1, a text-to-video (T2V) generation model capable of producing realistic scenes from textual descriptions.
no code implementations • 11 Dec 2023 • Lifu Tu, Semih Yavuz, Jin Qu, Jiacheng Xu, Rui Meng, Caiming Xiong, Yingbo Zhou
Large Language Models (LLMs) have demonstrated a powerful ability for text generation.
1 code implementation • 7 Sep 2023 • Erik Nijkamp, Tian Xie, Hiroaki Hayashi, Bo Pang, Congying Xia, Chen Xing, Jesse Vig, Semih Yavuz, Philippe Laban, Ben Krause, Senthil Purushwalkam, Tong Niu, Wojciech Kryściński, Lidiya Murakhovs'ka, Prafulla Kumar Choubey, Alex Fabbri, Ye Liu, Rui Meng, Lifu Tu, Meghana Bhat, Chien-Sheng Wu, Silvio Savarese, Yingbo Zhou, Shafiq Joty, Caiming Xiong
Most open-source LLMs, on the other hand, are limited in their ability to support longer sequence lengths, which is a key requirement for many tasks that require inference over an input context.
1 code implementation • 3 Apr 2023 • Lifu Tu, Jin Qu, Semih Yavuz, Shafiq Joty, Wenhao Liu, Caiming Xiong, Yingbo Zhou
Our results demonstrate the strong and efficient modeling ability of NLI-based classifiers and the large cross-lingual transfer improvements achieved by our aligned prompts, particularly in few-shot settings.
1 code implementation • 17 Dec 2022 • Rui Meng, Ye Liu, Semih Yavuz, Divyansh Agarwal, Lifu Tu, Ning Yu, JianGuo Zhang, Meghana Bhat, Yingbo Zhou
In this study, we aim to develop unsupervised methods for improving dense retrieval models.
2 code implementations • 22 Oct 2022 • Lifu Tu, Caiming Xiong, Yingbo Zhou
Pre-trained multilingual language models show significant performance gains for zero-shot cross-lingual model transfer on a wide range of natural language understanding (NLU) tasks.
7 code implementations • 25 Mar 2022 • Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong
To democratize this, we train and release a family of large language models up to 16. 1B parameters, called CODEGEN, on natural language and programming language data, and open source the training library JAXFORMER.
Ranked #96 on Code Generation on HumanEval
no code implementations • 27 Aug 2021 • Lifu Tu
In this dissertation, we discuss the concept of the energy function and structured models with different energy functions.
1 code implementation • EMNLP 2020 • Lifu Tu, Tianyu Liu, Kevin Gimpel
Many tasks in natural language processing involve predicting structured outputs, e. g., sequence labeling, semantic role labeling, parsing, and machine translation.
1 code implementation • 14 Jul 2020 • Lifu Tu, Garima Lalwani, Spandana Gella, He He
Recent work has shown that pre-trained language models such as BERT improve robustness to spurious correlations in the dataset.
1 code implementation • ACL 2020 • Lifu Tu, Richard Yuanzhe Pang, Sam Wiseman, Kevin Gimpel
We propose to train a non-autoregressive machine translation model to minimize the energy defined by a pretrained autoregressive model.
no code implementations • EMNLP (spnlp) 2020 • Lifu Tu, Richard Yuanzhe Pang, Kevin Gimpel
Deep energy-based models are powerful, but pose challenges for learning and inference (Belanger and McCallum, 2016).
no code implementations • WS 2019 • Lifu Tu, Xiaoan Ding, Dong Yu, Kevin Gimpel
We propose a simple and effective modeling framework for controlled generation of multiple, diverse outputs.
1 code implementation • NAACL 2019 • Lifu Tu, Kevin Gimpel
One approach is to perform gradient descent with respect to the output structure directly (Belanger and McCallum, 2016).
no code implementations • SEMEVAL 2018 • Manasvi Sagarkar, John Wieting, Lifu Tu, Kevin Gimpel
We study the problem of measuring the quality of automatically-generated stories.
3 code implementations • ICLR 2018 • Lifu Tu, Kevin Gimpel
Prior work used gradient descent for inference, relaxing the structured output to a set of continuous variables and then optimizing the energy with respect to them.
no code implementations • ACL 2017 • Zheng Cai, Lifu Tu, Kevin Gimpel
We consider the ROC story cloze task (Mostafazadeh et al., 2016) and present several findings.
no code implementations • WS 2017 • Lifu Tu, Kevin Gimpel, Karen Livescu
We present models for embedding words in the context of surrounding words.
no code implementations • 7 Feb 2016 • Qingming Tang, Lifu Tu, Weiran Wang, Jinbo Xu
We propose a novel method for network inference from partially observed edges using a node-specific degree prior.