no code implementations • 3 Feb 2024 • Haitao Mao, Guangliang Liu, Yao Ma, Rongrong Wang, Jiliang Tang
In-Context Learning (ICL) empowers Large Language Models (LLMs) with the capacity to learn in context, achieving downstream generalization without gradient updates but with a few in-context examples.
no code implementations • 26 Oct 2023 • Guangliang Liu, Zhiyu Xue, Xitong Zhang, Kristen Marie Johnson, Rongrong Wang
Fine-tuning pretrained language models (PLMs) for downstream tasks is a large-scale optimization problem, in which the choice of the training algorithm critically determines how well the trained model can generalize to unseen test data, especially in the context of few-shot learning.
no code implementations • 30 May 2023 • Xitong Zhang, Avrajit Ghosh, Guangliang Liu, Rongrong Wang
It is widely recognized that the generalization ability of neural networks can be greatly enhanced through carefully designing the training procedure.