no code implementations • 22 Mar 2024 • Chen Cecilia Liu, Iryna Gurevych
Prior research has found that differences in the early period of neural network training significantly impact the performance of in-distribution (ID) tasks.
1 code implementation • 15 Sep 2023 • Chen Cecilia Liu, Fajri Koto, Timothy Baldwin, Iryna Gurevych
Large language models (LLMs) are highly adept at question answering and reasoning tasks, but when reasoning in a situational context, human expectations vary depending on the relevant cultural common ground.
1 code implementation • 13 Jan 2023 • Chen Cecilia Liu, Jonas Pfeiffer, Ivan Vulić, Iryna Gurevych
Our experiments reveal that scheduled unfreezing induces different learning dynamics compared to standard fine-tuning, and provide evidence that the dynamics of Fisher Information during training correlate with cross-lingual generalization performance.
no code implementations • 12 Oct 2022 • Gregor Geigle, Chen Cecilia Liu, Jonas Pfeiffer, Iryna Gurevych
While many VEs -- of different architectures, trained on different data and objectives -- are publicly available, they are not designed for the downstream V+L tasks.