no code implementations • 16 Oct 2024 • Yong Xie, Karan Aggarwal, Aitzaz Ahmad, Stephen Lau
Hallucination pattern guidance leverages the most important task-specific hallucination patterns while language style alignment aligns the style of the synthetic dataset with benchmark text.
no code implementations • 23 Sep 2024 • Mingqi Li, Karan Aggarwal, Yong Xie, Aitzaz Ahmad, Stephen Lau
As LLMs evolve, significant effort is spent on manually crafting prompts.
no code implementations • 14 Nov 2023 • Yong Xie, Karan Aggarwal, Aitzaz Ahmad
We further explore simple but effective data selection strategies for continual pre-training.