Search Results for author: Anton Alexandrov

Found 4 papers, 2 papers with code

BgGPT 1.0: Extending English-centric LLMs to other languages

no code implementations14 Dec 2024 Anton Alexandrov, Veselin Raychev, Dimitar I. Dimitrov, Ce Zhang, Martin Vechev, Kristina Toutanova

We present BgGPT-Gemma-2-27B-Instruct and BgGPT-Gemma-2-9B-Instruct: continually pretrained and fine-tuned versions of Google's Gemma-2 models, specifically optimized for Bulgarian language understanding and generation.

Continual Learning

Mitigating Catastrophic Forgetting in Language Transfer via Model Merging

no code implementations11 Jul 2024 Anton Alexandrov, Veselin Raychev, Mark Niklas Müller, Ce Zhang, Martin Vechev, Kristina Toutanova

As open-weight large language models (LLMs) achieve ever more impressive performances across a wide range of tasks in English, practitioners aim to adapt these models to different languages.

Cannot find the paper you are looking for? You can Submit a new open access paper.