Search Results for author: Alexandre Marques

Found 4 papers, 0 papers with code

Enabling High-Sparsity Foundational Llama Models with Efficient Pretraining and Deployment

no code implementations6 May 2024 Abhinav Agarwalla, Abhay Gupta, Alexandre Marques, Shubhra Pandit, Michael Goin, Eldar Kurtic, Kevin Leong, Tuan Nguyen, Mahmoud Salem, Dan Alistarh, Sean Lie, Mark Kurtz

We achieve this for the LLaMA-2 7B model by combining the SparseGPT one-shot pruning method and sparse pretraining of those models on a subset of the SlimPajama dataset mixed with a Python subset of The Stack dataset.

Arithmetic Reasoning Code Generation +2

oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes

no code implementations30 Mar 2023 Daniel Campos, Alexandre Marques, Mark Kurtz, ChengXiang Zhai

In this paper, we introduce the range of oBERTa language models, an easy-to-use set of language models which allows Natural Language Processing (NLP) practitioners to obtain between 3. 8 and 24. 3 times faster models without expertise in model compression.

Knowledge Distillation Model Compression +3

Sparse*BERT: Sparse Models Generalize To New tasks and Domains

no code implementations25 May 2022 Daniel Campos, Alexandre Marques, Tuan Nguyen, Mark Kurtz, ChengXiang Zhai

Our experimentation shows that models that are pruned during pretraining using general domain masked language models can transfer to novel domains and tasks without extensive hyperparameter exploration or specialized approaches.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.