Search Results for author: Christopher Carothers

Found 1 papers, 0 papers with code

A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts

no code implementations26 May 2024 Mohammed Nowaz Rabbani Chowdhury, Meng Wang, Kaoutar El Maghraoui, Naigang Wang, Pin-Yu Chen, Christopher Carothers

The sparsely gated mixture of experts (MoE) architecture sends different inputs to different subnetworks, i. e., experts, through trainable routers.

Binary Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.