Search Results for author: Arash Ahmadian

Found 5 papers, 2 papers with code

Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs

no code implementations22 Feb 2024 Arash Ahmadian, Chris Cremer, Matthias Gallé, Marzieh Fadaee, Julia Kreutzer, Olivier Pietquin, Ahmet Üstün, Sara Hooker

AI alignment in the shape of Reinforcement Learning from Human Feedback (RLHF) is increasingly treated as a crucial ingredient for high performance large language models.

Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning

1 code implementation11 Sep 2023 Ted Zadouri, Ahmet Üstün, Arash Ahmadian, Beyza Ermiş, Acyr Locatelli, Sara Hooker

The Mixture of Experts (MoE) is a widely known neural architecture where an ensemble of specialized sub-models optimizes overall performance with a constant computational cost.

$λ$-models: Effective Decision-Aware Reinforcement Learning with Latent Models

no code implementations30 Jun 2023 Claas A Voelcker, Arash Ahmadian, Romina Abachi, Igor Gilitschenski, Amir-Massoud Farahmand

The idea of decision-aware model learning, that models should be accurate where it matters for decision-making, has gained prominence in model-based reinforcement learning.

Continuous Control Decision Making +2

Pseudo-Inverted Bottleneck Convolution for DARTS Search Space

1 code implementation31 Dec 2022 Arash Ahmadian, Louis S. P. Liu, Yue Fei, Konstantinos N. Plataniotis, Mahdi S. Hosseini

Our proposed architecture is much less sensitive to evaluation layer count and outperforms a DARTS network with similar size significantly, at layer counts as small as 2.

Neural Architecture Search

Cannot find the paper you are looking for? You can Submit a new open access paper.