Search Results for author: Nikhil Sardana

Found 5 papers, 3 papers with code

Sparse Upcycling: Inference Inefficient Finetuning

no code implementations13 Nov 2024 Sasha Doubov, Nikhil Sardana, Vitaliy Chiley

Small, highly trained, open-source large language models are widely used due to their inference efficiency, but further improving their quality remains a challenge.

Beyond Chinchilla-Optimal: Accounting for Inference in Language Model Scaling Laws

no code implementations31 Dec 2023 Nikhil Sardana, Jacob Portes, Sasha Doubov, Jonathan Frankle

We modify the Chinchilla scaling laws to calculate the optimal LLM parameter count and pre-training data size to train and deploy a model of a given quality and inference demand.

Language Modelling Large Language Model

Autonomous Reinforcement Learning: Formalism and Benchmarking

2 code implementations ICLR 2022 Archit Sharma, Kelvin Xu, Nikhil Sardana, Abhishek Gupta, Karol Hausman, Sergey Levine, Chelsea Finn

In this paper, we aim to address this discrepancy by laying out a framework for Autonomous Reinforcement Learning (ARL): reinforcement learning where the agent not only learns through its own experience, but also contends with lack of human supervision to reset between trials.

Benchmarking reinforcement-learning +2

Bayesian Meta-Learning Through Variational Gaussian Processes

1 code implementation21 Oct 2021 Vivek Myers, Nikhil Sardana

This problem setting can be extended to the Bayesian context, wherein rather than predicting a single label for each query data point, a model predicts a distribution of labels capturing its uncertainty.

Gaussian Processes Meta-Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.