Search Results for author: Prajjwal Bhargava

Found 10 papers, 8 papers with code

Effective Long-Context Scaling of Foundation Models

1 code implementation27 Sep 2023 Wenhan Xiong, Jingyu Liu, Igor Molybog, Hejia Zhang, Prajjwal Bhargava, Rui Hou, Louis Martin, Rashi Rungta, Karthik Abinav Sankararaman, Barlas Oguz, Madian Khabsa, Han Fang, Yashar Mehdad, Sharan Narang, Kshitiz Malik, Angela Fan, Shruti Bhosale, Sergey Edunov, Mike Lewis, Sinong Wang, Hao Ma

We also examine the impact of various design choices in the pretraining process, including the data mix and the training curriculum of sequence lengths -- our ablation experiments suggest that having abundant long texts in the pretrain dataset is not the key to achieving strong performance, and we empirically verify that long context continual pretraining is more efficient and similarly effective compared to pretraining from scratch with long sequences.

Continual Pretraining Language Modelling

When should we prefer Decision Transformers for Offline Reinforcement Learning?

1 code implementation23 May 2023 Prajjwal Bhargava, Rohan Chitnis, Alborz Geramifard, Shagun Sodhani, Amy Zhang

Three popular algorithms for offline RL are Conservative Q-Learning (CQL), Behavior Cloning (BC), and Decision Transformer (DT), from the class of Q-Learning, Imitation Learning, and Sequence Modeling respectively.

D4RL Imitation Learning +5

AUTODIAL: Efficient Asynchronous Task-Oriented Dialogue Model

no code implementations10 Mar 2023 Prajjwal Bhargava, Pooyan Amini, Shahin Shayandeh, Chinnadhurai Sankar

As large dialogue models become commonplace in practice, the problems surrounding high compute requirements for training, inference and larger memory footprint still persists.

Dialogue State Tracking

DiscoSense: Commonsense Reasoning with Discourse Connectives

1 code implementation22 Oct 2022 Prajjwal Bhargava, Vincent Ng

We present DiscoSense, a benchmark for commonsense reasoning via understanding a wide variety of discourse connectives.

Sentence Completion

Commonsense Knowledge Reasoning and Generation with Pre-trained Language Models: A Survey

no code implementations28 Jan 2022 Prajjwal Bhargava, Vincent Ng

While commonsense knowledge acquisition and reasoning has traditionally been a core research topic in the knowledge representation and reasoning community, recent years have seen a surge of interest in the natural language processing community in developing pre-trained models and testing their ability to address a variety of newly designed commonsense knowledge reasoning and generation tasks.

Generalization in NLI: Ways (Not) To Go Beyond Simple Heuristics

1 code implementation EMNLP (insights) 2021 Prajjwal Bhargava, Aleksandr Drozd, Anna Rogers

Much of recent progress in NLU was shown to be due to models' learning dataset-specific heuristics.

Adaptive Transformers for Learning Multimodal Representations

2 code implementations ACL 2020 Prajjwal Bhargava

The usage of transformers has grown from learning about language semantics to forming meaningful visiolinguistic representations.

Computational Efficiency

On Generalizing Detection Models for Unconstrained Environments

1 code implementation28 Sep 2019 Prajjwal Bhargava

We address the problem of incremental learning in object detection on the India Driving Dataset (IDD).

Incremental Learning Object +3

Incremental Learning in Person Re-Identification

1 code implementation20 Aug 2018 Prajjwal Bhargava

In this paper, we propose a model that can be used for multiple tasks in Person Re-Identification, provide state-of-the-art results on a variety of tasks and still achieve considerable accuracy subsequently.

Incremental Learning Person Re-Identification

Cannot find the paper you are looking for? You can Submit a new open access paper.