Search Results for author: Yatin Nandwani

Found 10 papers, 5 papers with code

BRAIn: Bayesian Reward-conditioned Amortized Inference for natural language generation from feedback

no code implementations4 Feb 2024 Gaurav Pandey, Yatin Nandwani, Tahira Naseem, Mayank Mishra, Guangxuan Xu, Dinesh Raghu, Sachindra Joshi, Asim Munawar, Ramón Fernandez Astudillo

Following the success of Proximal Policy Optimization (PPO) for Reinforcement Learning from Human Feedback (RLHF), new techniques such as Sequence Likelihood Calibration (SLiC) and Direct Policy Optimization (DPO) have been proposed that are offline in nature and use rewards in an indirect manner.

Text Generation

Pointwise Mutual Information Based Metric and Decoding Strategy for Faithful Generation in Document Grounded Dialogs

1 code implementation20 May 2023 Yatin Nandwani, Vineet Kumar, Dinesh Raghu, Sachindra Joshi, Luis A. Lastras

PMI quantifies the extent to which the document influences the generated response -- with a higher PMI indicating a more faithful response.

Response Generation

A Solver-Free Framework for Scalable Learning in Neural ILP Architectures

1 code implementation17 Oct 2022 Yatin Nandwani, Rishabh Ranjan, Mausam, Parag Singla

Experiments on several problems, both perceptual as well as symbolic, which require learning the constraints of an ILP, show that our approach has superior performance and scales much better compared to purely neural baselines and other state-of-the-art models that require solver-based training.

Matching Papers and Reviewers at Large Conferences

1 code implementation24 Feb 2022 Kevin Leyton-Brown, Mausam, Yatin Nandwani, Hedayat Zarkoob, Chris Cameron, Neil Newman, Dinesh Raghu

Peer-reviewed conferences, the main publication venues in CS, rely critically on matching highly qualified reviewers for each paper.

Neural Models for Output-Space Invariance in Combinatorial Problems

no code implementations ICLR 2022 Yatin Nandwani, Vidit Jain, Mausam, Parag Singla

One drawback of the proposed architectures, which are often based on Graph Neural Networks (GNN), is that they cannot generalize across the size of the output space from which variables are assigned a value, for example, set of colors in a GCP, or board-size in sudoku.

Node Classification

Neural Learning of One-of-Many Solutions for Combinatorial Problems in Structured Output Spaces

no code implementations ICLR 2021 Yatin Nandwani, Deepanshu Jindal, Mausam, Parag Singla

Our framework uses a selection module, whose goal is to dynamically determine, for every input, the solution that is most effective for training the network parameters in any given learning iteration.

A Primal Dual Formulation For Deep Learning With Constraints

1 code implementation NeurIPS 2019 Yatin Nandwani, Abhishek Pathak, Mausam, Parag Singla

In this paper, we present a constrained optimization formulation for training a deep network with a given set of hard constraints on output labels.

Entity Typing named-entity-recognition +4

Cannot find the paper you are looking for? You can Submit a new open access paper.