no code implementations • 19 Dec 2022 • Asish Ghoshal, Arash Einolghozati, Ankit Arun, Haoran Li, Lili Yu, Vera Gor, Yashar Mehdad, Scott Wen-tau Yih, Asli Celikyilmaz
Lack of factual correctness is an issue that still plagues state-of-the-art summarization systems despite their impressive progress on generating seemingly fluent summaries.
1 code implementation • 18 Nov 2022 • Minghan Li, Sheng-Chieh Lin, Barlas Oguz, Asish Ghoshal, Jimmy Lin, Yashar Mehdad, Wen-tau Yih, Xilun Chen
In this paper, we unify different multi-vector retrieval models from a token routing viewpoint and propose conditional token interaction via dynamic lexical routing, namely CITADEL, for efficient and effective multi-vector retrieval.
no code implementations • ICLR 2021 • Asish Ghoshal, Xilun Chen, Sonal Gupta, Luke Zettlemoyer, Yashar Mehdad
Training with soft targets instead of hard targets has been shown to improve performance and calibration of deep neural networks.
no code implementations • EMNLP 2021 • Kushal Lakhotia, Bhargavi Paranjape, Asish Ghoshal, Wen-tau Yih, Yashar Mehdad, Srinivasan Iyer
Natural language (NL) explanations of model predictions are gaining popularity as a means to understand and verify decisions made by large black-box pre-trained models, for NLP tasks such as Question Answering (QA) and Fact Verification.
1 code implementation • 29 Dec 2020 • Yilun Zhou, Adithya Renduchintala, Xian Li, Sida Wang, Yashar Mehdad, Asish Ghoshal
Active learning (AL) algorithms may achieve better performance with fewer data because the model guides the data selection process.
1 code implementation • EMNLP 2020 • Xilun Chen, Asish Ghoshal, Yashar Mehdad, Luke Zettlemoyer, Sonal Gupta
Task-oriented semantic parsing is a critical component of virtual assistants, which is responsible for understanding the user's intents (set reminder, play music, etc.).
no code implementations • 28 Jun 2019 • Asish Ghoshal, Kevin Bello, Jean Honorio
Discovering cause-effect relationships between variables from observational data is a fundamental challenge in many scientific disciplines.
no code implementations • 2 Jun 2019 • Kevin Bello, Asish Ghoshal, Jean Honorio
Structured prediction can be considered as a generalization of many standard supervised learning tasks, and is usually thought as a simultaneous prediction of multiple labels.
no code implementations • ICML 2018 • Asish Ghoshal, Jean Honorio
In this paper, we propose a provably polynomial time randomized algorithm for learning the parameters of perturbed MAP predictors.
no code implementations • 15 Jul 2017 • Asish Ghoshal, Jean Honorio
We develop a new algorithm --- which is computationally and statistically efficient and works in the high-dimensional regime --- for learning linear SEMs from purely observational data with arbitrary noise distribution.
no code implementations • 18 Jun 2017 • Asish Ghoshal, Jean Honorio
We also show that $\Omega(d \log (pm))$ samples are necessary for any method to consistently recover a game, with the same Nash-equilibria as the true game, from observations of strategic interactions.
no code implementations • NeurIPS 2017 • Asish Ghoshal, Jean Honorio
In this paper we propose a provably polynomial-time algorithm for learning sparse Gaussian Bayesian networks with equal noise variance --- a class of Bayesian networks for which the DAG structure can be uniquely identified from observational data --- under high-dimensional settings.
no code implementations • 3 Mar 2017 • Asish Ghoshal, Jean Honorio
In this paper we obtain sufficient and necessary conditions on the number of samples required for exact recovery of the pure-strategy Nash equilibria (PSNE) set of a graphical game from noisy observations of joint actions.
no code implementations • 11 Jul 2016 • Asish Ghoshal, Jean Honorio
In this paper we study the problem of exact recovery of the pure-strategy Nash equilibria (PSNE) set of a graphical game from noisy observations of joint actions of the players alone.
no code implementations • 27 Jan 2016 • Asish Ghoshal, Jean Honorio
In this paper, we study the information-theoretic limits of learning the structure of Bayesian networks (BNs), on discrete as well as continuous random variables, from a finite number of samples.